AWS Certified DAS-C01 – 80 Questions & Answers

AWS Certified Data Analytics Specialty (DAS-C01) – Questions & Answers

The AWS Certified Data Analytics – Specialty (DAS-C01) certification is designed for individuals who perform complex data analysis. This certification validates technical expertise in using AWS data lakes and analytics services to get insights from data. To earn this certification, candidates need to demonstrate knowledge of AWS data analytics services, the ability to design and build data analytics solutions, and understand the security best practices and compliance requirements.

AWS (DAS-C01) – 80 Questions & Answers

0

AWS Certified Data Analytics Specialty (DAS-C01)

AWS Certified Data Analytics Specialty (DAS-C01)

AWS Certified Data Analytics Specialty (DAS-C01)

Questions will be picked at random from the question bank.

You can use the NEXT button to move to the next question, use the PREV button to move to the previous question, the CLEAR button to clear any answer of your choice and you have the FINISH button to end the exam if you choose to.

Any question not answered before the end of the exam time, will be marked as wrong and the exam will end by itself. So try to attempt all questions on time.

Goodluck!

1 / 80

1. A technology company is creating a dashboard that will visualize and analyze time-sensitive data.
The data will come in through Amazon Kinesis Data Firehose with the butter interval set to 60
seconds. The dashboard must support near-real-time data. Which visualization solution will meet
these requirements?

2 / 80

2. A US-based sneaker retail company launched its global website. All the transaction
data is stored in Amazon RDS and curated historic transaction data is stored in
Amazon Redshift in the us-east-1 Region. The business intelligence (BI) team
wants to enhance the user experience by providing a dashboard for sneaker
trends. The BI team decides to use Amazon QuickSight to render the website
dashboards. During development, a team in Japan provisioned Amazon
QuickSight in ap-northeast-1. The team is having difficulty connecting Amazon
QuickSight from ap-northeast-1 to Amazon Redshift in us-east-1. Which solution
will solve this issue and meet the requirements?

3 / 80

3. A company currently uses Amazon Athena to query its global datasets. The
regional data is stored in Amazon S3 in the us-east-1 and us-west-2 Regions. The
data is not encrypted. To simplify the query process and manage it centrally, the
company wants to use Athena in us-west-2 to query data from Amazon S3 in both
Regions. The solution should be as low-cost as possible. What should the
company do to achieve this goal?

4 / 80

4. A large ride-sharing company has thousands of drivers globally serving millions of
unique customers every day. The company has decided to migrate an existing
data mart to Amazon Redshift. The existing schema includes the following tables.
A trips fact table for information on completed rides. A drivers dimension table for
driver profiles. A customers fact table holding customer profile information. The
company analyzes trip details by date and destination to examine profitability by
region. The drivers data rarely changes. The customers data frequently changes.
What table design provides optimal query performance?

5 / 80

5. A retail company is building its data warehouse solution using Amazon Redshift. As a part of that
effort, the company is loading hundreds of files into the fact table created in its Amazon Redshift
cluster. The company wants the solution to achieve the highest throughput and optimally use
cluster resources when loading data into the company's fact table. How should the company meet
these requirements?

6 / 80

6. A company has a business unit uploading .csv files to an Amazon S3 bucket. The
company's data platform team has set up an AWS Glue crawler to do discovery,
and create tables and schemas. An AWS Glue job writes processed data from the
created tables to an Amazon Redshift database. The AWS Glue job handles column
mapping and creating the Amazon Redshift table appropriately. When the AWS
Glue job is rerun for any reason in a day, duplicate records are introduced into the
Amazon Redshift table. Which solution will update the Redshift table without
duplicates when jobs are rerun?

7 / 80

7. A company's marketing team has asked for help in identifying a high performing long-term storage
service for their data based on the following requirements: The data size is approximately 32 TB
uncompressed. There is a low volume of single-row inserts each day. There is a high volume of
aggregation queries each day. Multiple complex joins are performed. The queries typically involve a
small subset of the columns in a table. Which storage service will provide the MOST performant
solution?

8 / 80

8. A global company has different sub-organizations, and each sub-organization sells its products and
services in various countries. The company's senior leadership wants to quickly identify which suborganization is the strongest performer in each country. All sales data is stored in Amazon S3 in
Parquet format. Which approach can provide the visuals that senior leadership requested with the
least amount of effort?

9 / 80

9. Once a month, a company receives a 100 MB .csv file compressed with gzip. The file contains
50,000 property listing records and is stored in Amazon S3 Glacier. The company needs its data
analyst to query a subset of the data for a specific vendor. What is the most cost-effective solution?

10 / 80

10. A large financial company is running its ETL process. Part of this process is to move data from
Amazon S3 into an Amazon Redshift cluster. The company wants to use the most cost-efficient
method to load the dataset into Amazon Redshift. Which combination of steps would meet these
requirements? (Choose TWO)

11 / 80

11. A banking company is currently using an Amazon Redshift cluster with dense
storage (DS) nodes to store sensitive data. An audit found that the cluster is
unencrypted. Compliance requirements state that a database with sensitive data
must be encrypted through a hardware security module (HSM) with automated
key rotation. Which combination of steps is required to achieve compliance?
(Choose TWO)

12 / 80

12. A data analyst is using Amazon QuickSight for data visualization across multiple
datasets generated by applications. Each application stores files within a separate
Amazon S3 bucket. AWS Glue Data Catalog is used as a central catalog across all
application data in Amazon S3. A new application stores its data within a separate
S3 bucket. After updating the catalog to include the new application data source,
the data analyst created a new Amazon QuickSight data source from an Amazon
Athena table, but the import into SPICE failed. How should the data analyst resolve
the issue?

13 / 80

13. A financial services company needs to aggregate daily stock trade data from the
exchanges into a data store. The company requires that data be streamed directly
into the data store, but also occasionally allows data to be modified using SQL.
The solution should integrate complex, analytic queries running with minimal
latency. The solution must provide a business intelligence dashboard that enables
viewing of the top contributors to anomalies in stock prices. Which solution meets
the company's requirements?

14 / 80

14. A data analyst is using AWS Glue to organize, cleanse, validate, and format a 200
GB dataset. The data analyst triggered the job to run with the Standard worker
type. After 3 hours, the AWS Glue job status is still RUNNING. Logs from the job
run show no error codes. The data analyst wants to improve the job execution
time without overprovisioning. Which actions should the data analyst take?

15 / 80

15. A mortgage company has a microservice for accepting payments. This microservice uses the
Amazon DynamoDB encryption client with AWS KMS managed keys to encrypt the sensitive data
before writing the data to DynamoDB. The finance team should be able to load this data into
Amazon Redshift and aggregate the values within the sensitive fields. The Amazon Redshift cluster
is shared with other data analysts from different business units. Which steps should a data analyst
take to accomplish this task efficiently and securely?

16 / 80

16. A company that produces network devices has millions of users. Data is collected
from the devices on an hourly basis and stored in an Amazon S3 data lake. The
company runs analyses on the last 24 hours of data flow logs for abnormality
detection and to troubleshoot and resolve user issues. The company also analyzes
historical logs dating back 2 years to discover patterns and look for improvement
opportunities. The data flow logs contain many metrics, such as date, timestamp,
source IP, and target IP. There are about 10 billion events every day. How should
this data be stored for optimal performance?

17 / 80

17. A company is planning to do a proof of concept for a machine learning (ML)
project using Amazon SageMaker with a subset of existing on-premises data
hosted in the company's 3 TB data warehouse. For part of the project, AWS Direct
Connect is established and tested. To prepare the data for ML, data analysts are
performing data curation. The data analysts want to perform multiple step,
including mapping, dropping null fields, resolving choice, and splitting fields. The
company needs the fastest solution to curate the data for this project. Which
solution meets these requirements?

18 / 80

18. A company wants to improve the data load time of a sales data dashboard. Data has been collected
as .csv files and stored within an Amazon S3 bucket that is partitioned by date. The data is then
loaded to an Amazon Redshift data warehouse for frequent analysis. The data volume is up to 500
GB per day. Which solution will improve the data loading performance?

19 / 80

19. A marketing company wants to improve its reporting and business intelligence
capabilities. During the planning phase, the company interviewed the relevant
stakeholders and discovered that: The operations team reports are run hourly for
the current month's data. The sales team wants to use multiple Amazon
QuickSight dashboards to show a rolling view of the last 30 days based on several
categories. The sales team also wants to view the data as soon as it reaches the
reporting backend. The finance team's reports are run daily for last month's data
and once a month for the last 24 months of data. Currently, there is 400 TB of data
in the system with an expected additional 100 TB added every month. The
company is looking for a solution that is as costeffective as possible. Which
solution meets the company's requirements?

20 / 80

20. A retail company's data analytics team recently created multiple product sales analysis dashboards
for the average selling price per product using Amazon QuickSight. The dashboards were created
from .csv files uploaded to Amazon S3. The team is now planning to share the dashboards with the
respective external product owners by creating individual users in Amazon QuickSight. For
compliance and governance reasons, restricting access is a key requirement. The product owners
should view only their respective product analysis in the dashboard reports. Which approach
should the data analytics team take to allow product owners to view only their products in the
dashboard?

21 / 80

21. A company uses Amazon Redshift as its data warehouse. A new table has columns that contain
sensitive data. The data in the table will eventually be referenced by several existing queries that
run many times a day. A data analyst needs to load 100 billion rows of data into the new table.
Before doing so, the data analyst must ensure that only members of the auditing group can read
the columns containing sensitive data. How can the data analyst meet these requirements with the
lowest maintenance overhead?

22 / 80

22. A company has developed an Apache Hive script to batch process data stared in Amazon S3. The
script needs to run once every day and store the output in Amazon S3. The company tested the
script, and it completes within 30 minutes on a small local three-node cluster. Which solution is the
MOST cost-effective for scheduling and executing the script?

23 / 80

23. A bank operates in a regulated environment. The compliance requirements for the country in which
the bank operates say that customer data for each state should only be accessible by the bank's
employees located in the same state. Bank employees in one state should NOT be able to access
data for customers who have provided a home address in a different state. The bank's marketing
team has hired a data analyst to gather insights from customer data for a new campaign being
launched in certain states. Currently, data linking each customer account to its home state is stored
in a tabular .csv file within a single Amazon S3 folder in a private S3 bucket. The total size of the S3
folder is 2 GB uncompressed. Due to the country's compliance requirements, the marketing team is
not able to access this folder. The data analyst is responsible for ensuring that the marketing team
gets one-time access to customer data for their campaign analytics project, while being subject to
all the compliance requirements and controls. Which solution should the data analyst implement to
meet the desired requirements with the LEAST amount of setup effort?

24 / 80

24. A mobile gaming company wants to capture data from its gaming app and make the data available
for analysis immediately. The data record size will be approximately 20 KB. The company is
concerned about achieving optimal throughput from each device. Additionally, the company wants
to develop a data stream processing application with dedicated throughput for each consumer.
Which solution would achieve this goal?

25 / 80

25. A data engineering team within a shared workspace company wants to build a centralized logging
system for all weblogs generated by the space reservation system. The company has a fleet of
Amazon EC2 instances that process requests for shared space reservations on its website. The data
engineering team wants to ingest all weblogs into a service that will provide a near-real-time search
engine. The team does not want to manage the maintenance and operation of the logging system.
Which solution allows the data engineering team to efficiently set up the web logging system
within AWS?

26 / 80

26. A company is streaming its high-volume billing data (100 MBps) to Amazon
Kinesis Data Streams. A data analyst partitioned the data on account_id to ensure
that all records belonging to an account go to the same Kinesis shard and order is
maintained. While building a custom consumer using the Kinesis Java SDK, the
data analyst notices that, sometimes, the messages arrive out of order for
account_id. Upon further investigation, the data analyst discovers the messages
that are out of order seem to be arriving from different shards for the same
account_id and are seen when a stream resize runs. What is an explanation for this
behavior and what is the solution?

27 / 80

27. A company is building a data lake and needs to ingest data from a relational database that has
time-series data. The company wants to use managed services to accomplish this. The process
needs to be scheduled daily and bring incremental data only from the source into Amazon S3.
What is the MOST cost-effective approach to meet these requirements?

28 / 80

28. A financial company hosts a data lake in Amazon S3 and a data warehouse on an
Amazon Redshift cluster. The company uses Amazon QuickSight to build
dashboards and wants to secure access from its on-premises Active Directory to
Amazon QuickSight. How should the data be secured?

29 / 80

29. A large university has adopted a strategic goal of increasing diversity among enrolled students. The
data analytics team is creating a dashboard with data visualizations to enable stakeholders to view
historical trends. All access must be authenticated using Microsoft Active Directory. All data in
transit and at rest must be encrypted. Which solution meets these requirements?

30 / 80

30. An ecommerce company stores customer purchase data in Amazon RDS. The company wants a
solution to store and analyze historical data. The most recent 6 months of data will be queried
frequently for analytics workloads. This data is several terabytes large. Once a month, historical data
for the last 5 years must be accessible and will be joined with the more recent data. The company
wants to optimize performance and cost. Which storage solution will meet these requirements?

31 / 80

31. A smart home automation company must efficiently ingest and process messages from various
connected devices and sensors. The majority of these messages are comprised of a large number of
small files. These messages are ingested using Amazon Kinesis Data Streams and sent to Amazon
S3 using a Kinesis data stream consumer application. The Amazon S3 message data is then passed
through a processing pipeline built on Amazon EMR running scheduled PySpark jobs. The data
platform team manages data processing and is concerned about the efficiency and cost of
downstream data processing. They want to continue to use PySpark. Which solution improves the
efficiency of the data processing jobs and is well architected?

32 / 80

32. A streaming application is reading data from Amazon Kinesis Data Streams and
immediately writing the data to an Amazon S3 bucket every 10 seconds. The
application is reading data from hundreds of shards. The batch interval cannot be
changed due to a separate requirement. The data is being accessed by Amazon
Athena. Users are seeing degradation in query performance as time progresses.
Which action can help improve query performance?

33 / 80

33. An online retail company with millions of users around the globe wants to improve its ecommerce
analytics capabilities. Currently, clickstream data is uploaded directly to Amazon S3 as compressed
files. Several times each day, an application running on Amazon EC2 processes the data and makes
search options and reports available for visualization by editors and marketers. The company wants
to make website clicks and aggregated data available to editors and marketers in minutes to enable
them to connect with users more effectively. Which options will help meet these requirements in
the MOST efficient way? (Choose TWO)

34 / 80

34. A team of data scientists plans to analyze market trend data for their company's
new investment strategy. The trend data comes from five different data sources in
large volumes. The team wants to utilize Amazon Kinesis to support their use case.
The team uses SQL-like queries to analyze trends and wants to send notifications
based on certain significant patterns in the trends. Additionally, the data scientists
want to save the data to Amazon S3 for archival and historical reprocessing, and
use AWS managed services wherever possible. The team wants to implement the
lowest-cost solution. Which solution meets these requirements?

35 / 80

35. A company that monitors weather conditions from remote construction sites is setting up a
solution to collect temperature data from the following two weather stations. Station A, which has
10 sensors. Station B, which has five sensors.These weather stations were placed by onsite subjectmatter experts. Each sensor has a unique ID. The data collected from each sensor will be collected
using Amazon Kinesis Data Streams. Based on the total incoming and outgoing data throughput, a
single Amazon Kinesis data stream with two shards is created. Two partition keys are created based
on the station names. During testing, there is a bottleneck on data coming from Station A, but not
from StationB. Upon review, it is confirmed that the total stream throughput is still less than the
allocated Kinesis Data Streams throughput. How can this bottleneck be resolved without increasing
the overall cost and complexity of the solution, while retaining the data collection quality
requirements?

36 / 80

36. A university intends to use Amazon Kinesis Data Firehose to collect JSON-formatted batches of
water quality readings in Amazon S3. The readings are from 50 sensors scattered across a local lake.
Students will query the stored data using Amazon Athena to observe changes in a captured metric
over time, such as water temperature or acidity. Interest has grown in the study, prompting the
university to reconsider how data will be stored. Which data format and partitioning choices will
MOST significantly reduce costs? (Choose TWO)

37 / 80

37. An insurance company has raw data in JSON format that is sent without a
predefined schedule through an Amazon Kinesis Data Firehose delivery stream to
an Amazon S3 bucket. An AWS Glue crawler is scheduled to run every 8 hours to
update the schema in the data catalog of the tables stored in the S3 bucket. Data
analysts analyze the data using Apache Spark SQL on Amazon EMR set up with
AWS Glue Data Catalog as the metastore. Data analysts say that, occasionally, the
data they receive is stale. A data engineer needs to provide access to the most up to-date data. Which solution meets these requirements?

38 / 80

38. A financial company uses Amazon S3 as its data lake and has set up a data warehouse using a
multi-node Amazon Redshift cluster. The data files in the data lake are organized in folders based
on the data source of each data file. All the data files are loaded to one table in the Amazon
Redshift cluster using a separate COPY command for each data file location. With this approach,
loading all the data files into Amazon Redshift takes a long time to complete. Users want a faster
solution with little or no increase in cost while maintaining the segregation of the data files in the
S3 data lake. Which solution meets these requirements?

39 / 80

39. Three teams of data analysts use Apache Hive on an Amazon EMR cluster with the EMR File System
(EMRFS) to query data stored within each teams Amazon S3 bucket. The EMR cluster has Kerberos
enabled and is configured to authenticate users from the corporate Active Directory. The data is
highly sensitive, so access must be limited to the members of each team. Which steps will satisfy
the security requirements?

40 / 80

40. A media company has been performing analytics on log data generated by its applications. There
has been a recent increase in the number of concurrent analytics jobs running, and the overall
performance of existing jobs is decreasing as the number of new jobs is increasing. The partitioned
data is stored in Amazon S3 One Zone-Infrequent Access (S3 One Zone-IA) and the analytic
processing is performed on Amazon EMR clusters using the EMR File System (EMRFS) with
consistent view enabled. A data analyst has determined that it is taking longer for the EMR task
nodes to list objects in Amazon S3. Which action would MOST likely increase the performance of
accessing log data in Amazon S3?

41 / 80

41. A company wants to optimize the cost of its data and analytics platform. The company is ingesting
a number of .csv and JSON files in Amazon S3 from various data sources. Incoming data is expected
to be 50 GB each day. The company is using Amazon Athena to query the raw data in Amazon S3
directly. Most queries aggregate data from the past 12 months, and data that is older than 5 years
is infrequently queried. The typical query scans about 500 MB of data and is expected to return
results in less than 1 minute. The raw data must be retained indefinitely for compliance
requirements. Which solution meets the company's requirements?

42 / 80

42. A media company wants to perform machine learning and analytics on the data
residing in its Amazon S3 data lake. There are two data transformation
requirements that will enable the consumers within the company to create reports:
Daily transformations of 300 GB of data with different file formats landing in
Amazon S3 at a scheduled time. One-time transformations of terabytes of
archived data residing in the S3 data lake. Which combination of solutions costeffectively meets the company's requirements for transforming the data? (Choose
THREE)

43 / 80

43. A company is planning to create a data lake in Amazon S3. The company wants to create tiered
storage based on access patterns and cost objectives. The solution must include support for JDBC
connections from legacy clients, metadata management that allows federation for access control,
and batch-based ETL using PySpark and Scala Operational management should be limited. Which
combination of components can meet these requirements? (Choose THREE)

44 / 80

44. A regional energy company collects voltage data from sensors attached to buildings. To address
any known dangerous conditions, the company wants to be alerted when a sequence of two
voltage drops is detected within 10 minutes of a voltage spike at the same building. It is important
to ensure that all messages are delivered as quickly as possible. The system must be fully managed
and highly available. The company also needs a solution that will automatically scale up as it covers
additional cites with this monitoring feature. The alerting system is subscribed to an Amazon SNS
topic for remediation. Which solution meets these requirements?

45 / 80

45. A hospital uses wearable medical sensor devices to collect data from patients. The
hospital is architecting a near-real-time solution that can ingest the data securely
at scale. The solution should also be able to remove the patient's protected health
information (PHI) from the streaming data and store the data in durable storage.
Which solution meets these requirements with the least operational overhead?

46 / 80

46. A company wants to improve user satisfaction for its smart home system by
adding more features to its recommendation engine. Each sensor asynchronously
pushes its nested JSON data into Amazon Kinesis Data Streams using the Kinesis
Producer Library (KPL) in Java. Statistics from a set of failed sensors showed that,
when a sensor is malfunctioning, its recorded data is not always sent to the cloud.
The company needs a solution that offers near-real-time analytics on the data
from the most updated sensors. Which solution enables the company to meet
these requirements?

47 / 80

47. A company's data analyst needs to ensure that queries run in Amazon Athena cannot scan more
than a prescribed amount of data for cost control purposes. Queries that exceed the prescribed
threshold must be canceled immediately. What should the data analyst do to achieve this?

48 / 80

48. A marketing company is using Amazon EMR clusters for its workloads. The company manually
installs third-party libraries on the clusters by logging in to the master nodes. A data analyst needs
to create an automated solution to replace the manual process. Which options can fulfill these
requirements? (Choose TWO)

49 / 80

49. A company has a data warehouse in Amazon Redshift that is approximately 500 TB in size. New
data is imported every few hours and read-only queries are run throughout the day and evening.
There is a particularly heavy load with no writes for several hours each morning on business days.
During those hours, some queries are queued and take a long time to execute. The company needs
to optimize query execution and avoid any downtime. What is the MOST cost-effective solution?

50 / 80

50. A company uses the Amazon Kinesis SDK to write data to Kinesis Data Streams. Compliance
requirements state that the data must be encrypted at rest using a key that can be rotated. The
company wants to meet this encryption requirement with minimal coding effort. How can these
requirements be met?

51 / 80

51. A banking company wants to collect large volumes of transactional data using Amazon Kinesis Data
Streams for real-time analytics. The company uses PutRecord to send data to Amazon Kinesis, and
has observed network outages during certain times of the day. The company wants to obtain
exactly once semantics for the entire processing pipeline. What should the company do to obtain
these characteristics?

52 / 80

52. A media content company has a streaming playback application. The company wants to collect and
analyze the data to provide near-real-time feedback on playback issues. The company needs to
consume this data and return results within 30 seconds according to the service-level agreement
(SLA). The company needs the consumer to identify playback issues, such as quality during a
specified timeframe. The data will be emitted as JSON and may change schemas over time. Which
solution will allow the company to collect data for processing while meeting these requirements?

53 / 80

53. A media analytics company consumes a stream of social media posts. The posts are sent to an
Amazon Kinesis data stream partitioned on user_id. An AWS Lambda function retrieves the records
and validates the content before loading the posts into an Amazon OpenSearch Service (Amazon
Elasticsearch Service) cluster. The validation process needs to receive the posts for a given user in
the order they were received by the Kinesis data stream. During peak hours, the social media posts
take more than an hour to appear in the Amazon OpenSearch Service (Amazon ES) cluster. A data
analytics specialist must implement a solution that reduces this latency with the least possible
operational overhead. Which solution meets these requirements?

54 / 80

54. An airline has .csv-formatted data stored in Amazon S3 with an AWS Glue Data
Catalog. Data analysts want to join this data with call center data stored in
Amazon Redshift as part of a dally batch process. The Amazon Redshift cluster is
already under a heavy load. The solution must be managed, serverless, wellfunctioning, and minimize the load on the existing Amazon Redshift cluster. The
solution should also require minimal effort and development activity. Which
solution meets these requirements?

55 / 80

55. A company has 1 million scanned documents stored as image files in Amazon S3. The documents
contain typewritten application forms with information including the applicant first name, applicant
last name, application date, application type, and application text. The company has developed a
machine learning algorithm to extract the metadata values from the scanned documents. The
company wants to allow internal data analysts to analyze and find applications using the applicant
name, application date, or application text. The original images should also be downloadable. Cost
control is secondary to query performance. Which solution organizes the images and metadata to
drive insights while meeting the requirements?

56 / 80

56. A company leverages Amazon Athena for ad-hoc queries against data stored in Amazon S3. The
company wants to implement additional controls to separate query execution and query history
among users, teams, or applications running in the same AWS account to comply with internal
security policies. Which solution meets these requirements?

57 / 80

57. A company analyzes its data in an Amazon Redshift data warehouse, which currently has a cluster
of three dense storage nodes. Due to a recent business acquisition, the company needs to load an
additional 4 TB of user data into Amazon Redshift. The engineering team will combine all the user
data and apply complex calculations that require I/O intensive resources. The company needs to
adjust the cluster's capacity to support the change in analytical and storage requirements. Which
solution meets these requirements?

58 / 80

58. A company wants to enrich application logs in near-real-time and use the enriched dataset for
further analysis. The application is running on Amazon EC2 instances across multiple Availability
Zones and storing its logs using Amazon CloudWatch Logs. The enrichment source is stored in an
Amazon DynamoDB table. Which solution meets the requirements for the event collection and
enrichment?

59 / 80

59. An airline has been collecting metrics on flight activities for analytics. A recently completed proof of
concept demonstrates how the company provides insights to data analysts to improve on-time
departures. The proof of concept used objects in Amazon S3, which contained the metrics in .csv
format, and used Amazon Athena for querying the data. As the amount of data increases, the data
analyst wants to optimize the storage solution to improve query performance. Which options
should the data analyst use to improve performance as the data lake grows? (Choose THREE)

60 / 80

60. A company launched a service that produces millions of messages every day and uses Amazon
Kinesis Data Streams as the streaming service. The company uses the Kinesis SDK to write data to
Kinesis Data Streams. A few months after launch, a data analyst found that write performance is
significantly reduced. The data analyst investigated the metrics and determined that Kinesis is
throttling the write requests. The data analyst wants to address this issue without significant
changes to the architecture. Which actions should the data analyst take to resolve this issue?
(Choose TWO)

61 / 80

61. An Amazon Redshift database contains sensitive user data. Logging is necessary to meet
compliance requirements. The logs must contain database authentication attempts, connections,
and disconnections. The logs must also contain each query run against the database and record
which database user ran each query. Which steps will create the required logs?

62 / 80

62. A company is planning to do a proof of concept for a machine earning (ML)
project using Amazon SageMaker with a subset of existing on-premises data
hosted in the company's 3 TB data warehouse. For part of the project, AWS Direct
Connect is established and tested. To prepare the data for ML, data analysts are
performing data curation. The data analysts want to perform multiple step,
including mapping, dropping null fields, resolving choice, and splitting fields. The
company needs the fastest solution to curate the data for this project. Which
solution meets these requirements?

63 / 80

63. A company is migrating its existing on-premises ETL jobs to Amazon EMR. The
code consists of a series of jobs written in Java. The company needs to reduce
overhead for the system administrators without changing the underlying code.
Due to the sensitivity of the data, compliance requires that the company use root
device volume encryption on all nodes in the cluster. Corporate standards require
that environments be provisioned though AWS CloudFormation when possible.
Which solution satisfies these requirements?

64 / 80

64. A company uses Amazon OpenSearch Service (Amazon Elasticsearch Service) to
store and analyze its website clickstream data. The company ingests 1 TB of data
daily using Amazon Kinesis Data Firehose and stores one day's worth of data in an
Amazon ES cluster. The company has very slow query performance on the Amazon
ES index and occasionally sees errors from Kinesis Data Firehose when attempting
to write to the index. The Amazon ES cluster has 10 nodes running a single index
and 3 dedicated master nodes. Each data node has 1.5 TB of Amazon EBS storage
attached and the cluster is configured with 1,000 shards. Occasionally,
JVMMemoryPressure errors are found in the cluster logs. Which solution will
improve the performance of Amazon ES?

65 / 80

65. A company has developed several AWS Glue jobs to validate and transform its data from Amazon
S3 and load it into Amazon RDS for MySQL in batches once every day. The ETL jobs read the S3
data using a DynamicFrame. Currently, the ETL developers are experiencing challenges in
processing only the incremental data on every run, as the AWS Glue job processes all the S3 input
data on each run. Which approach would allow the developers to solve the issue with minimal
coding effort?

66 / 80

66. A real estate company has a mission-critical application using Apache HBase in
Amazon EMR. Amazon EMR is configured with a single master node. The company
has over 5 TB of data stored on an Hadoop Distributed File System (HDFS). The
company wants a cost-effective solution to make its HBase data highly available.
Which architectural pattern meets company's requirements?

67 / 80

67. A data analyst is designing a solution to interactively query datasets with SQL using a JDBC
connection. Users will join data stored in Amazon S3 in Apache ORC format with data stored in
Amazon Elasticsearch Service (Amazon ES) and Amazon Aurora MySQL. Which solution will provide
the MOST up-to-date results?

68 / 80

68. A transportation company uses IoT sensors attached to trucks to collect vehicle
data for its global delivery fleet. The company currently sends the sensor data in
small .csv files to Amazon S3. The files are then loaded into a 10-node Amazon
Redshift cluster with two slices per node and queried using both Amazon Athena
and Amazon Redshift. The company wants to optimize the files to reduce the cost
of querying and also improve the speed of data loading into the Amazon Redshift
cluster. Which solution meets these requirements?

69 / 80

69. A healthcare company uses AWS data and analytics tools to collect, ingest, and store electronic
health record (EHR) data about its patients. The raw EHR data is stored in Amazon S3 in JSON
format partitioned by hour, day, and year and is updated every hour. The company wants to
maintain the data catalog and metadata in an AWS Glue Data Catalog to be able to access the data
using Amazon Athena or Amazon Redshift Spectrum for analytics. When defining tables in the Data
Catalog, the company has the following requirements: Choose the catalog table name and do not
rely on the catalog table naming algorithm. Keep the table updated with new partitions loaded in
the respective S3 bucket prefixes. Which solution meets these requirements with minimal effort?

70 / 80

70. A company wants to use an automatic machine learning (ML) Random Cut Forest (RCF) algorithm
to visualize complex real-word scenarios, such as detecting seasonality and trends, excluding
outers, and imputing missing values. The team working on this project is non-technical and is
looking for an out-of-the-box solution that will require the LEAST amount of management
overhead. Which solution will meet these requirements?

71 / 80

71. A manufacturing company has been collecting IoT sensor data from devices on its
factory floor for a year and is storing the data in Amazon Redshift for daily
analysis. A data analyst has determined that, at an expected ingestion rate of
about 2 TB per day, the cluster will be undersized in less than 4 months. A longterm solution is needed. The data analyst has indicated that most queries only
reference the most recent 13 months of data, yet there are also quarterly reports
that need to query all the data generated from the past 7 years. The chief
technology officer (CTO) is concerned about the costs, administrative effort, and
performance of a long-term solution. Which solution should the data analyst use
to meet these requirements?

72 / 80

72. An energy company collects voltage data in real time from sensors that are
attached to buildings. The company wants to receive notifications when a
sequence of two voltage drops is detected within 10 minutes of a sudden voltage
increase at the same building. All notifications must be delivered as quickly as
possible. The system must be highly available. The company needs a solution that
will automatically scale when this monitoring feature is implemented in other
cities. The notification system is subscribed to an Amazon Simple Notification
Service (Amazon SNS) topic for remediation. Which solution will meet these
requirements?

73 / 80

73. A media company has a streaming playback application. The company needs to collect and analyze
data to provide near-real-time feedback on playback issues within 30 seconds. The company
requires a consumer application to identify playback issues, such as decreased quality during a
specified time frame. The data will be streamed in JSON format. The schema can change over time.
Which solution will meet these requirements?

74 / 80

74. A software company hosts an application on AWS, and new features are released
weekly. As part of the application testing process, a solution must be developed
that analyzes logs from each Amazon EC2 instance to ensure that the application
is working as expected after each deployment. The collection and analysis solution
should be highly available with the ability to display new information with minimal
delays. Which method should the company use to collect and analyze the logs?

75 / 80

75. A company stores its sales and marketing data that includes personally identifiable information (PII)
in Amazon S3. The company allows its analysts to launch their own Amazon EMR cluster and run
analytics reports with the data. To meet compliance requirements, the company must ensure the
data is not publicly accessible throughout this process. A data engineer has secured Amazon S3 but
must ensure the individual EMR clusters created by the analysts are not exposed to the public
internet. Which solution should the data engineer to meet this compliance requirement with LEAST
amount of effort?

76 / 80

76. A financial company uses Apache Hive on Amazon EMR for ad-hoc queries. Users are complaining
of sluggish performance. A data analyst notes the following: Approximately 90% of queries are
submitted 1 hour after the market opens. Hadoop Distributed File System (HDFS) utilization never
exceeds 10%. Which solution would help address the performance issues?

77 / 80

77. A large company has a central data lake to run analytics across different
departments. Each department uses a separate AWS account and stores its data in
an Amazon S3 bucket in that account. Each AWS account uses the AWS Glue Data
Catalog as its data catalog. There are different data lake access requirements
based on roles. Associate analysts should only have read access to their
departmental data. Senior data analysts can have access in multiple departments
including theirs, but for a subset of columns only. Which solution achieves these
required access patterns to minimize costs and administrative tasks?

78 / 80

78. A company developed a new elections reporting website that uses Amazon Kinesis Data Firehose to
deliver full logs from AWS WAF to an Amazon S3 bucket. The company is now seeking a low-cost
option to perform this infrequent data analysis with visualizations of logs in a way that requires
minimal development effort. Which solution meets these requirements? (Choose TWO)

79 / 80

79. A large company receives files from external parties in Amazon EC2 throughout
the day. At the end of the day, the files are combined into a single file,
compressed into a gzip file, and uploaded to Amazon S3. The total size of all the
files is close to 100 GB daily. Once the files are uploaded to Amazon S3, an AWS
Batch program executes a COPY command to load the files into an Amazon
Redshift cluster. Which program modification will accelerate the COPY process?

80 / 80

80. A company wants to research user turnover by analyzing the past 3 months of user activities. With
millions of users, 1.5 TB of uncompressed data is generated each day. A 30-node Amazon Redshift
cluster with 2.56 TB of solid state drive (SSD) storage for each node is required to meet the query
performance goals. The company wants to run an additional analysis on a year's worth of historical
data to examine trends indicating which features are most popular. This analysis will be done once
a week. What is the MOST cost-effective solution?

Your score is

0%

To prepare effectively for the DAS-C01 exam, it’s crucial to practice past questions and review answers. This preparation strategy helps you become familiar with the exam format, understand the types of questions that are commonly asked, and identify areas where you need further study. Practicing past questions also enhances your ability to manage time during the actual exam, as you become accustomed to the pace at which you need to answer questions.

There are various resources available for practicing past questions, including online platforms, study guides, and practice exams. These resources often come with detailed explanations for each answer, helping you understand the reasoning behind correct answers and learn from your mistakes. Additionally, participating in study groups or forums can provide valuable insights and tips from those who have already passed the exam.

Consistency and thoroughness in preparation are key. Dedicate regular study time, review AWS documentation, and stay updated with any changes in AWS services. By systematically practicing past questions and understanding the underlying concepts, you can build confidence and increase your chances of passing the AWS Certified Data Analytics – Specialty exam. Achieving this certification can significantly enhance your career prospects in the field of data analytics, showcasing your expertise in leveraging AWS services to derive valuable insights from data.

Practice these:

AWS Certified Cloud Practitioner -100 Questions & Answers (Part 2)

AWS Certified Cloud Practitioner -100 Questions & Answers (Part 1)

Free AWS Certified Cloud Practitioner Questions & Answers

AWS Certified Advanced Networking – Specialty | 90 Questions & Answers

Free AWS Certified Advanced Networking (ANS-C01) Questions & Answers

Leave a Comment

Your email address will not be published. Required fields are marked *

error: Content is protected !!
Scroll to Top