Amazon DAS-C01 Exam (page: 4)
Amazon AWS Certified Data Analytics - Specialty (DAS-C01)
Updated on: 29-Mar-2026

Viewing Page 4 of 22

A media company has a streaming playback application. The company needs to collect and analyze data to provide near-real-time feedback on playback issues within 30 seconds. The company requires a consumer application to identify playback issues, such as decreased quality during a speci ed time frame. The data will be streamed in JSON format. The schema can change over time.
Which solution will meet these requirements?

  1. Send the data to Amazon Kinesis Data Firehose with delivery to Amazon S3. Con gure an S3 event to invoke an AWS Lambda function to process and analyze the data.
  2. Send the data to Amazon Managed Streaming for Apache Kafka. Con gure Amazon Kinesis Data Analytics for SQL Application as the consumer application to process and analyze the data.
  3. Send the data to Amazon Kinesis Data Firehose with delivery to Amazon S3. Con gure Amazon S3 to initiate an event for AWS Lambda to process and analyze the data.
  4. Send the data to Amazon Kinesis Data Streams. Con gure an Amazon Kinesis Data Analytics for Apache Flink application as the consumer application to process and analyze the data.

Answer(s): D

Explanation:

D) Correct because Kinesis Data Streams with a Kinesis Data Analytics for Apache Flink app supports real-time, low-latency processing of streaming JSON with evolving schemas, suitable for near-instant playback issue detection within 30 seconds.
A) Firehose is designed for near-real-time delivery to destinations like S3, not low-latency analytics; Lambda from S3 events introduces additional latency and lacks streaming windowing for continuous analysis.
B) Managed Streaming for Kafka plus Kinesis Data Analytics SQL is not ideal for real-time, per-event analytic latency and may complicate schema evolution handling.
C) Firehose to S3 with Lambda is batch-oriented and unsuitable for continuous, low-latency streaming analytics and schema changes.



An ecommerce company stores customer purchase data in Amazon RDS. The company wants a solution to store and analyze historical data. The most recent 6 months of data will be queried frequently for analytics workloads. This data is several terabytes large. Once a month, historical data for the last 5 years must be accessible and will be joined with the more recent data. The company wants to optimize performance and cost.
Which storage solution will meet these requirements?

  1. Create a read replica of the RDS database to store the most recent 6 months of data. Copy the historical data into Amazon S3. Create an AWS Glue Data Catalog of the data in Amazon S3 and Amazon RDS. Run historical queries using Amazon Athena.
  2. Use an ETL tool to incrementally load the most recent 6 months of data into an Amazon Redshift cluster. Run more frequent queries against this cluster. Create a read replica of the RDS database to run queries on the historical data.
  3. Incrementally copy data from Amazon RDS to Amazon S3. Create an AWS Glue Data Catalog of the data in Amazon S3. Use Amazon Athena to query the data.
  4. Incrementally copy data from Amazon RDS to Amazon S3. Load and store the most recent 6 months of data in Amazon Redshift. Con gure an Amazon Redshift Spectrum table to connect to all historical data.

Answer(s): D

Explanation:

The correct choice combines cost-effective storage for long-term historical data with fast analytics on recent data and seamless cross-join capability: Redshift Spectrum enables querying both recent data in Redshift and historical data in S3 without moving data, providing performance for frequent 6-month analytics and scalable cost for multi-year history.
A) Incorrect: separates historical data in S3 but still relies on Athena for history; lacks integrated fast join with recent 6 months in Redshift, reducing join performance and analytics speed.
B) Incorrect: uses Redshift for recent data but keeps RDS read replica for history, adding complexity and potential latency; not a unified analytical model.
C) Incorrect: only S3 with Athena; no active storage for recent 6 months in a fast-access data warehouse, hindering performance for frequent queries.



A company leverages Amazon Athena for ad-hoc queries against data stored in Amazon S3. The company wants to implement additional controls to separate query execution and query history among users, teams, or applications running in the same AWS account to comply with internal security policies.
Which solution meets these requirements?

  1. Create an S3 bucket for each given use case, create an S3 bucket policy that grants permissions to appropriate individual IAM users. and apply the S3 bucket policy to the S3 bucket.
  2. Create an Athena workgroup for each given use case, apply tags to the workgroup, and create an IAM policy using the tags to apply appropriate permissions to the workgroup.
  3. Create an IAM role for each given use case, assign appropriate permissions to the role for the given use case, and add the role to associate the role with Athena.
  4. Create an AWS Glue Data Catalog resource policy for each given use case that grants permissions to appropriate individual IAM users, and apply the resource policy to the speci c tables used by Athena.

Answer(s): B

Explanation:

Athena workgroups isolate query execution and history per use case, allowing per-group settings, quotas, and history separation while sharing the same account and catalog.
A) Incorrect: S3 bucket policies alone separate data access but do not isolate query execution history or control per-user query context in Athena.
B) Correct: Athena workgroups provide distinct execution contexts and query history per use case; tagging and IAM policies scoped to workgroups enforce separation.
C) Incorrect: IAM roles per use case don’t inherently partition Athena query history or execution isolation within a single account unless combined with dedicated workspaces; complexity increases.
D) Incorrect: Glue Data Catalog resource policies govern table-level access but do not offer per-use-case query execution isolation or history separation in Athena.


Reference:

https://aws.amazon.com/athena/faqs/



A company wants to use an automatic machine learning (ML) Random Cut Forest (RCF) algorithm to visualize complex real-world scenarios, such as detecting seasonality and trends, excluding outers, and imputing missing values. The team working on this project is non-technical and is looking for an out-of-the-box solution that will require the LEAST amount of management overhead.
Which solution will meet these requirements?

  1. Use an AWS Glue ML transform to create a forecast and then use Amazon QuickSight to visualize the data.
  2. Use Amazon QuickSight to visualize the data and then use ML-powered forecasting to forecast the key business metrics.
  3. Use a pre-build ML AMI from the AWS Marketplace to create forecasts and then use Amazon QuickSight to visualize the data.
  4. Use calculated elds to create a new forecast and then use Amazon QuickSight to visualize the data.

Answer(s): B

Explanation:

A concise justification: QuickSight’s built-in ML-powered forecasting provides an out-of-the-box, low-management solution suitable for non-technical users, enabling visualization of seasonality, trends, missing value handling, and basic imputation without extra infrastructure.
A) Incorrect: AWS Glue ML transforms introduce additional ETL/ML setup and management, not as turnkey for visualization and forecasting as QuickSight forecasting.
B) Correct: QuickSight forecasting is integrated, managed, and user-friendly, minimizing overhead while delivering time series insights.
C) Incorrect: Pre-built ML AMIs require provisioning, maintenance, and potentially more technical skill than native QuickSight forecasting.
D) Incorrect: Calculated fields are manual and limited; they do not provide an automated ML-based forecast with seasonality/trend handling.


Reference:

https://aws.amazon.com/blogs/big-data/query-visualize-and-forecast-trufactor-web-session-intelligence-with-aws-data-exchange/



A retail company's data analytics team recently created multiple product sales analysis dashboards for the average selling price per product using Amazon
QuickSight. The dashboards were created from .csv les uploaded to Amazon S3. The team is now planning to share the dashboards with the respective external product owners by creating individual users in Amazon QuickSight. For compliance and governance reasons, restricting access is a key requirement. The product owners should view only their respective product analysis in the dashboard reports. Which approach should the data analytics team take to allow product owners to view only their products in the dashboard?

  1. Separate the data by product and use S3 bucket policies for authorization.
  2. Separate the data by product and use IAM policies for authorization.
  3. Create a manifest le with row-level security.
  4. Create dataset rules with row-level security.

Answer(s): D

Explanation:

A short summary: Use row-level security at the dataset level to filter data per product for each user.
A) Separate the data by product and use S3 bucket policies for authorization. — Incorrect: S3 bucket policies control access to storage, not per-user data filtering within QuickSight dashboards.
B) Separate the data by product and use IAM policies for authorization. — Incorrect: IAM policies grant access to AWS resources, not fine-grained row-level data visibility inside QuickSight dashboards.
C) Create a manifest le with row-level security. — Incorrect: manifests control data source metadata for ingestion, not per-user data access within QuickSight analyses.
D) Create dataset rules with row-level security. — Correct: DS-level row-level security (RLS) restricts data rows based on user identity or attributes, allowing product owners to see only their products in dashboards.



A company has developed an Apache Hive script to batch process data stared in Amazon S3. The script needs to run once every day and store the output in
Amazon S3. The company tested the script, and it completes within 30 minutes on a small local three-node cluster. Which solution is the MOST cost-effective for scheduling and executing the script?

  1. Create an AWS Lambda function to spin up an Amazon EMR cluster with a Hive execution step. Set KeepJobFlowAliveWhenNoSteps to false and disable the termination protection ag. Use Amazon CloudWatch Events to schedule the Lambda function to run daily.
  2. Use the AWS Management Console to spin up an Amazon EMR cluster with Python Hue. Hive, and Apache Oozie. Set the termination protection ag to true and use Spot Instances for the core nodes of the cluster. Con gure an Oozie work ow in the cluster to invoke the Hive script daily.
  3. Create an AWS Glue job with the Hive script to perform the batch operation. Con gure the job to run once a day using a time-based schedule.
  4. Use AWS Lambda layers and load the Hive runtime to AWS Lambda and copy the Hive script. Schedule the Lambda function to run daily by creating a work ow using AWS Step Functions.

Answer(s): A

Explanation:

A) The most cost-effective approach is to use a Lambda function to spin up a transient EMR cluster for the Hive job, then terminate after completion. This minimizes idle cost since clusters run only on demand, aligns with a daily batch window, and CloudWatch Events provides simple scheduling without maintaining always-on infrastructure.
B) Using Hue, Oozie, and termination protection with Spot Instances adds management overhead and potential reliability concerns; it’s more costly and complex for a daily, self-contained Hive batch. Oozie workflows on EMR are heavier to maintain than a serverless scheduling pattern.
C) Glue supports ETL but is not optimized for pure Hive scripts running on S3 data; it may introduce unnecessary data catalog and job orchestration overhead for a simple Hive batch.
D) Lambda layers for Hive runtime in Lambda is not a typical or reliable deployment model for Hive workloads, and Step Functions adds orchestration cost and complexity without clear cost benefits for a once-daily batch.



A company wants to improve the data load time of a sales data dashboard. Data has been collected as .csv les and stored within an Amazon S3 bucket that is partitioned by date. The data is then loaded to an Amazon Redshift data warehouse for frequent analysis. The data volume is up to 500 GB per day.
Which solution will improve the data loading performance?

  1. Compress .csv les and use an INSERT statement to ingest data into Amazon Redshift.
  2. Split large .csv les, then use a COPY command to load data into Amazon Redshift.
  3. Use Amazon Kinesis Data Firehose to ingest data into Amazon Redshift.
  4. Load the .csv les in an unsorted key order and vacuum the table in Amazon Redshift.

Answer(s): B

Explanation:

A brief strengthens load performance: B is correct because a COPY command parallelizes ingestion from S3, efficiently loading split CSV files into Redshift with proper formatting and optional compression, dramatically improving throughput for large daily volumes. A is incorrect because INSERTs are row-by-row and far slower for bulk loads. C is incorrect because Kinesis Data Firehose is designed for streaming data ingestion, not optimized bulk loads from S3 into Redshift. D is incorrect because loading unsorted data and VACUUMing post-load adds unnecessary overhead; proper data organization and automatic sort keys during COPY yield better performance.


Reference:

https://aws.amazon.com/blogs/big-data/using-amazon-redshift-spectrum-amazon-athena-and-aws-glue-with-node-js-in-production/



A company has a data warehouse in Amazon Redshift that is approximately 500 TB in size. New data is imported every few hours and read-only queries are run throughout the day and evening. There is a particularly heavy load with no writes for several hours each morning on business days. During those hours, some queries are queued and take a long time to execute. The company needs to optimize query execution and avoid any downtime.
What is the MOST cost-effective solution?

  1. Enable concurrency scaling in the workload management (WLM) queue.
  2. Add more nodes using the AWS Management Console during peak hours. Set the distribution style to ALL.
  3. Use elastic resize to quickly add nodes during peak times. Remove the nodes when they are not needed.
  4. Use a snapshot, restore, and resize operation. Switch to the new target cluster.

Answer(s): A

Explanation:

Concurrency scaling in Redshift automatically adds transient capacity to handle bursts of read queries without downtime, reducing queue wait times while only incurring charges for usage during those bursts, making it cost-effective for sporadic heavy workloads.
A) Correct. Concurrency scaling spins up additional cluster capacity to serve high-demand read queries with no downtime and pay-per-use pricing.
B) Incorrect. Scaling by adding nodes with ALL distribution is expensive and can cause data movement; not ideal for intermittent bursts and increases baseline cost.
C) Incorrect. Elastic resize adds permanent capacity and involves data redistribution; not ideal for short-lived peak periods and can cause downtime or performance variability.
D) Incorrect. Snapshot/restore/resize is disruptive and not suitable for rapid, transient workload bursts; also more complex and slower.



Viewing Page 4 of 22



Share your comments for Amazon DAS-C01 exam with other users:

Satish 11/6/2023 4:27:00 AM

it is very useful, thank you
Anonymous


Chinna 7/30/2023 8:37:00 AM

need safe rte dumps
FRANCE


1234 6/30/2023 3:40:00 AM

can you upload the cis - cpg dumps
Anonymous


Did 1/12/2024 3:01:00 AM

q6 = 1. download odt application 2. create a configuration file (xml) 3. setup.exe /download to download the installation files 4. setup.exe /configure to deploy the application
FRANCE


John 10/12/2023 12:30:00 PM

great material
Anonymous


Dinesh 8/1/2023 2:26:00 PM

could you please upload sap c_arsor_2302 questions? it will be very much helpful.
Anonymous


LBert 6/19/2023 10:23:00 AM

vraag 20c: rsa veilig voor symmtrische cryptografie? antwoord c is toch fout. rsa is voor asymmetrische cryptogafie??
NETHERLANDS


g 12/22/2023 1:51:00 PM

so far good
UNITED STATES


Milos 8/4/2023 9:33:00 AM

question 31 has obviously wrong answers. tls and ssl are used to encrypt data at transit, not at rest.
Serbia And Montenegro


Diksha 9/25/2023 2:32:00 AM

pls provide dump for 1z0-1080-23 planning exams
Anonymous


H 7/17/2023 4:28:00 AM

could you please upload the exam?
Anonymous


Anonymous 9/14/2023 4:47:00 AM

please upload this
UNITED STATES


Naveena 1/13/2024 9:55:00 AM

good material
Anonymous


WildWilly 1/19/2024 10:43:00 AM

lets see if this is good stuff...
Anonymous


Lavanya 11/2/2023 1:53:00 AM

useful information
UNITED STATES


Moussa 12/12/2023 5:52:00 AM

intéressant
BURKINA FASO


Madan 6/22/2023 9:22:00 AM

thank you for making the interactive questions
Anonymous


Vavz 11/2/2023 6:51:00 AM

questions are accurate
Anonymous


Su 11/23/2023 4:34:00 AM

i need questions/dumps for this exam.
Anonymous


LuvSN 7/16/2023 11:19:00 AM

i need this exam, when will it be uploaded
ROMANIA


Mihai 7/19/2023 12:03:00 PM

i need the dumps !
Anonymous


Wafa 11/13/2023 3:06:00 AM

very helpful
Anonymous


Alokit 7/3/2023 2:13:00 PM

good source
Anonymous


Show-Stopper 7/27/2022 11:19:00 PM

my 3rd test and passed on first try. hats off to this brain dumps site.
UNITED STATES


Michelle 6/23/2023 4:06:00 AM

please upload it
Anonymous


Lele 11/20/2023 11:55:00 AM

does anybody know if are these real exam questions?
EUROPEAN UNION


Girish Jain 10/9/2023 12:01:00 PM

are these questions similar to actual questions in the exam? because they seem to be too easy
Anonymous


Phil 12/8/2022 11:16:00 PM

i have a lot of experience but what comes in the exam is totally different from the practical day to day tasks. so i thought i would rather rely on these brain dumps rather failing the exam.
GERMANY


BV 6/8/2023 4:35:00 AM

good questions
NETHERLANDS


krishna 12/19/2023 2:05:00 AM

valied exam dumps. they were very helpful and i got a pretty good score. i am very grateful for this service and exam questions
Anonymous


Pie 9/3/2023 4:56:00 AM

will it help?
INDIA


Lucio 10/6/2023 1:45:00 PM

very useful to verify knowledge before exam
POLAND


Ajay 5/17/2023 4:54:00 AM

good stuffs
Anonymous


TestPD1 8/10/2023 12:19:00 PM

question 17 : responses arent b and c ?
EUROPEAN UNION