AWS Certified Data Engineer Associate DEA-C01 Practice Question

A company ingests daily JSON files into the s3://sales/raw/ prefix. An AWS Glue Spark job converts the files to Parquet and loads the results into an Amazon Redshift table. The job now takes hours because it reprocesses two years of files every night. You must limit processing to only files added since the last run without changing code. What should you do?

  • Reconfigure the job as an AWS Glue streaming job that reads from a Kinesis Data Stream.

  • Add an S3 event notification that invokes an AWS Lambda function to call StartJobRun for each new object key.

  • Enable AWS Glue job bookmarks for the existing job and keep the default run schedule.

  • Create time-based folders in Amazon S3, define an Athena external table with partition projection, and query it from Redshift Spectrum instead of using ETL.

AWS Certified Data Engineer Associate DEA-C01
Data Operations and Support
Your Score:
Settings & Objectives
Random Mixed
Questions are selected randomly from all chosen topics, with a preference for those you haven’t seen before. You may see several questions from the same objective or domain in a row.
Rotate by Objective
Questions cycle through each objective or domain in turn, helping you avoid long streaks of questions from the same area. You may see some repeat questions, but the distribution will be more balanced across topics.

Check or uncheck an objective to set which questions you will receive.

Bash, the Crucial Exams Chat Bot
AI Bot