AWS Certified Data Engineer Associate DEA-C01 Practice Question

Your company ingests daily CSV files into an S3 data lake, runs an AWS Glue Spark job to denormalize the data, and then loads the result into an Amazon Redshift table that has a primary key on order_id. Duplicate order_id values occasionally appear in the source data and cause the Redshift load to fail. You must add an automated step to the existing Glue workflow that verifies the order_id column contains only unique values and stops the workflow if the rule is violated. Which approach satisfies these requirements with minimal custom code?

  • Use AWS Database Migration Service with a full-load task followed by validation to compare S3 data with Redshift and detect any duplicate records before loading.

  • Enable versioning on the S3 bucket so duplicate files are stored as separate object versions, ensuring the downstream load receives only the latest data.

  • Create an AWS Glue Data Quality ruleset that uses an IsUnique rule on the order_id column and configure the evaluation action to fail the workflow when the rule fails.

  • Turn on AWS Glue job bookmarks so previously processed rows are skipped and duplicates are automatically removed during the next job run.

AWS Certified Data Engineer Associate DEA-C01
Data Operations and Support
Your Score:
Settings & Objectives
Random Mixed
Questions are selected randomly from all chosen topics, with a preference for those you haven’t seen before. You may see several questions from the same objective or domain in a row.
Rotate by Objective
Questions cycle through each objective or domain in turn, helping you avoid long streaks of questions from the same area. You may see some repeat questions, but the distribution will be more balanced across topics.

Check or uncheck an objective to set which questions you will receive.

Bash, the Crucial Exams Chat Bot
AI Bot