AWS Certified Data Engineer Associate DEA-C01 Practice Question

A retail company runs an AWS Glue Spark job that reads daily order files from an S3 bucket, transforms the data, and loads it into an Amazon Redshift fact table. The business requires that the order_id column be unique each day; if any duplicate order_id values are found, the load must abort and an alert must be sent. Which approach most efficiently enforces this rule inside the existing Glue job while minimizing additional services?

  • Attach an AWS Lake Formation row-level access filter to drop duplicate order_id rows during the load.

  • Add an AWS Glue Data Quality transform to the job, create a ruleset that asserts order_id is unique, and configure the job to fail if the rule is violated.

  • Define a UNIQUE constraint on order_id in the Redshift table and let the COPY command reject duplicate rows after the data load.

  • Enable Amazon S3 Object Lock in compliance mode on the raw bucket so duplicate objects cannot be written.

AWS Certified Data Engineer Associate DEA-C01
Data Operations and Support
Your Score:
Settings & Objectives
Random Mixed
Questions are selected randomly from all chosen topics, with a preference for those you haven’t seen before. You may see several questions from the same objective or domain in a row.
Rotate by Objective
Questions cycle through each objective or domain in turn, helping you avoid long streaks of questions from the same area. You may see some repeat questions, but the distribution will be more balanced across topics.

Check or uncheck an objective to set which questions you will receive.

Bash, the Crucial Exams Chat Bot
AI Bot