AWS Certified Data Engineer Associate DEA-C01 Practice Question

A company needs to move about 100 GB of new sales records from its Amazon Redshift cluster to an Amazon S3 data lake every night at 01:00. The files must be stored as partitioned Parquet, and previously exported rows must not be included in subsequent runs. The solution should be fully managed, low-cost, and require as little custom code as possible. Which approach meets these requirements?

  • Configure an Amazon EventBridge rule that triggers an AWS Lambda function at 01:00; the function uses the Redshift Data API to run SELECT queries and streams the results to Amazon S3.

  • Use Redshift Spectrum to create an external table that points to an S3 location, then run a nightly CTAS command to export the sales data into Parquet files in that location.

  • Set up an AWS DMS task with Amazon Redshift as the source and Amazon S3 as the target, enable change data capture, and run the task on a nightly schedule.

  • Create an AWS Glue ETL job that reads the Redshift table through a JDBC connection, enable job bookmarks on the sales_date column, write the output as partitioned Parquet to Amazon S3, and schedule the job with an AWS Glue workflow.

AWS Certified Data Engineer Associate DEA-C01
Data Ingestion and Transformation
Your Score:
Settings & Objectives
Random Mixed
Questions are selected randomly from all chosen topics, with a preference for those you haven’t seen before. You may see several questions from the same objective or domain in a row.
Rotate by Objective
Questions cycle through each objective or domain in turn, helping you avoid long streaks of questions from the same area. You may see some repeat questions, but the distribution will be more balanced across topics.

Check or uncheck an objective to set which questions you will receive.

Bash, the Crucial Exams Chat Bot
AI Bot