AWS Certified Data Engineer Associate DEA-C01 Practice Question

An e-commerce company ingests 20 GB of click-stream logs every 15 minutes into AWS. An AWS Glue Spark job will convert the logs from JSON to partitioned Parquet before loading them into Amazon Redshift. The solution must minimize cost, decouple ingestion from transformation jobs, and scale automatically with data volume without manual capacity management. Where should the logs be placed as the intermediate staging location?

  • Persist the logs on an Amazon EFS file system mounted on an ingestion EC2 instance.

  • Store the raw logs in an Amazon S3 bucket that serves as the landing zone.

  • Load the logs directly into Amazon Redshift staging tables in a separate schema.

  • Write the logs into an Amazon DynamoDB table using binary (BLOB) attributes.

AWS Certified Data Engineer Associate DEA-C01
Data Ingestion and Transformation
Your Score:
Settings & Objectives
Random Mixed
Questions are selected randomly from all chosen topics, with a preference for those you haven’t seen before. You may see several questions from the same objective or domain in a row.
Rotate by Objective
Questions cycle through each objective or domain in turn, helping you avoid long streaks of questions from the same area. You may see some repeat questions, but the distribution will be more balanced across topics.

Check or uncheck an objective to set which questions you will receive.

Bash, the Crucial Exams Chat Bot
AI Bot