AWS Certified Data Engineer Associate DEA-C01 Practice Question

A company ingests 50 000 JSON events per second from IoT sensors into an Amazon Kinesis Data Stream. The analytics team needs each record converted to Apache Parquet with sub-second latency and written to Amazon S3. The solution must scale automatically with the unpredictable event rate and require minimal infrastructure management. Which approach meets these requirements most effectively?

  • Configure an Amazon EMR cluster with Spark Structured Streaming to poll the stream and convert data to Parquet in Amazon S3.

  • Use AWS Lambda with Kinesis Data Streams as the event source; each invocation converts the JSON record to Parquet and writes it to Amazon S3.

  • Deliver the stream to Amazon S3 through Kinesis Data Firehose with a Lambda transformation that converts incoming records to Parquet format.

  • Create an AWS Glue streaming ETL job that reads from the Kinesis Data Stream and writes Parquet files to Amazon S3.

AWS Certified Data Engineer Associate DEA-C01
Data Ingestion and Transformation
Your Score:
Settings & Objectives
Random Mixed
Questions are selected randomly from all chosen topics, with a preference for those you haven’t seen before. You may see several questions from the same objective or domain in a row.
Rotate by Objective
Questions cycle through each objective or domain in turn, helping you avoid long streaks of questions from the same area. You may see some repeat questions, but the distribution will be more balanced across topics.

Check or uncheck an objective to set which questions you will receive.

Bash, the Crucial Exams Chat Bot
AI Bot