GCP Professional Data Engineer Practice Question

Your team is designing the first landing layer for a new pipeline that ingests about 10 TB per day of compressed JSON and image files through Pub/Sub and Dataflow. Compliance requires that the raw, immutable data be retained for at least seven years at the lowest possible cost. The data will later be transformed in Dataproc and occasionally queried in BigQuery, so it must remain accessible to multiple analytics engines and be able to transition automatically to colder, cheaper tiers as it ages. Which Google Cloud sink best satisfies these requirements?

  • A Cloud Bigtable instance with a wide-row schema storing each file as a cell value

  • A BigQuery dataset that relies on BigQuery long-term storage pricing after 90 days

  • A Cloud Storage bucket configured with lifecycle rules to serve as the organization's data lake

  • A regional Cloud Spanner database using a BLOB column for each ingested object

GCP Professional Data Engineer
Ingesting and processing the data
Your Score:
Settings & Objectives
Random Mixed
Questions are selected randomly from all chosen topics, with a preference for those you haven’t seen before. You may see several questions from the same objective or domain in a row.
Rotate by Objective
Questions cycle through each objective or domain in turn, helping you avoid long streaks of questions from the same area. You may see some repeat questions, but the distribution will be more balanced across topics.

Check or uncheck an objective to set which questions you will receive.

Bash, the Crucial Exams Chat Bot
AI Bot