GCP Professional Data Engineer Practice Question

A healthcare analytics provider stores daily CSV exports of patient monitoring metrics in a Google Cloud Storage bucket. New files are added around 02:00 UTC and are never modified after creation. Business analysts query the data in BigQuery and require the previous day's data to be available by 06:00 UTC. Operations wants a single managed service that can schedule a daily load, detect only newly arrived objects, and avoid the need for custom code or customer-managed VMs. Which design meets these requirements?

  • Configure BigQuery Data Transfer Service with the Cloud Storage connector to run daily at 03:00 UTC and load the files from the bucket prefix into a date-partitioned BigQuery table.

  • Build a Dataflow batch pipeline orchestrated by Cloud Composer that processes new objects from the bucket and writes them to BigQuery.

  • Create a Cloud Scheduler job that invokes a Cloud Function each morning to run a bq load command for all objects in the bucket.

  • Ship the bucket's contents each night on a Transfer Appliance and load the files into BigQuery after arrival.

GCP Professional Data Engineer
Ingesting and processing the data
Your Score:
Settings & Objectives
Random Mixed
Questions are selected randomly from all chosen topics, with a preference for those you haven’t seen before. You may see several questions from the same objective or domain in a row.
Rotate by Objective
Questions cycle through each objective or domain in turn, helping you avoid long streaks of questions from the same area. You may see some repeat questions, but the distribution will be more balanced across topics.

Check or uncheck an objective to set which questions you will receive.

Bash, the Crucial Exams Chat Bot
AI Bot