GCP Professional Data Engineer Practice Question

You deployed a streaming Dataflow pipeline in us-central1-a that ingests Pub/Sub events and writes to BigQuery. A zonal outage in us-central1-a terminated all worker VMs, and the job remained in a FAILED state until an engineer manually relaunched it, breaking the two-minute processing‐lag SLA. You must redesign the deployment so that it can survive a future zone outage without human intervention while requiring the least possible code change. What should you do?

  • Redeploy the pipeline to the regional endpoint us-central1, letting Dataflow place and restart workers in healthy zones automatically.

  • Keep the current setup but configure the Cloud Composer task to retry the Dataflow job up to three times with exponential backoff.

  • Migrate the job to a Dataproc cluster with high-availability masters spread across zones and run it with Spark Structured Streaming.

  • Enable FlexRS on the existing Dataflow job so interrupted workers are restarted on preemptible VMs in the same zone.

GCP Professional Data Engineer
Maintaining and automating data workloads
Your Score:
Settings & Objectives
Random Mixed
Questions are selected randomly from all chosen topics, with a preference for those you haven’t seen before. You may see several questions from the same objective or domain in a row.
Rotate by Objective
Questions cycle through each objective or domain in turn, helping you avoid long streaks of questions from the same area. You may see some repeat questions, but the distribution will be more balanced across topics.

Check or uncheck an objective to set which questions you will receive.

Bash, the Crucial Exams Chat Bot
AI Bot