Your retail company runs a nightly Dataflow batch job that loads 3 TB of CSV files from a Cloud Storage bucket in europe-west1 into BigQuery. Corporate policy mandates that if the primary region becomes unavailable the pipeline must be restarted in another region within 30 minutes, without recompiling or rebuilding the code. Which design best satisfies this disaster-recovery objective while aligning with Google-recommended practices?
Store the pipeline as a Dataflow Flex Template in a multi-regional Cloud Storage bucket, monitor the job with Cloud Monitoring, and trigger a Cloud Function to launch the same template in europe-west4 when the primary job fails.
Redesign the pipeline to use Cloud Spanner for state management; Spanner's multi-region replication will allow the existing job to keep running even if europe-west1 fails.
Start the job with the --automaticFailover flag so Dataflow transparently restarts the pipeline in the nearest healthy region during an outage.
Convert the batch job to a streaming pipeline, enable hourly Dataflow snapshots, and restore the latest snapshot in europe-west4 if europe-west1 becomes unavailable.
Packaging the pipeline as a Dataflow Flex Template stores an executable Docker image and JSON specification in Cloud Storage. Because templates are region-agnostic, the same image can be launched in any Dataflow-supported region simply by specifying a different --region parameter. By placing the template artifacts in a multi-regional bucket they remain accessible even if europe-west1 is down. Cloud Monitoring can watch the running job's state and emit an alert when it enters a Failed or Cancelled state; a Cloud Function (or Cloud Run service) triggered by that alert can call the Dataflow REST API to start a new job from the same template in europe-west4, meeting the 30-minute RTO without human compilation work.
The other options rely on features that do not exist (automatic region fail-over flag), misuse snapshots (which are limited to streaming pipelines and cannot be restored across regions), or depend on an unrelated service's replication (Cloud Spanner) rather than enabling rapid redeployment of the Dataflow job itself.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What are Dataflow Flex Templates and how do they differ from traditional templates?
Open an interactive chat with Bash
What role does Cloud Monitoring play in this disaster recovery design?
Open an interactive chat with Bash
Why is storing pipeline artifacts in a multi-regional Cloud Storage bucket important?
Open an interactive chat with Bash
What is a Dataflow Flex Template?
Open an interactive chat with Bash
How does multi-regional Cloud Storage ensure disaster recovery?
Open an interactive chat with Bash
How does Cloud Monitoring and Cloud Functions enable failover in this scenario?
Open an interactive chat with Bash
GCP Professional Data Engineer
Designing data processing systems
Your Score:
Report Issue
Bash, the Crucial Exams Chat Bot
AI Bot
Loading...
Loading...
Loading...
Pass with Confidence.
IT & Cybersecurity Package
You have hit the limits of our free tier, become a Premium Member today for unlimited access.
Military, Healthcare worker, Gov. employee or Teacher? See if you qualify for a Community Discount.
Monthly
$19.99
$19.99/mo
Billed monthly, Cancel any time.
3 Month Pass
$44.99
$14.99/mo
One time purchase of $44.99, Does not auto-renew.
MOST POPULAR
Annual Pass
$119.99
$9.99/mo
One time purchase of $119.99, Does not auto-renew.
BEST DEAL
Lifetime Pass
$189.99
One time purchase, Good for life.
What You Get
All IT & Cybersecurity Package plans include the following perks and exams .