GCP Professional Data Engineer Practice Question

Your team schedules an hourly Cloud Composer DAG that triggers a Dataflow batch job to aggregate the previous hour's click-stream data and then archives the raw files. After an upcoming four-hour maintenance pause you must ensure that Composer does not automatically create four delayed DAG runs, yet any individual task in the next scheduled run must retry up to three times at 10-minute intervals if it fails. With minimal custom code, how should you configure the DAG?

  • Use schedule_interval='@hourly', keep catchup=True, and add depends_on_past=True with the same retry settings.

  • Leave catchup enabled, set execution_timeout=600 seconds, and disable retries to avoid duplicate processing.

  • Set schedule_interval='@hourly', configure catchup=False, and add retries: 3 and retry_delay: timedelta(minutes=10) in default_args.

  • Disable catchup but set max_active_runs=24, and require any failed run to be relaunched manually.

GCP Professional Data Engineer
Maintaining and automating data workloads
Your Score:
Settings & Objectives
Random Mixed
Questions are selected randomly from all chosen topics, with a preference for those you haven’t seen before. You may see several questions from the same objective or domain in a row.
Rotate by Objective
Questions cycle through each objective or domain in turn, helping you avoid long streaks of questions from the same area. You may see some repeat questions, but the distribution will be more balanced across topics.

Check or uncheck an objective to set which questions you will receive.

Bash, the Crucial Exams Chat Bot
AI Bot