GCP Professional Data Engineer Practice Question

A retail analytics team deploys the following daily Cloud Composer workflow:

  • LocalFilesystemToGCSOperator uploads the previous day's CSV files
  • DataflowTemplateOperator transforms the files
  • GCSToBigQueryOperator loads the results into a partitioned table

After the DAG is first published, Airflow immediately tries to schedule every daily interval back to the start_date, quickly exceeding the project's Dataflow quota. The team wants the DAG to run only for intervals that begin after the deployment moment, while keeping the existing schedule_interval and start_date. What modification to the DAG definition will accomplish this goal?

  • Change schedule_interval to None to prevent historical and future automatic scheduling.

  • Set depends_on_past=True so each run waits for the previous one to finish before starting.

  • Configure max_active_runs=1 to limit the DAG to a single concurrent run.

  • Add catchup=False to the DAG constructor to disable back-filling of missed intervals.

GCP Professional Data Engineer
Maintaining and automating data workloads
Your Score:
Settings & Objectives
Random Mixed
Questions are selected randomly from all chosen topics, with a preference for those you haven’t seen before. You may see several questions from the same objective or domain in a row.
Rotate by Objective
Questions cycle through each objective or domain in turn, helping you avoid long streaks of questions from the same area. You may see some repeat questions, but the distribution will be more balanced across topics.

Check or uncheck an objective to set which questions you will receive.

Bash, the Crucial Exams Chat Bot
AI Bot