🔥 40% Off Crucial Exams Memberships — Deal ends today!

2 hours, 31 minutes remaining!

GCP Professional Data Engineer Practice Question

Your team must automate a nightly data pipeline that 1) launches a parameterized Dataflow Flex template, 2) executes a BigQuery stored procedure on completion, and 3) calls an external Jira REST endpoint to update a ticket. The run order must change if the Dataflow job fails, and business analysts want to reuse existing Python libraries for custom retry logic. All resources will run in a private VPC with no public IPs. Which orchestration approach best meets these requirements?

  • Chain Cloud Functions with Pub/Sub topics: the first function launches the Dataflow template, a second polls Dataflow and invokes the BigQuery procedure, and a third updates Jira.

  • Schedule the Dataflow Flex template with Cloud Scheduler, embed BigQuery execution and Jira calls inside the pipeline, and rely on Dataflow's built-in retries for error handling.

  • Use Cloud Workflows triggered nightly by Cloud Scheduler; call Dataflow, BigQuery, and Jira REST endpoints with HTTP steps and handle failures with conditional states.

  • Create a private-IP Cloud Composer environment and build a Python DAG that uses built-in Dataflow and BigQuery operators plus a PythonOperator that posts to the Jira REST API.

GCP Professional Data Engineer
Ingesting and processing the data
Your Score:
Settings & Objectives
Random Mixed
Questions are selected randomly from all chosen topics, with a preference for those you haven’t seen before. You may see several questions from the same objective or domain in a row.
Rotate by Objective
Questions cycle through each objective or domain in turn, helping you avoid long streaks of questions from the same area. You may see some repeat questions, but the distribution will be more balanced across topics.

Check or uncheck an objective to set which questions you will receive.

Bash, the Crucial Exams Chat Bot
AI Bot