GCP Professional Data Engineer Practice Question

Your analytics team uses Dataform in BigQuery to orchestrate an ELT pipeline. Raw daily POS files land in a staging table, then several downstream tables are published to production datasets. You must ensure the run is aborted-and nothing is published-if any row in the staging table has a sale_date older than 90 days. The control must live in the same Dataform repo, without extra services or custom orchestration. Which approach meets these requirements?

  • Add a SQLX file of type "assertion" that selects rows where sale_date < DATE_SUB(CURRENT_DATE(), INTERVAL 90 DAY); Dataform automatically fails the run if the query returns rows.

  • Embed a BEGIN … RAISE ERROR script in the pre_operation block of every downstream table so the build stops when invalid dates are detected.

  • Create a Cloud Monitoring alert on the staging table's age metric and configure Cloud Build to cancel the Dataform run when the alert fires.

  • Set a 90-day partition expiration on the staging table so older records are automatically deleted before downstream tables are built.

GCP Professional Data Engineer
Designing data processing systems
Your Score:
Settings & Objectives
Random Mixed
Questions are selected randomly from all chosen topics, with a preference for those you haven’t seen before. You may see several questions from the same objective or domain in a row.
Rotate by Objective
Questions cycle through each objective or domain in turn, helping you avoid long streaks of questions from the same area. You may see some repeat questions, but the distribution will be more balanced across topics.

Check or uncheck an objective to set which questions you will receive.

Bash, the Crucial Exams Chat Bot
AI Bot