🔥 40% Off Crucial Exams Memberships — Deal ends today!

27 minutes, 23 seconds remaining!

GCP Professional Data Engineer Practice Question

Your analytics team operates a Dataplex lake that ingests raw logs into Cloud Storage, processes them with Dataflow streaming jobs, and writes curated tables to BigQuery. They must:

  1. continuously evaluate data quality on each zone asset,
  2. trace lineage from raw objects to curated tables, and
  3. receive automatic email alerts when any Dataflow job throughput falls below a threshold. Which combination of Google Cloud services best satisfies all three monitoring needs?
  • Catalog assets with Data Catalog, detect anomalies using Vertex AI Model Monitoring, and use Cloud Monitoring only for email alerts.

  • Instrument pipelines with Stackdriver Trace, collect metrics with the Ops Agent, and trigger alerts by Cloud Scheduler cron jobs.

  • Use Dataplex data quality rules and built-in lineage tracking for requirements 1-2, and configure Cloud Monitoring alert policies on Dataflow metrics with an email notification channel for requirement 3.

  • Query BigQuery INFORMATION_SCHEMA for quality checks, publish lineage metadata to Cloud Logging, and export logs to Pub/Sub to trigger email notifications.

GCP Professional Data Engineer
Storing the data
Your Score:
Settings & Objectives
Random Mixed
Questions are selected randomly from all chosen topics, with a preference for those you haven’t seen before. You may see several questions from the same objective or domain in a row.
Rotate by Objective
Questions cycle through each objective or domain in turn, helping you avoid long streaks of questions from the same area. You may see some repeat questions, but the distribution will be more balanced across topics.

Check or uncheck an objective to set which questions you will receive.

Bash, the Crucial Exams Chat Bot
AI Bot