GCP Professional Data Engineer Practice Question

A multinational retailer wants to modernize its on-premises analytics stack by moving to Google Cloud. Requirements are:

  • Land raw click-stream and IoT data (JSON, images) with virtually unlimited scale.
  • Provide sub-minute dashboards fed by a streaming pipeline.
  • Run monthly company-wide SQL analytics without managing infrastructure.
  • Enforce centralized security, data quality, and lineage controls, while letting regional business units own their datasets. Which high-level design best satisfies these goals with minimal operational overhead?
  • Land raw data in Cloud Storage, register it in Dataplex raw zones, process streams with Dataflow into curated BigQuery tables that are also governed by Dataplex.

  • Stream JSON payloads straight into a single BigQuery dataset, store images as BASE64 strings, and manage access with manual dataset-level ACLs across regions.

  • Ingest all data directly into a global Spanner database to serve both real-time dashboards and analytical SQL queries, enforcing governance through IAM on Spanner tables.

  • Write raw events into Bigtable, schedule Cloud Composer DAGs to copy data into BigQuery, and use Data Catalog in each project for discovery and policy control.

GCP Professional Data Engineer
Storing the data
Your Score:
Settings & Objectives
Random Mixed
Questions are selected randomly from all chosen topics, with a preference for those you haven’t seen before. You may see several questions from the same objective or domain in a row.
Rotate by Objective
Questions cycle through each objective or domain in turn, helping you avoid long streaks of questions from the same area. You may see some repeat questions, but the distribution will be more balanced across topics.

Check or uncheck an objective to set which questions you will receive.

Bash, the Crucial Exams Chat Bot
AI Bot