GCP Professional Data Engineer Practice Question

A media company stores 5 TB of clickstream events in a date-partitioned, user_id and campaign_id-clustered BigQuery table. Analysts explore the data with Looker Studio dashboards that join this fact table to two small dimension tables and apply flexible filters such as "last 7 / 30 / 90 days," specific campaigns, and ad formats. Each dashboard widget currently takes about 8-10 seconds to render. You need to achieve sub-second latency without changing existing SQL or rebuilding the dashboards, while keeping incremental cost low and remaining entirely within BigQuery. Which approach best meets these requirements?

  • Build a scheduled Dataflow job that writes daily aggregated click metrics to a new partitioned table and have analysts query it with table decorators.

  • Create a materialized view that pre-aggregates daily clicks by user_id and campaign_id, then modify dashboards to query the materialized view instead of the base table.

  • Purchase a BigQuery BI Engine reservation large enough to hold the most-queried partitions of the fact table and the two dimension tables, letting Looker Studio queries be automatically accelerated in memory.

  • Export the fact table to Cloud Storage, load it into Cloud Bigtable, and connect Looker Studio through a Bigtable connector for low-latency lookups.

GCP Professional Data Engineer
Preparing and using data for analysis
Your Score:
Settings & Objectives
Random Mixed
Questions are selected randomly from all chosen topics, with a preference for those you haven’t seen before. You may see several questions from the same objective or domain in a row.
Rotate by Objective
Questions cycle through each objective or domain in turn, helping you avoid long streaks of questions from the same area. You may see some repeat questions, but the distribution will be more balanced across topics.

Check or uncheck an objective to set which questions you will receive.

Bash, the Crucial Exams Chat Bot
AI Bot