🔥 40% Off Crucial Exams Memberships — Deal ends today!

1 hour, 52 minutes remaining!

GCP Professional Data Engineer Practice Question

Your media company ingests roughly 5 TB of clickstream events into BigQuery each day. Analysts frequently aggregate metrics filtered by event_date and country and occasionally need to drill into a nested JSON payload per event. The user profile dimension changes less than 1 % daily, and the team wants to avoid the latency of runtime joins while keeping query costs low. Which BigQuery table design best meets these requirements?

  • Create a single BigQuery table partitioned on event_date, clustered on country, and store the user profile and event payload as nested STRUCT and ARRAY fields in each row.

  • Store events in individual daily sharded tables that contain only primitive columns; analysts combine shards using wildcard queries.

  • Implement a star schema with a partitioned fact_events table and separate dimension tables for users and countries, relying on joins at query time.

  • Build a snowflake schema with fully normalized dimensions and generate daily flattened materialized views to satisfy analyst queries.

GCP Professional Data Engineer
Storing the data
Your Score:
Settings & Objectives
Random Mixed
Questions are selected randomly from all chosen topics, with a preference for those you haven’t seen before. You may see several questions from the same objective or domain in a row.
Rotate by Objective
Questions cycle through each objective or domain in turn, helping you avoid long streaks of questions from the same area. You may see some repeat questions, but the distribution will be more balanced across topics.

Check or uncheck an objective to set which questions you will receive.

Bash, the Crucial Exams Chat Bot
AI Bot