🔥 40% Off Crucial Exams Memberships — Deal ends today!

1 hour, 26 minutes remaining!

GCP Professional Data Engineer Practice Question

Your team must design a storage layout for an e-commerce platform that streams about 50 000 clickstream events per second (≈5 TB/day). The recommendation engine needs single-digit millisecond access to the most recent seven days of events, while data scientists expect to run large ad-hoc SQL analyses on the full event history every morning. You want a fully managed solution that minimizes operational overhead and storage cost. Which approach best satisfies these requirements?

  • Persist events in Cloud SQL for PostgreSQL with logical replication into BigQuery; query Cloud SQL for the latest seven days and BigQuery for historical analysis.

  • Ingest events directly into BigQuery partitioned tables and build materialized views on the last seven days for the recommendation engine.

  • Write all events to Cloud Storage Nearline; define a BigQuery external table for analytics and use Memorystore to cache the most recent events for recommendations.

  • Stream events into Cloud Bigtable for operational access; use a Dataflow job to write each event to a partitioned BigQuery table and delete data in Bigtable after seven days.

GCP Professional Data Engineer
Storing the data
Your Score:
Settings & Objectives
Random Mixed
Questions are selected randomly from all chosen topics, with a preference for those you haven’t seen before. You may see several questions from the same objective or domain in a row.
Rotate by Objective
Questions cycle through each objective or domain in turn, helping you avoid long streaks of questions from the same area. You may see some repeat questions, but the distribution will be more balanced across topics.

Check or uncheck an objective to set which questions you will receive.

Bash, the Crucial Exams Chat Bot
AI Bot