GCP Professional Data Engineer Practice Question

Your genomics research institute must move 3 PB of raw sequencing files from an on-premises HPC cluster to a new data lake in a regional Cloud Storage bucket. The data center's WAN link averages 150 Mbps and cannot be upgraded in time. Compliance prohibits transferring this data across the public internet, yet the migration must finish within one month so downstream analytics in BigQuery can begin on schedule. Which Google solution best meets these constraints while requiring the least change to existing infrastructure?

  • Configure BigQuery Data Transfer Service to pull the sequencing files directly into BigQuery as external data sources.

  • Provision Dedicated Interconnect and use gsutil -m rsync to stream the data from the cluster to Cloud Storage.

  • Install Storage Transfer Service agents on the HPC cluster and schedule continuous transfers to Cloud Storage over the existing link.

  • Order a Google Transfer Appliance, copy the datasets locally, then return it for bulk ingestion into Cloud Storage.

GCP Professional Data Engineer
Designing data processing systems
Your Score:
Settings & Objectives
Random Mixed
Questions are selected randomly from all chosen topics, with a preference for those you haven’t seen before. You may see several questions from the same objective or domain in a row.
Rotate by Objective
Questions cycle through each objective or domain in turn, helping you avoid long streaks of questions from the same area. You may see some repeat questions, but the distribution will be more balanced across topics.

Check or uncheck an objective to set which questions you will receive.

Bash, the Crucial Exams Chat Bot
AI Bot