GCP Professional Data Engineer Practice Question

Your company operates a private data center that stores 400 TB of historical sensor data in an on-premises Hadoop HDFS cluster. Leadership has approved a one-time migration of this archive into a Coldline Cloud Storage bucket for long-term retention and occasional analytics in BigQuery. The data center's dedicated internet connection is limited to 200 Mbps and must remain available for business traffic, so extended online transfers are unacceptable. Which ingestion approach best meets the constraints?

  • Deploy Storage Transfer Service agents in the data center to continuously stream the HDFS data to Cloud Storage over the public internet.

  • Use Storage Transfer Service to pull the data from the data center to Cloud Storage during nightly off-peak hours over the existing 200 Mbps link.

  • Request a Google Transfer Appliance, copy the HDFS data onto the device onsite, and ship it to Google for bulk import into Cloud Storage.

  • Create a Dataproc cluster in Google Cloud, establish a VPN to the on-premises Hadoop cluster, and run a DistCp job to copy the data.

GCP Professional Data Engineer
Ingesting and processing the data
Your Score:
Settings & Objectives
Random Mixed
Questions are selected randomly from all chosen topics, with a preference for those you haven’t seen before. You may see several questions from the same objective or domain in a row.
Rotate by Objective
Questions cycle through each objective or domain in turn, helping you avoid long streaks of questions from the same area. You may see some repeat questions, but the distribution will be more balanced across topics.

Check or uncheck an objective to set which questions you will receive.

Bash, the Crucial Exams Chat Bot
AI Bot