Hard1 markMultiple Choice
Subtask 4.1: Technical ProcessesMachine LearningDataflowBigQueryVertex AI
This question is part of a case study — click to read the full scenario(Case 11)

CASE STUDY: TerramEarth

Company Overview: TerramEarth manufactures heavy equipment. 2 million vehicles in the field.
Current Environment: Vehicles send telemetry via cellular. Processing 100,000 msgs/sec. On-prem Hadoop cluster.
Business Requirements: Predict equipment failure. Reduce warranty costs. Provide fleet dashboard.
Executive Statements: CEO: 'Monetize data.' CFO: 'Storage costs spiraling.' CTO: 'Need scalable ingestion and ML.'
Technical Requirements: Ingest 500,000 msgs/sec. Store petabytes cost-effectively. Train ML models. Real-time anomaly detection.
Constraints: Intermittent connectivity. Strict vehicle authentication.

QUESTION:
Which architecture should you design to handle the ingestion of 500,000 messages per second from vehicles with intermittent connectivity?

GCP PCA · Question 14 · Technical Processes

CASE STUDY: TerramEarth

Company Overview: TerramEarth manufactures heavy equipment. 2 million vehicles in the field.
Current Environment: Vehicles send telemetry via cellular. Processing 100,000 msgs/sec. On-prem Hadoop cluster.
Business Requirements: Predict equipment failure. Reduce warranty costs. Provide fleet dashboard.
Executive Statements: CEO: 'Monetize data.' CFO: 'Storage costs spiraling.' CTO: 'Need scalable ingestion and ML.'
Technical Requirements: Ingest 500,000 msgs/sec. Store petabytes cost-effectively. Train ML models. Real-time anomaly detection.
Constraints: Intermittent connectivity. Strict vehicle authentication.

QUESTION:
Which THREE GCP services should you combine to build the pipeline for real-time anomaly detection and predictive maintenance ML training? (Select THREE)

Answer options:

A.

Cloud Dataflow.

B.

Cloud SQL.

C.

BigQuery.

D.

Vertex AI.

E.

Cloud Spanner.

F.

Cloud Dataprep.

How to approach this question

Identify the stream processor, the data warehouse, and the ML platform in GCP.

Full Answer

Cloud Dataflow., BigQuery., Vertex AI.
The standard GCP data-to-AI pipeline uses Cloud Dataflow for real-time stream processing (applying anomaly detection on the fly), BigQuery for storing the massive datasets, and Vertex AI for training and deploying the predictive maintenance machine learning models using the data in BigQuery.

Common mistakes

Selecting Cloud SQL or Spanner instead of BigQuery for analytics.

Practice the full GCP Professional Cloud Architect Practice Exam 1

50 questions · hints · full answers · grading

More questions from this exam