Medium1 markMultiple Choice
This question is part of a case study — click to read the full scenario(Case 16)

CASE STUDY: AutoIoT

Overview: Connected car manufacturer. 1M vehicles sending telemetry every 5 seconds.
Business: Predictive maintenance alerts, real-time fleet tracking, monetize anonymized data.
Executives:

  • CEO: "Leverage AI to predict failures."
  • CTO: "Current MQTT brokers crashing. Need fully managed, scalable ingestion."
  • DPO: "Vehicle location is sensitive. Strip PII before analytics."
    Tech: Ingest millions of msgs/sec, real-time stream processing for anomalies, store raw data for ML, sub-second queries for dashboards.
    Constraints: Vehicles lose connection and send late batch data. ML models updated weekly. Strict analytics budget.

Which architecture should you design for the data ingestion and processing layer to replace the crashing MQTT brokers?

GCP PCA · Question 17 · Domain 2: Managing and Provisioning a Solution Infrastructure

CASE STUDY: AutoIoT

Overview: Connected car manufacturer. 1M vehicles sending telemetry every 5 seconds.
Business: Predictive maintenance alerts, real-time fleet tracking, monetize anonymized data.
Executives:

  • CEO: "Leverage AI to predict failures."
  • CTO: "Current MQTT brokers crashing. Need fully managed, scalable ingestion."
  • DPO: "Vehicle location is sensitive. Strip PII before analytics."
    Tech: Ingest millions of msgs/sec, real-time stream processing for anomalies, store raw data for ML, sub-second queries for dashboards.
    Constraints: Vehicles lose connection and send late batch data. ML models updated weekly. Strict analytics budget.

How should you handle the constraint where vehicles lose connection and send late batch data?

Answer options:

A.

Use processing-time windowing in Dataflow to ensure all data is processed immediately upon arrival.

B.

Use Cloud Dataflow with event-time windowing and configure allowed lateness to process delayed messages.

C.

Drop the late data at the Pub/Sub layer to maintain real-time dashboard performance.

D.

Store all data in Cloud Storage first, then run a daily batch job to sort the timestamps.

How to approach this question

Understand the difference between event-time (when it happened) and processing-time (when the server saw it). Late IoT data requires event-time processing.

Full Answer

B.Use Cloud Dataflow with event-time windowing and configure allowed lateness to process delayed messages.✓ Correct
Use Cloud Dataflow with event-time windowing and configure allowed lateness to process delayed messages.
In stream processing, handling late data is a core capability of Cloud Dataflow (Apache Beam). By using 'event-time' (the timestamp generated by the IoT device) rather than 'processing-time' (when Dataflow receives it), and configuring 'allowed lateness', Dataflow can correctly update aggregations and windows even when vehicles reconnect and upload batch data late.

Common mistakes

Choosing processing-time (A). This is a classic stream processing error. Processing time ignores when the event actually occurred.

Practice the full GCP Professional Cloud Architect Practice Exam 4

50 questions · hints · full answers · grading

More questions from this exam