Business Impact

  • 15s

    Near real-time inference of time series telemetry data

  • 5M

    Autoscaled data ingestion (in records per second)

Customer Key Facts

  • Location : North America
  • Industry : Oil & Energy

Problem Context

The customer is an American multinational energy corporation engaged in oil, natural gas, and geothermal energy and produces approximately 160 thousand barrels of oil per day across ten oil rigs, also known as facilities. Approximately 7 to 8 percent of facility downtime in their oil rigs could be attributed to malfunctioning compressors, leading to significant increases in resource and maintenance costs. They wanted to minimize facility downtime by using predictive tools to understand causes of such failures.



  • Incredibly large volumes of input data
  • High throughput required in streaming data with milliseconds of lag
  • Customizing third-party application OpenTSDB to increase performance
Cloud Storage
Cloud Pub/Sub
Google's BigQuery
Cloud ML
Google Cloud Dataflow
Cloud Bigtable

Data Pipeline and Data Science Workflow to Streamline Processes, Make Predictions, and Visualize Telemetry Data


Quantiphi built an end-to-end data pipeline from the client’s on-prem to Google Cloud Platform (GCP) that routes high throughput telemetry data to GCP APIs, data stores, and visualization platforms. In addition to building machine learning training and serving pipeline to identify patterns in the ingested data and inference, the data science workflow facilitates predictive maintenance.


  • Near real-time ingestion of telemetry data
  • Minimized facility downtime
  • Increased cost savings
  • Improved safety of operators

Looking for similar project?

Let's Talk

Get your digital transformation started

Let's Talk