Manufacturing data pipelines from PLC to lakehouse — OPC UA ingestion, Historian integration, MES and ERP CDC, and the curated layer that powers analytics, AI, and the models that actually go to production. Built by engineers who've been on call for a 3am ingestion failure.
A typical SaaS data pipeline ingests JSON from APIs into a warehouse. A manufacturing data pipeline ingests time-series from PLCs through OPC UA gateways into a Historian, then exfiltrates that to the cloud, while also pulling transactional events from MES, master data from ERP, and unstructured quality records from the QMS — all with timestamps that need to align to within seconds and tag names that change when the controls engineer reflashes a PLC. The volume is high (hundreds of millions of tag readings per day per plant), the latency requirements vary by use case, and a broken pipeline at 3am means tomorrow's shift report is wrong.
Manufacturing data engineering done right uses the medallion pattern (bronze raw, silver curated, gold business-ready) on a lakehouse — but with extras: tag-name mapping tables that track changes over time, time-window partitioning calibrated to the analytical use case (not just calendar partitioning), schema-evolution handling for the day a new line gets added, and data quality checks that flag drift before it corrupts downstream models. Pipelines built this way survive PLC firmware updates, MES upgrades, and the inevitable controls engineering changes.
OPC UA ingestion from PLCs and SCADA into an edge gateway, edge to cloud egress with store-and-forward for connectivity drops, then into a Bronze lakehouse table partitioned by tag and time window. The pattern that scales from one line to one hundred plants without re-architecting.
CDC from your MES (Wonderware, Rockwell, Siemens) and ERP (SAP, Oracle, D365) into the lakehouse with proper handling of late-arriving data, soft deletes, and the cross-system joins that downstream analytics need. Daily reconciliation against the source of record.
Silver and gold tables organized by manufacturing concept — production events, quality events, maintenance events, master data — with consistent schemas, documented lineage, and the data contracts that keep downstream consumers from breaking when source schemas change.
Manufacturing data engineering delivered for production reliability: medallion lakehouse architecture, OPC UA ingestion to Bronze, CDC from MES and ERP, curated Silver/Gold layers with data contracts, monitoring and alerting at every stage, runbooks for the on-call engineer, and the lineage documentation that lets your team trace any anomaly back to source within minutes.
The full Data Engineering Consulting practice across industries.
All manufacturing technology services from Xylity.
Industry-specific consulting across the verticals we serve.
Discrete manufacturing plant with 50 lines and a Historian: typically 50-200 GB of raw tag data per day, plus a few GB of MES events and ERP transactions. Process plants with deeper instrumentation can hit 1-2 TB/day. Storage is rarely the bottleneck on modern lakehouses; query patterns and partitioning are what determines whether the data is usable.
Almost always keep the Historian. It's purpose-built for sub-second time-series and the controls engineering team knows it. The lakehouse sits downstream as the analytical layer — pulling from the Historian via scheduled or change-driven extraction. We've never recommended replacing a working Historian with a lakehouse.
Yes. Pre-qualified data engineers with manufacturing domain experience — OPC UA, Historian integration, lakehouse patterns (Databricks, Fabric, Snowflake), MES/ERP CDC, and the on-call discipline to keep production pipelines running. 92% first-match acceptance.
OPC UA, lakehouse medallion, data contracts, and the operational discipline that keeps shift reports running through PLC updates.