I did a webinar with Confluent‘s partner Expero about “Apache Kafka and Machine Learning for Real Time Supply Chain Optimization“. This is a great example for anybody in automation industry / Industrial IoT (IIoT) like automotive, manufacturing, logistics, etc.
We explain how a real time event streaming platform can integrate in real time with the legacy world and proprietary IIoT protocols (like Siemens S7, Modbus, Beckhoff ADS, OPC-UA, et al). You can process the data at scale and then ingest it into a modern database (like AWS S3, Snowflake or MongoDB) or analytic / machine learning framework (like TensorFlow, PyTorch or Azure Machine Learning Service).
Here is the architecture we use to discuss and implement the supply chain optimization use case leveraging real time stream processing and machine learning:
We leverage various components from the Apache Kafka ecosystem. This includes:
Automating multifaceted, complex workflows requires hybrid solutions including streaming analytics of IOT data and batch analytics. This includes machine learning solutions and real time visualization. Leaders in organizations who are responsible for global supply chain planning are responsible for working with and integrating with data from disparate sources around the world. Many of these data sources output information in real time. This assists planners in operationalizing plans and interacting with manufacturing output. IOT sensors on manufacturing equipment and inventory control systems feed real time processing pipelines to match actuals productions figures against planned schedules to calculate yield efficiency.
Using information from both real time systems and batch optimization, supply chain managers are able to economize operations and automate tedious inventory and manufacturing accounting processes. Sitting on top of all of these systems is a supply chain visualization tool. This enables users’ visibility over the global supply chain. If you are responsible for key data integration initiatives, join for a detailed walk through of a customer’s use of this system built using Confluent and Expero tools.
What will you learn?
Here is the slide deck and video recording. Have fun watching it. Please let me know if you have any feedback or questions:
You are currently viewing a placeholder content from Default. To access the actual content, click the button below. Please note that doing so will share data with third-party providers.
Complex Event Processing is the most underused capability in Apache Flink. It detects meaningful event…
MCP, REST/HTTP APIs, and Apache Kafka are not alternatives. They solve different problems at different…
The Enterprise Agentic AI Landscape 2026 maps every major AI vendor across two dimensions that…
Agentic AI without governed processes is fast but ungoverned. Event-driven integration without process intelligence moves…
Two toolchains, two skill sets, two CI/CD pipelines — that has been the reality for…
The Shift Left Architecture moves data integration logic into an event-driven architecture where governed data…