I am happy that my first official Confluent blog post was published and want to link to it from by blog:
How to Build and Deploy Scalable Machine Learning in Production with Apache Kafka
The post explains in detail how you can leverage Apache Kafka and its Streams API to deploy analytic models to a lightweight, but scalable, mission-critical streaming appilcation.
If you want to take a look directly at the source code, go to my Github project about Kafka + Machine Learning. It contains several examples how to combine Kafka Streams with frameworks like TensorFlow, H2O or DeepLearning4J.
The mainframe is evolving—not fading. With cloud-native features, AI acceleration, and quantum-safe encryption, platforms like…
OpenAI revealed how it builds and scales the real-time data streaming infrastructure that powers its…
Durable execution engines like Temporal and Restate are redefining how developers orchestrate long-running, stateful workflows…
Real-time visibility has become essential in logistics. As supply chains grow more complex, providers must…
SAP Sapphire 2025 in Madrid brought together global SAP users, partners, and technology leaders to…
Agentic AI is emerging as a powerful pattern for building autonomous, intelligent, and collaborative systems.…