The Future of AI-Ready Integration with Data Streaming powered by Apache Kafka and Flink
Read More

How Siemens, SAP, and Confluent Shape the Future of AI Ready Integration – Highlights from the Rojo Event in Amsterdam

Many enterprises want to become AI ready but are limited by slow, batch based integration platforms that prevent real time insight and automation. The Rojo “Future of Integration” event in Amsterdam addressed this challenge by bringing together Siemens, SAP, Rojo, and Confluent to show how event driven and intelligent data architectures solve it. The discussions revealed how data streaming with Apache Kafka and Flink complements traditional integration tools, enabling continuous data flow, scalability, and the foundation for AI and automation. This blog summarizes the key learnings from the event, including my presentation “AI Ready Integration with Data Streaming,” and insights from Siemens, SAP, and Rojo on how enterprises can build truly connected, AI ready ecosystems.
Read More
Event-Driven Agentic AI with Data Streaming using Apache Kafka and Flink
Read More

How Apache Kafka and Flink Power Event-Driven Agentic AI in Real Time

Agentic AI marks a major evolution in artificial intelligence—shifting from passive analytics to autonomous, goal-driven systems capable of planning and executing complex tasks in real time. To function effectively, these intelligent agents require immediate access to consistent, trustworthy data. Traditional batch processing architectures fall short of this need, introducing delays, data staleness, and rigid workflows. This blog post explores why event-driven architecture (EDA)—powered by Apache Kafka and Apache Flink—is essential for building scalable, reliable, and adaptive AI systems. It introduces key concepts such as Model Context Protocol (MCP) and Google’s Agent-to-Agent (A2A) protocol, which are redefining interoperability and context management in multi-agent environments. Real-world use cases from finance, healthcare, manufacturing, and more illustrate how Kafka and Flink provide the real-time backbone needed for production-grade Agentic AI. The post also highlights why popular frameworks like LangChain and LlamaIndex must be complemented by robust streaming infrastructure to support stateful, event-driven AI at scale.
Read More
De Facto Standard API - Amazon S3 for Object Storage and Apache Kafka for Event Streaming
Read More

Kafka API is the De Facto Standard API for Event Streaming like Amazon S3 for Object Storage

Real-time beats slow data in most use cases across industries. The rise of event-driven architectures and data in motion powered by Apache Kafka enables enterprises to build real-time infrastructure and applications. This blog post explores why the Kafka API became the de facto standard API for event streaming like Amazon S3 for object storage, and the tradeoffs of these standards and corresponding frameworks, products, and cloud services.
Read More