Categories: Uncategorized

WJAX 2013 Slides online: Big Data beyond Apache Hadoop – How to integrate ALL your Data with Camel and Talend

I just want to publish my updated slide deck from WJAX 2013 in Munich. I have integrated some real world use cases for showing Hadoop integration with Apache Camel and Talend Open Studio for Big Data.

Abstract

Big data represents a significant paradigm shift in enterprise technology. Big data radically changes the nature of the data management profession as it introduces new concerns about the volume, velocity and variety of corporate data. Apache Hadoop is the open source defacto standard for implementing big data solutions on the Java platform. Hadoop consists of its kernel, MapReduce, and the Hadoop Distributed Filesystem (HDFS). A challenging task is to send all data to Hadoop for processing and storage (and then get it back to your application later), because in practice data comes from many different applications (SAP, Salesforce, Siebel, etc.) and databases (File, SQL, NoSQL), uses different technologies and concepts for communication (e.g. HTTP, FTP, RMI, JMS), and consists of different data formats using CSV, XML, binary data, or other alternatives. This session shows different open source frameworks and products to solve this challenging task. Learn how to use every thinkable data with Hadoop – without plenty of complex or redundant boilerplate code.

Slides

You are currently viewing a placeholder content from Default. To access the actual content, click the button below. Please note that doing so will share data with third-party providers.

More Information

Kai Waehner

bridging the gap between technical innovation and business value for real-time data streaming and applied AI.

Share
Published by
Kai Waehner

Recent Posts

The Trinity of Modern Data Architecture: Process Intelligence, Event-Driven Integration, and Trusted Agentic AI

Agentic AI without governed processes is fast but ungoverned. Event-driven integration without process intelligence moves…

3 hours ago

dbt Meets Apache Flink: One Workflow for Data Engineers on Snowflake, BigQuery, Databricks, and Confluent

Two toolchains, two skill sets, two CI/CD pipelines — that has been the reality for…

6 days ago

The Shift Left Architecture 2.0: Operational, Analytical and AI Interfaces for Real-Time Data Products

The Shift Left Architecture moves data integration logic into an event-driven architecture where governed data…

1 week ago

UFC VIP Experience Worth the Price? Fan Review. Business Perspective. Tech Vision.

The Ultimate Fighting Championship (UFC) held Fight Night London on March 21, 2026, at The…

1 week ago

Dashboards and Queries for Apache Kafka: Operational, Explorative, and the Role of the Context Engine

Dashboards are a popular way to make streaming data visible and useful, but they are…

3 weeks ago

Data Streaming at MWC 2026: How Apache Kafka, Flink and Agentic AI Power Telecom Trends

Mobile World Congress (MWC) 2026 highlights the shift from batch systems to real time data…

4 weeks ago