“Hadoop and Data Warehouse (DWH) – Friends, Enemies or Profiteers? What about Real Time?” – Slides (including TIBCO Examples) from JAX 2014 Online

Slides from my talk “Hadoop and Data Warehouse (DWH) – Friends, Enemies or Profiteers? What about Real Time?” at JAX 2014 (Twitter #jaxcon) in Mainz are online. JAX is a great conference with interesting topics and many good speakers!

Content (Data Warehouse, Business Intelligence, Hadoop, Stream Processing)

Big data represents a significant paradigm shift in enterprise technology. Big data radically changes the nature of the data management profession as it introduces new concerns about the volume, velocity and variety of corporate data. New business models based on predictive analytics, such as recommendation systems or fraud detection, are relevant more than ever before. Apache Hadoop seems to become the de facto standard for implementing big data solutions. For that reason, solutions from many different vendors emerged on top of Hadoop.

But hold on… Companies have spent a lot of many to implement a data warehouse for the same reason in the last decades. Both, Apache Hadoop and data warehouse were invented to store and analyze big data. This session explains the different architectural and technical concepts of Apache Hadoop and a data warehouse. The following questions will be answered: When to use which alternative? Does a data warehouse even have a future at all? Or how can we combine both alternatives?

However, Hadoop and a Data Warehouse cannot solve every big data problem. Complex event processing and real-time analytics have to be solved in another way. So, in-memory computing and streaming platforms are good alternatives or complements to Hadoop for processing and analyzing big data. For that reasons, an almost unimaginable number of solutions for big data emerged on the market. This session shows and compares the most important concepts and solutions for processing and analyzing big data, and discusses how they complement each other.

TIBCO Products (Spotfire, StreamBase, BusinessEvents, BusinessWorks) and Real World Examples

I discuss a good big data architecture which includes Data Warehouse / Business Intelligence + Apache Hadoop + Real Time / Stream Processing. Several real world example are shown. TIBCO offers some very nice products for realizing these use cases, e.g. Spotfire (Business Intelligence / BI), StreamBase (Stream Processing), BusinessEvents (Complex Event Processing / CEP) and BusinessWorks (Integration / ESB). TIBCO is also ready for Hadoop by offering connectors and plugins for many important Hadoop frameworks / interfaces such as HDFS, Pig, Hive, Impala, Apache Flume and more.

Slides

Here are the slides:

Click on the button to load the content from www.slideshare.net.

Load content

As always, I appreciate feedback and discussions.

Kai Wähner

Kai Waehner

bridging the gap between technical innovation and business value for real-time data streaming, processing and analytics

Recent Posts

Mainframe Integration with Data Streaming: Architecture, Business Value, Real-World Success

The mainframe is evolving—not fading. With cloud-native features, AI acceleration, and quantum-safe encryption, platforms like…

2 days ago

How OpenAI uses Apache Kafka and Flink for GenAI

OpenAI revealed how it builds and scales the real-time data streaming infrastructure that powers its…

6 days ago

­­The Rise of the Durable Execution Engine (Temporal, Restate) in an Event-driven Architecture (Apache Kafka)

Durable execution engines like Temporal and Restate are redefining how developers orchestrate long-running, stateful workflows…

1 week ago

How Penske Logistics Transforms Fleet Intelligence with Data Streaming and AI

Real-time visibility has become essential in logistics. As supply chains grow more complex, providers must…

2 weeks ago

Data Streaming Meets the SAP Ecosystem and Databricks – Insights from SAP Sapphire Madrid

SAP Sapphire 2025 in Madrid brought together global SAP users, partners, and technology leaders to…

3 weeks ago

Agentic AI with the Agent2Agent Protocol (A2A) and MCP using Apache Kafka as Event Broker

Agentic AI is emerging as a powerful pattern for building autonomous, intelligent, and collaborative systems.…

3 weeks ago