In this section, we will create the following three pipelines:
- Streaming pipeline with Kafka - Storm - MySQL
- Streaming pipeline with Kafka - Storm - HDFS - Hive
In this section, we will see how data streams flow from Kafka to Storm to MySQL table.
The whole pipeline will work as follows:
- We will ingest customer records (customer_firstname and customer_lastname) in Kafka using the Kafka console-producer API.
- After that, Storm will pull the messages from Kafka.
- A connection to MySQL will be established.
- Storm will use MySQL-Bolt to ingest records into MySQL table. MySQL will automatically generate customer_id.
- The MySQL table data (customer_id, customer_firstname, and customer_lastname ...