Ingestion is the process of bringing the data to the target system, in this case, a big data storage system such as Hadoop, Mongo, or Cassandra, or any other system that can handle that amount of data efficiently. This data can either arrive in bulk format (batch ingestion) or can be a continuous stream of data (event or stream ingestion):
- Batch ingestion: Typically used when you want to ingest data from one source to another, for example, when you want to bring your CRM data into Hadoop. Typically, what you would do is do a data dump from your relational database and load this data dump into your big data storage platform. Hence, you are doing a bulk or batch ingest of data.
- Stream ingestion: When you have a continuous source ...