Chapter 12. Hive Integration with Other Frameworks

In this chapter, you will learn the following topics:

  • Working with Apache Spark
  • Working with Accumulo
  • Working with HBase
  • Working with Google Drill

Working with Apache Spark

In this recipe, you will learn how to integrate Hive with Apache Spark. Apache Spark is an open source cluster computing framework. It is used as a replacement of the MapReduce framework.

Getting ready

In this topic, we will cover the use of Hive and Apache Spark. You must have Apache Spark installed on your system before going further in the topic.

  1. Once the Spark is installed, start the Spark master server by executing the following command:
    ./sbin/start-master.sh
    
  2. Check whether the Spark master server has been started or not by issuing ...

Get Apache Hive Cookbook now with the O’Reilly learning platform.

O’Reilly members experience books, live events, courses curated by job role, and more from O’Reilly and nearly 200 top publishers.