Spark QuickStart Guide
Also available as:

Chapter 2. Prerequisites

Before installing Spark, make sure your cluster meets the following prerequisites.

Table 2.1. Prerequisites for running Spark 1.3.1

Cluster Stack Version
  • HDP 2.2.6 or later

(Optional) Ambari
  • Version 2.1 or later

Software dependencies
  • Spark requires HDFS and YARN

  • PySpark requires Python to be installed on all nodes


If you installed the tech preview, save any configuration changes you made to the tech preview environment. Install Spark, and then update the configuration with your changes.