Creating a Compute Cluster and Data Context

Compute cluster needs Data Context to communicate with the Base cluster and you must create one.

Minimum Required Role: Cluster Administrator (also provided by Full Administrator)

To create a Compute cluster, you must have a Base cluster. The Base cluster contains data that is used by the Compute cluster.

To create a Compute cluster:

  1. On the Cloudera Manager home page, click Clusters > Add Cluster
    The Add Cluster Welcome page displays.
  2. Click Continue. .
    The Cluster Basics page displays
  3. Select Compute cluster.
  4. If you already have a Data Context defined, select it from the drop-down list.
  5. To create a new Data Context:
    1. Select Create Data Context from the drop-down list.
      The Create Data Context dialog box displays.
    2. Enter a unique name for the Data Context.
    3. Select the Base cluster from the drop-down list.
    4. Select the services you want to expose in the Data Context. You can choose from the following:
      • HDFS (required)
      • Hive Metadata Service
      • Atlas
      • Ranger
      • Ozone
    5. Click Create.
      The Cluster Basics page displays your selections.
    6. Click Continue.
  6. Continue with the next steps in the Add Cluster Wizard to specify hosts and credentials, and install the Agent and CDH software.
    The Select Repository screen examines the CDH version of the base cluster and recommend a supported version. Cloudera recommends that your Base and Compute clusters each run the same version of CDH. The Add Cluster Wizard offers the option to choose other versions, but these combinations have not been tested and are not supported for production use.
  7. On the Select Services screen, choose any of the pre-configured combinations of services listed on this page, or you can select Custom Services and choose the services you want to install.

    Service combinations for Compute Clusters:

    The following services can be installed on a Compute cluster:
    • Hive Execution Service (This service supplies the HiveServer2 role only.)
    • Hue
    • Kafka
    • Spark 2
    • Oozie (only when Hue is available, and is a requirement for Hue)
    • YARN
    • HDFS
    • Stub DFS (Stub DFS replaces Core Settings and requires the Storage Operations role.)
  8. If you have enabled Kerberos authentication on the Base cluster, you must also enable Kerberos on the Compute cluster.