Creating a Compute Cluster and Data Context
Compute cluster needs Data Context to communicate with the Base cluster and you must create one.
Minimum Required Role: Cluster Administrator (also provided by Full Administrator)
To create a Compute cluster, you must have a Base cluster. The Base cluster contains data that is used by the Compute cluster.
To create a Compute cluster:
- On the Cloudera Manager home page, click
Clusters > Add
ClusterThe Add Cluster Welcome page displays.
- Click Continue. .The Cluster Basics page displays
- Select Compute cluster.
- If you already have a Data Context defined, select it from the drop-down list.
- To create a new Data Context:
- Select Create Data Context from
the drop-down list.The Create Data Context dialog box displays.
- Enter a unique name for the Data Context.
- Select the Base cluster from the drop-down list.
-
Select the services you want to expose in the Data Context. You can choose from the
following:
- HDFS (required)
- Hive Metadata Service
- Atlas
- Ranger
- Ozone
- Click
Create.The Cluster Basics page displays your selections.
- Click Continue.
- Select Create Data Context from
the drop-down list.
- Continue with the next steps in the Add Cluster Wizard to
specify hosts and credentials, and install the Agent and CDH software.
The Select Repository screen examines the CDH version of the base cluster and recommend a supported version. Cloudera recommends that your Base and Compute clusters each run the same version of CDH. The Add Cluster Wizard offers the option to choose other versions, but these combinations have not been tested and are not supported for production use.
-
On the Select Services screen, choose any of the pre-configured combinations of
services listed on this page, or you can select Custom Services and
choose the services you want to install.
Service combinations for Compute Clusters:
The following services can be installed on a Compute cluster:- Hive Execution Service (This service supplies the HiveServer2 role only.)
- Hue
- Kafka
- Spark 2
- Oozie (only when Hue is available, and is a requirement for Hue)
- YARN
- HDFS
- Stub DFS (Stub DFS replaces Core Settings and requires the Storage Operations role.)
- If you have enabled Kerberos authentication on the Base cluster, you must also enable Kerberos on the Compute cluster.
- If you don't select a service from Step 7 but if you want to use integrations related to these services (such as Hive, Spark, Kafka, etc.) on the Compute cluster, then you need to add all the necessary gateway roles to use the connections properly.