Compute cluster needs Data Context to communicate with the Cloudera Base on premises cluster and you must create one.
Minimum Required Role:
Cluster Administrator (also
provided by Full Administrator)
To create a Compute cluster, you must have a Cloudera Base on premises cluster. The Cloudera Base on premises cluster contains data that is used by
the Compute cluster.
To create a Compute cluster:
-
On the Cloudera Manager home page, click
The Add Cluster Welcome page displays.
- Click Continue. .
The
Cluster Basics page
displays
- Select Compute cluster.
- If you already have a Data Context defined, select it from the
drop-down list.
- To create a new Data Context:
- Select Create Data Context from
the drop-down list.
The Create Data
Context dialog box displays.
- Enter a unique name for the Data
Context.
- Select the Base cluster from the drop-down
list.
-
Select the services you want to expose in the Data Context. You can choose from the
following:
- HDFS (required)
- Hive Metadata Service
- Atlas
- Ranger
- Ozone
- Click
Create.
The Cluster Basics page displays your selections.
- Click Continue.
- Continue with the next steps in the Add Cluster Wizard to
specify hosts and credentials, and install the Agent and CDH software.
The Select Repository screen examines the CDH version of the base cluster and recommend
a supported version. Cloudera recommends that your
Cloudera Base on premises and Compute clusters each run the
same version of CDH. The Add Cluster Wizard offers the option to choose other versions,
but these combinations have not been tested and are not supported for production
use.
-
On the Select Services screen, choose any of the pre-configured combinations of
services listed on this page, or you can select Custom Services and
choose the services you want to install.
Service combinations for Compute Clusters:
The following services can be installed on a Compute cluster:
- Hive Execution Service (This service supplies the
HiveServer2 role only.)
- Hue
- Kafka
- Spark
- Oozie (only when Hue is available, and is a requirement
for Hue)
- YARN
- HDFS
- Stub DFS (Stub DFS replaces Core Settings and requires the Storage Operations
role.)
- If you have enabled Kerberos authentication on the Base
cluster, you must also enable Kerberos on the Compute
cluster.
-
If you don't select a service from Step
7 but if you want to use integrations related to these services (such as Hive,
Spark, Kafka, etc.) on the Compute cluster, then you need to add all the necessary gateway
roles to use the connections properly.