Configuring TLS/SSL for HDFS, YARN and MapReduce
Required Role: Configurator, Cluster Administrator, or Full Administrator
TLS/SSL for the core Hadoop services—HDFS, MapReduce, and YARN—must be enabled as a group. Because most clusters run either MapReduce or YARN, not both, you will typically enable HDFS and YARN, or HDFS and MapReduce. Enabling TLS/SSL on HDFS is required before it can be enabled on either MapReduce or YARN.
The steps below include enabling Kerberos authentication for HTTP Web-Consoles. Enabling TLS/SSL for the core Hadoop services on a cluster without enabling authentication displays a warning.
Before You Begin
- Before enabling TLS/SSL, keystores containing certificates bound to the appropriate domain names will need to be accessible on all hosts on which at least one HDFS, MapReduce, or YARN daemon role is running.
- Since HDFS, MapReduce, and YARN daemons act as TLS/SSL clients as well as TLS/SSL servers, they must have access to truststores. In many cases, the most practical approach is to deploy truststores to all hosts in the cluster, as it may not be desirable to determine in advance the set of hosts on which clients will run.
- Keystores for HDFS, MapReduce and YARN must be owned by the hadoop group, and have permissions 0440 (that is, readable by owner and group). Truststores must have permissions 0444 (that is, readable by all)
- Cloudera Manager supports TLS/SSL configuration for HDFS, MapReduce and YARN at the service level. For each of these services, you must specify absolute paths to the keystore and
truststore files. These settings apply to all hosts on which daemon roles of the service in question run. Therefore, the paths you choose must be valid on all hosts.
An implication of this is that the keystore file names for a given service must be the same on all hosts. If, for example, you have obtained separate certificates for HDFS daemons on hosts node1.example.com and node2.example.com, you might have chosen to store these certificates in files called hdfs-node1.keystore and hdfs-node2.keystore (respectively). When deploying these keystores, you must give them both the same name on the target host — for example, hdfs.keystore.
- Multiple daemons running on a host can share a certificate. For example, in case there is a DataNode and an Oozie server running on the same host, they can use the same certificate.
Configuring TLS/SSL for HDFS
- Go to the HDFS service.
- Click the Configuration tab.
- Select .
- Select .
- In the Search field, type TLS/SSL to show the TLS/SSL properties (found under the Service-Wide > Security category).
- Edit the following properties according to your cluster configuration:
Property Description Hadoop TLS/SSL Server Keystore File Location Path to the keystore file containing the server certificate and private key. Hadoop TLS/SSL Server Keystore File Password Password for the server keystore file. Hadoop TLS/SSL Server Keystore Key Password Password that protects the private key contained in the server keystore. - If you are not using the default truststore, configure TLS/SSL client truststore properties:
Property Description Cluster-Wide Default TLS/SSL Client Truststore Location Path to the client truststore file. This truststore contains certificates of trusted servers, or of Certificate Authorities trusted to identify servers. Cluster-Wide Default TLS/SSL Client Truststore Password Password for the client truststore file. - (Optional) Cloudera recommends you enable web UI authentication for the HDFS service. Web UI authentication uses SPNEGO. After enabling this, you cannot
access the Hadoop web consoles without a valid Kerberos ticket and proper client-side configuration. For more information, see How to Configure Browsers for Kerberos Authentication.
To enable web UI authentication, enter web consoles in the Search field to bring up the Enable Authentication for HTTP Web-Consoles property (found under the Service-Wide>Security category). Check the property to enable web UI authentication.
Enable Authentication for HTTP Web-Consoles Enables authentication for Hadoop HTTP web-consoles for all roles of this service. - Click Save Changes.
- Follow the procedure described in the following Configuring TLS/SSL for YARN and MapReduce section, at the end of which you will be instructed to restart all the affected services (HDFS, MapReduce and YARN).
Configuring TLS/SSL for YARN or MapReduce
- Go to the YARN or MapReduce service.
- Click the Configuration tab.
- Select .
- Select .
- Locate the <property name> property or search for it by typing its name in the Search box.
- In the Search field, type TLS/SSL to show the TLS/SSL properties (found under the Service-Wide > Security category).
- Edit the following properties according to your cluster configuration:
Property Description Hadoop TLS/SSL Server Keystore File Location Path to the keystore file containing the server certificate and private key. Hadoop TLS/SSL Server Keystore File Password Password for the server keystore file. Hadoop TLS/SSL Server Keystore Key Password Password that protects the private key contained in the server keystore. - Configure the following TLS/SSL client truststore properties for MRv1 or YARN only if you want to override the cluster-wide defaults set by the HDFS properties configured above.
Property Description TLS/SSL Client Truststore File Location Path to the client truststore file. This truststore contains certificates of trusted servers, or of Certificate Authorities trusted to identify servers. TLS/SSL Client Truststore File Password Password for the client truststore file. - Cloudera recommends you enable Web UI authentication for the service in question.
Enter web consoles in the Search field to bring up the Enable Authentication for HTTP Web-Consoles property (found under the Service-Wide>Security category). Check the property to enable web UI authentication.
Enable Authentication for HTTP Web-Consoles Enables authentication for Hadoop HTTP web-consoles for all roles of this service. - Enter a Reason for change, and then click Save Changes to commit the changes.
- Go to the HDFS service
- Click the Configuration tab.
- Type Hadoop SSL Enabled in the Search box.
- Select the Hadoop SSL Enabled property to enable SSL communication for HDFS, MapReduce, and YARN.
Property Description Hadoop TLS/SSL Enabled Enable TLS/SSL encryption for HDFS, MapReduce, and YARN web UIs, as well as encrypted shuffle for MapReduce and YARN. - Enter a Reason for change, and then click Save Changes to commit the changes.
- Restart all affected services (HDFS, MapReduce and YARN), as well as their dependent services.
Configuring HSTS for HDFS
Configuring the HTTP Strict Transport Security (HSTS) for HDFS ensures that a web browser does not load the service information using HTTP. Additionally, all attempts to load the information using HTTP will automatically be converted to HTTPS.
Perform the following steps to configure HSTS for HDFS:
- Go to the HDFS service.
- Click the Configuration tab.
- Set the HSTS credentials in Cluster-wide Advanced Configuration Snippet (Safety Valve) for core-site.xml.
<property> <name>hadoop.http.header.Strict_Transport_Security</name> <value>max-age=63072000;includeSubDomains;preload</value> </property>
- If required, configure additional headers by using the safety value specified in the previous step for the hadoop.http.header.http-header property.
- Enter a Reason for change, and then click Save Changes to commit the changes.
- Restart the HDFS service.