Decommissioning and Recommissioning Hosts
Decommissioning a host decommissions and stops all roles on the host without having to go to each service and individually decommission the roles. Decommissioning applies to only to HDFS DataNode, MapReduce TaskTracker, YARN NodeManager, and HBase RegionServer roles. If the host has other roles running on it, those roles are stopped.
Once all roles on the host have been decommissioned and stopped, the host can be removed from service. You can decommission multiple hosts in parallel.
- Decommissioning Hosts
- Tuning Prior to Decommissioning a Large Number of Hosts
- Recommissioning Hosts
- Restarting All The Roles on a Recommissioned Host
Decommissioning Hosts
If you are planning to decommission a large number of hosts, see Tuning Prior to Decommissioning a Large Number of Hosts.
- Click the Hosts tab.
- Select one or more hosts to decommission.
- Select .
A confirmation pop-up informs you of the roles that will be decommissioned or stopped on the hosts you have selected. To proceed with the decommissioning, click Confirm.
A Command Details window appears that will show each stop or decommission command as it is run, service by service. You can click one of the decommission links to see the subcommands that are run for decommissioning a given role. Depending on the role, the steps may include adding the host to an "exclusions list" and refreshing the NameNode, JobTracker, or NodeManager, stopping the Balancer (if it is running), and moving data blocks or regions. Roles that do not have specific decommission actions are stopped.
While decommissioning is in progress, the host displays the icon. Once all roles have been decommissioned or stopped, the host displays the icon. If one host in a cluster has been decommissioned, the DECOMMISSIONED facet displays in the Filters on the Hosts page and you can filter the hosts according to their decommission status.
You cannot start roles on a decommissioned host.
Tuning Prior to Decommissioning a Large Number of Hosts
- Raise the heap size of the DataNodes. DataNodes should be configured
with at least 4 GB heap size to allow for the increase in iterations and max
streams.
- Go to the HDFS service page.
- Click the Configuration tab.
- Under each DataNode role group (DataNode Default Group and any additional DataNode role groups) go to the Resource Management category, and set the Java Heap Size of DataNode in Bytes property as recommended.
- Increase the replication work multiplier to a larger number (the
default is 2, however 10 is recommended):
- Go to the NameNode Default Group > Advanced category.
- Search for dfs.namenode.replication and increase the dfs.namenode.replication.work.multiplier.per.iteration property to a value such as 10.
- Click Save Changes.
- Set the replication maximum streams to a value such as 100.
- Go to the Service-Wide > Advanced category.
- In the HDFS
Service Configuration Advanced Configuration Snippet for
hdfs-site.xml, paste in the
following:
<property> <name>dfs.namenode.replication.max-streams</name> <value>100</value> </property
- Click Save Changes.
- Restart the HDFS service.
Recommissioning Hosts
Only hosts that are decommissioned using Cloudera Manager can be recommissioned.
- Click the Hosts tab.
- Select one or more hosts to recommission.
- Select .
The icon is removed from the host and from the roles that reside on the host. However, the roles themselves are not restarted.
Restarting All The Roles on a Recommissioned Host
- Click the Hosts tab.
- Select one or more hosts on which to start recommissioned roles.
- Select .
<< Maintenance Mode | Deleting Hosts >> | |