Use the following instructions to decommission DataNodes in your cluster:
On the NameNode host machine, edit the
file and add the list of DataNodes hostnames (separated by newline character).$HADOOP_CONF_DIR
/dfs.excludewhere
$HADOOP_CONF_DIR
is the directory for storing the Hadoop configuration files. For example,/etc/hadoop/conf
.Update the NameNode with the new set of permitted DataNodes. On the NameNode host machine, execute the following command:
su - $HDFS_USER % hadoop dfsadmin –refreshNodes
where
$HDFS_USER
is the user owning the HDFS services. For example,hdfs
.Open the NameNode web UI (http://
) and navigate to page.$NameNode_FQDN
:50070Check whether the admin state has changed to Decommission In Progress for the DataNodes being decommissioned.
When all the DataNodes report their state as Decommissioned then all the blocks have been replicated. Shut down the decommissioned nodes.
On the NameNode host machine, remove the decommissioned nodes from the
file and execute the following command:$HADOOP_CONF_DIR
/dfs.includesu - $HDFS_USER % hadoop dfsadmin –refreshNodes