Define Environment Parameters
You must set up specific users and directories for your HDP installation by using the following instructions:
Define directories.
The following table describes the directories you need for installation, configuration, data storage, process IDs, and log information based on the Apache Hadoop Services you plan to install. Use this table to define what you are going to use to set up your environment.
Note The scripts.zip file that you downloaded in the supplied companion files includes a script, directories.sh, for setting directory environment parameters.
You should edit and source (or copy the contents to your ~/.bash_profile) to set up these environment variables in your environment.
Table 1.1. Directories Needed to Install Core Hadoop
Hadoop Service
Parameter
Definition
HDFS
DFS_NAME_DIR
Space separated list of directories to which NameNode should store the file system image: for example,
/grid/hadoop/hdfs/nn /grid1/hadoop/hdfs/nn
.HDFS
DFS_DATA_DIR
Space separated list of directories where DataNodes should store the blocks. For example,
/grid/hadoop/hdfs/dn /grid1/hadoop/hdfs/dn /grid2/hadoop/hdfs/dn
HDFS
FS_CHECKPOINT_DIR
Space separated list of directories where SecondaryNameNode should store the checkpoint image. For example,
/grid/hadoop/hdfs/snn /grid1/hadoop/hdfs/snn /grid2/hadoop/hdfs/snn
HDFS
HDFS_LOG_DIR
Directory for storing the HDFS logs. This directory name is a combination of a directory and the $HDFS_USER. For example,
/var/log/hadoop/hdfs
, where hdfs is the $HDFS_USER.HDFS
HDFS_PID_DIR
Directory for storing the HDFS process ID. This directory name is a combination of a directory and the $HDFS_USER. For example,
/var/run/hadoop/hdfs
, where hdfs is the $HDFS_USER.HDFS
HADOOP_CONF_DIR
Directory for storing the Hadoop configuration files. For example, /etc/hadoop/conf.
YARN
YARN_LOCAL_DIR
Space-separated list of directories where YARN should store temporary data. For example,
/grid/hadoop/yarn /grid1/hadoop/yarn /grid2/hadoop/yarn
YARN
YARN_LOG_DIR
Directory for storing the YARN logs. For example,
/var/log/hadoop/yarn
. This directory name is a combination of a directory and the $YARN_USER. In the example yarn is the $YARN_USER.YARN
YARN_LOCAL_LOG_DIR
Space-separated list of directories where YARN stores container log data. For example,
/grid/hadoop/yarn/logs /grid1/hadoop/yarn/log
.YARN
YARN_PID_DIR
Directory for storing the YARN process ID. For example,
/var/run/hadoop/yarn
. This directory name is a combination of a directory and the $YARN_USER. In the example, yarn is the $YARN_USER.MapReduce
MAPRED_LOG_DIR
Directory for storing the JobHistory Server logs. For example, /var/log/hadoop/mapred. This directory name is a combination of a directory and the $MAPRED_USER. In the example, mapred is the $MAPRED_USER.
Table 1.2. Directories Needed to Install Ecosystem Components
Hadoop Service
Parameter
Definition
Pig
PIG_CONF_DIR
Directory in which to store the Apache Pig configuration files: for example, /etc/pig/conf.
Pig
PIG_LOG_DIR
Directory to store the Pig logs. For example, /var/log/pig.
Pig
PIG_PID_DIR
Directory to store the Pig process ID. For example, /var/run/pig.
Oozie
OOZIE_CONF_DIR
Directory to store the Oozie configuration files. For example, /etc/oozie/conf.
Oozie
OOZIE_DATA
Directory to store the Oozie data. For example, /var/db/oozie.
Oozie
OOZIE_LOG_DIR
Directory to store the Oozie logs. For example, /var/log/oozie.
Oozie
OOZIE_PID_DIR
Directory to store the Oozie process ID. For example, /var/run/oozie.
Oozie
OOZIE_TMP_DIR
Directory to store the Oozie temporary files. For example, /var/tmp/oozie.
Hive
HIVE_CONF_DIR
Directory to store the Hive configuration files. For example, /etc/hive/conf.
Hive
HIVE_LOG_DIR
Directory to store the Hive logs. For example, /var/log/hive.
Hive
HIVE_PID_DIR
Directory to store the Hive process ID. For example, /var/run/hive.
WebHCat
WEBHCAT_CONF_DIR
Directory to store the WebHCat configuration files. For example, /etc/hcatalog/conf/webhcat.
WebHCat
WEBHCAT_LOG_DIR
Directory to store the WebHCat logs. For example, var/log/webhcat.
WebHCat
WEBHCAT_PID_DIR
Directory to store the WebHCat process ID. For example, /var/run/webhcat.
HBase
HBASE_CONF_DIR
Directory to store the Apache HBase configuration files. For example, /etc/hbase/conf.
HBase
HBASE_LOG_DIR
Directory to store the HBase logs. For example, /var/log/hbase.
HBase
HBASE_PID_DIR
Directory to store the HBase process ID. For example, /var/run/hbase.
ZooKeeper
ZOOKEEPER_DATA_DIR
Directory where Apache ZooKeeper stores data. For example, /grid/hadoop/zookeeper/data
ZooKeeper
ZOOKEEPER_CONF_DIR
Directory to store the ZooKeeper configuration files. For example, /etc/zookeeper/conf.
ZooKeeper
ZOOKEEPER_LOG_DIR
Directory to store the ZooKeeper logs. For example, /var/log/zookeeper.
ZooKeeper
ZOOKEEPER_PID_DIR
Directory to store the ZooKeeper process ID. For example, /var/run/zookeeper.
Sqoop
SQOOP_CONF_DIR
Directory to store the Apache Sqoop configuration files. For example, /etc/sqoop/conf.
If you use the companion files, the following screen provides a snapshot of how your
directories.sh
file should look after you edit the TODO variables:#!/bin/sh # # Directories Script # # 1. To use this script, you must edit the TODO variables below for your environment. # # 2. Warning: Leave the other parameters as the default values. Changing these default values requires you to # change values in other configuration files. # # # Hadoop Service - HDFS # # Space separated list of directories where NameNode stores file system image. For example, /grid/hadoop/hdfs/nn /grid1/hadoop/hdfs/nn DFS_NAME_DIR="TODO-LIST-OF-NAMENODE-DIRS"; # Space separated list of directories where DataNodes stores the blocks. For example, /grid/hadoop/hdfs/dn /grid1/hadoop/hdfs/dn /grid2/hadoop/hdfs/dn DFS_DATA_DIR="TODO-LIST-OF-DATA-DIRS"; # Space separated list of directories where SecondaryNameNode stores checkpoint image. For example, /grid/hadoop/hdfs/snn /grid1/hadoop/hdfs/snn /grid2/hadoop/hdfs/snn FS_CHECKPOINT_DIR="TODO-LIST-OF-SECONDARY-NAMENODE-DIRS"; # Directory to store the HDFS logs. HDFS_LOG_DIR="/var/log/hadoop/hdfs"; # Directory to store the HDFS process ID. HDFS_PID_DIR="/var/run/hadoop/hdfs"; # Directory to store the Hadoop configuration files. HADOOP_CONF_DIR="/etc/hadoop/conf"; # # Hadoop Service - YARN # # Space separated list of directories where YARN stores temporary data. For example, /grid/hadoop/yarn/local /grid1/hadoop/yarn/local /grid2/hadoop/yarn/local YARN_LOCAL_DIR="TODO-LIST-OF-YARN-LOCAL-DIRS"; # Directory to store the YARN logs. YARN_LOG_DIR="/var/log/hadoop/yarn"; # Space separated list of directories where YARN stores container log data. For example, /grid/hadoop/yarn/logs /grid1/hadoop/yarn/logs /grid2/hadoop/yarn/logs YARN_LOCAL_LOG_DIR="TODO-LIST-OF-YARN-LOCAL-LOG-DIRS"; # Directory to store the YARN process ID. YARN_PID_DIR="/var/run/hadoop/yarn"; # # Hadoop Service - MAPREDUCE # # Directory to store the MapReduce daemon logs. MAPRED_LOG_DIR="/var/log/hadoop/mapred"; # Directory to store the mapreduce jobhistory process ID. MAPRED_PID_DIR="/var/run/hadoop/mapred"; # # Hadoop Service - Hive # # Directory to store the Hive configuration files. HIVE_CONF_DIR="/etc/hive/conf"; # Directory to store the Hive logs. HIVE_LOG_DIR="/var/log/hive"; # Directory to store the Hive process ID. HIVE_PID_DIR="/var/run/hive"; # # Hadoop Service - WebHCat (Templeton) # # Directory to store the WebHCat (Templeton) configuration files. WEBHCAT_CONF_DIR="/etc/hcatalog/conf/webhcat"; # Directory to store the WebHCat (Templeton) logs. WEBHCAT_LOG_DIR="var/log/webhcat"; # Directory to store the WebHCat (Templeton) process ID. WEBHCAT_PID_DIR="/var/run/webhcat"; # # Hadoop Service - HBase # # Directory to store the HBase configuration files. HBASE_CONF_DIR="/etc/hbase/conf"; # Directory to store the HBase logs. HBASE_LOG_DIR="/var/log/hbase"; # Directory to store the HBase logs. HBASE_PID_DIR="/var/run/hbase"; # # Hadoop Service - ZooKeeper # # Directory where ZooKeeper stores data. For example, /grid1/hadoop/zookeeper/data ZOOKEEPER_DATA_DIR="TODO-ZOOKEEPER-DATA-DIR"; # Directory to store the ZooKeeper configuration files. ZOOKEEPER_CONF_DIR="/etc/zookeeper/conf"; # Directory to store the ZooKeeper logs. ZOOKEEPER_LOG_DIR="/var/log/zookeeper"; # Directory to store the ZooKeeper process ID. ZOOKEEPER_PID_DIR="/var/run/zookeeper"; # # Hadoop Service - Pig # # Directory to store the Pig configuration files. PIG_CONF_DIR="/etc/pig/conf"; # Directory to store the Pig logs. PIG_LOG_DIR="/var/log/pig"; # Directory to store the Pig process ID. PIG_PID_DIR="/var/run/pig"; # # Hadoop Service - Oozie # # Directory to store the Oozie configuration files. OOZIE_CONF_DIR="/etc/oozie/conf" # Directory to store the Oozie data. OOZIE_DATA="/var/db/oozie" # Directory to store the Oozie logs. OOZIE_LOG_DIR="/var/log/oozie" # Directory to store the Oozie process ID. OOZIE_PID_DIR="/var/run/oozie" # Directory to store the Oozie temporary files. OOZIE_TMP_DIR="/var/tmp/oozie" # # Hadoop Service - Sqoop # SQOOP_CONF_DIR="/etc/sqoop/conf" # # Hadoop Service - Accumulo # ACCUMULO_CONF_DIR="/etc/accumulo/conf"; ACCUMULO_LOG_DIR="/var/log/accumulo"
The following table describes system user account and groups. Use this table to define what you are going to use in setting up your environment. These users and groups should reflect the accounts you create in Create System Users and Groups. The
scripts.zip
file you downloaded includes a script, usersAndGroups.sh, for setting user and group environment parameters.Table 1.3. Define Users and Groups for Systems
Parameter
Definition
HDFS_USER
User that owns the Hadoop Distributed File Sysem (HDFS) services. For example, hdfs.
YARN_USER
User that owns the YARN services. For example, yarn.
ZOOKEEPER_USER
User that owns the ZooKeeper services. For example, zookeeper.
HIVE_USER
User that owns the Hive services. For example, hive.
WEBHCAT_USER
User that owns the WebHCat services. For example, hcat.
HBASE_USER
User that owns the HBase services. For example, hbase.
FALCON_USER
User that owns the Apache Falcon services. For example, falcon.
SQOOP_USER
User owning the Sqoop services. For example, sqoop.
KAFKA_USER
User owning the Apache Kafka services. For example, kafka.
OOZIE_USER
User owning the Oozie services. For example oozie.
STORM_USER
User owning the Storm Services. For example, storm.
HADOOP_GROUP
A common group shared by services. For example, hadoop.
ACCUMULO_USER
User that owns the Accumulo services. For example, accumulo.
KNOX_USER
User that owns the Knox Gateway services. For example, knox.
NAGIOS_USER
User that owns the Nagios services. For example, nagios.