Command Line Upgrade
Also available as:
PDF
loading table of contents...

Abstract

The Hortonworks Data Platform, powered by Apache Hadoop, is a massively scalable and 100% open source platform for storing, processing and analyzing large volumes of data. It is designed to deal with data from many sources and formats in a very quick, easy and cost-effective manner. The Hortonworks Data Platform consists of the essential set of Apache Software Foundation projects that focus on the storage and processing of Big Data, along with operations, security, and governance for the resulting system. This includes Apache Hadoop -- which includes MapReduce, Hadoop Distributed File System (HDFS), and Yet Another Resource Negotiator (YARN) -- along with Ambari, Falcon, Flume, HBase, Hive, Kafka, Knox, Oozie, Phoenix, Pig, Ranger, Slider, Spark, Sqoop, Storm, Tez, and ZooKeeper. Hortonworks is the major contributor of code and patches to many of these projects. These projects have been integrated and tested as part of the Hortonworks Data Platform release process and installation and configuration tools have also been included.

Unlike other providers of platforms built using Apache Hadoop, Hortonworks contributes 100% of our code back to the Apache Software Foundation. The Hortonworks Data Platform is Apache-licensed and completely open source. We sell only expert technical support, training and partner-enablement services. All of our technology is, and will remain, free and open source.

Please visit the Hortonworks Data Platform page for more information on Hortonworks technology. For more information on Hortonworks services, please visit either the Support or Training page. Feel free to contact us directly to discuss your specific needs.


Contents

1. Upgrade from HDP 2.4 to Manually
Getting Ready to Upgrade
Upgrade HDP 2.4 Components
Symlink Directories with hdp-select
Configure and Start Apache ZooKeeper
Configure Hadoop
Start Hadoop Core
Verify HDFS Filesystem Health
Configure YARN and MapReduce
Start YARN/MapReduce Services
Run Hadoop Smoke Tests
Configure and Start Apache HBase
Configure Apache Phoenix
Configure and Start Apache Accumulo
Configure and Start Apache Tez
Configure and Start Apache Hive and Apache HCatalog
Configure and Start Apache Oozie
Configure and Start Apache WebHCat
Configure Apache Pig
Configure and Start Apache Sqoop
Configure, Start, and Validate Apache Flume
Configure, Start, and Validate Apache Mahout
Configure and Start Hue
Configure and Start Apache Knox
Upgrade the Knox Gateway
Verify the Knox Upgrade
Downgrade the Knox Gateway to the Previous Version
Verify the Knox Downgrade Was Successful
Configure and Validate Apache Falcon
Configure and Start Apache Storm
Configure and Start Apache Ranger
Preparing Your Cluster to Upgrade Ranger
Stop the Ranger Services
Preparing the Cluster for Upgrade
Registering the Repo
Install the Ranger Components
Restart the Ranger Services
Enable Ranger Plugins
Enable KMS Configuration
Configure and Start Apache Ranger on a Kerberized Cluster
Configuring and Upgrading Apache Spark
Upgrade Apache Slider
Upgrade Apache Kafka
Downgrading Kafka
Finalize the Upgrade
Migrate the Audit Logs from DB to Solr
Install New Services
2. Upgrade from HDP 2.3 to Manually
Getting Ready to Upgrade
Upgrade HDP 2.3 Components
Symlink Directories with hdp-select
Configure and Start Apache ZooKeeper
Configure Hadoop
Start Hadoop Core
Verify HDFS Filesystem Health
Configure YARN and MapReduce
Start YARN/MapReduce Services
Run Hadoop Smoke Tests
Configure and Start Apache HBase
Configure Apache Phoenix
Configure and Start Apache Accumulo
Configure and Start Apache Tez
Configure and Start Apache Hive and Apache HCatalog
Configure and Start Apache Oozie
Configure and Start Apache WebHCat
Configure Apache Pig
Configure and Start Apache Sqoop
Configure, Start, and Validate Apache Flume
Configure, Start, and Validate Apache Mahout
Configure and Start Hue
Configure and Start Apache Knox
Upgrade the Knox Gateway
Verify the Knox Upgrade
Downgrade the Knox Gateway to the Previous Version
Verify the Knox Downgrade Was Successful
Configure and Validate Apache Falcon
Configure and Start Apache Storm
Configure and Start Apache Ranger
Preparing Your Cluster to Upgrade Ranger
Stop the Ranger Services
Preparing the Cluster for Upgrade
Registering the Repo
Install the Ranger Components
Restart the Ranger Services
Enable Ranger Plugins
Enable KMS Configuration
Configure and Start Apache Ranger on a Kerberized Cluster
Configuring and Upgrading Apache Spark
Upgrade Apache Slider
Upgrade Apache Kafka
Downgrading Kafka
Finalize the Upgrade
Migrate the Audit Logs from DB to Solr
Install New Services
3. Upgrade from HDP 2.2 to Manually
Getting Ready to Upgrade
Upgrade HDP 2.2 Components
Symlink Directories with hdp-select
Configure and Start Apache ZooKeeper
Configure Hadoop
Start Hadoop Core
Verify HDFS Filesystem Health
Configure YARN and MapReduce
Start YARN/MapReduce Services
Run Hadoop Smoke Tests
Configure and Start Apache HBase
Configure Apache Phoenix
Configure and Start Apache Accumulo
Configure and Start Apache Tez
Configure and Start Apache Hive and Apache HCatalog
Configure and Start Apache Oozie
Configure and Start Apache WebHCat
Configure Apache Pig
Configure and Start Apache Sqoop
Configure, Start, and Validate Apache Flume
Configure, Start, and Validate Apache Mahout
Configure and Start Hue
Configure and Start Apache Knox
Upgrade the Knox Gateway
Verify the Knox Upgrade
Configure and Validate Apache Falcon
Configure and Start Apache Storm
Configure and Start Apache Ranger
Preparing Your Cluster to Upgrade Ranger
Stop the Ranger Services
Preparing the Cluster for Upgrade
Registering the HDP 2.4 Repo
Install the Ranger Components
Restart the Ranger Services
Enable Ranger Plugins
Enable KMS Configuration
Configure and Start Apache Ranger on a Kerberized Cluster
Configuring and Upgrading Apache Spark
Upgrade Apache Slider
Upgrade Apache Kafka
Downgrading Kafka
Finalize the Upgrade
Migrate the Audit Logs from DB to Solr
Install New Services
4. Upgrade from HDP 2.1 to Manually
Getting Ready to Upgrade
Upgrade HDP 2.1 Components
Symlink Directories with hdp-select
Configure and Start Apache ZooKeeper
Configure Hadoop
Start Hadoop Core
Verify HDFS Filesystem Health
Configure YARN and MapReduce
Start YARN/MapReduce Services
Run Hadoop Smoke Tests
Configure and Start Apache HBase
Configure Apache Phoenix
Configure and Start Apache Accumulo
Configure and Start Apache Tez
Configure and Start Apache Hive and Apache HCatalog
Configure and Start Apache Oozie
Configure and Start Apache WebHCat
Configure Apache Pig
Configure and Start Apache Sqoop
Configure, Start, and Validate Apache Flume
Configure and Validate Apache Mahout
Configure and Start Hue
Configure and Start Apache Knox
Upgrade the Knox Gateway
Verify the Knox Upgrade
Downgrade the Knox Gateway to the Previous Version
Verify the Knox Downgrade Was Successful
Configure and Validate Apache Falcon
Configure and Start Apache Storm
Configure and Start Apache Ranger
Preparing Your Cluster to Upgrade Ranger
Stop the Ranger Services
Install the Ranger Components
Restart the Ranger Services
Remove Existing Startup Files and Symbolic Links
Enable Ranger Plugins
Configuring and Upgrading Apache Spark
Upgrade Apache Slider
Upgrade Apache Kafka
Downgrading Kafka
Finalize the Upgrade
Install New HDP 2.5 Services
5. Upgrade from HDP 2.0 to Manually
Getting Ready to Upgrade
Upgrade HDP 2.0 Components
Symlink Directories with hdp-select
Configure and Start Apache ZooKeeper
Configure Hadoop
Set RPC Authentication
Start Hadoop Core
Verify HDFS Filesystem Health
Configure YARN and MapReduce
Start YARN/MapReduce Services
Run Hadoop Smoke Tests
Configure and Start Apache HBase
Configure and Start Apache Hive and Apache HCatalog
Configure and Start Apache Oozie
Configure and Start Apache WebHCat (Templeton)
Configure and Start Apache Pig
Configure and Start Apache Sqoop
Configure, Start, and Validate Apache Flume
Configure, Start, and Validate Apache Mahout
Configure and Start Hue
Finalize the Upgrade
Install New HDP 2.5 Services
6. Upgrade from HDP 1.3 to Manually
Getting Ready to Upgrade
Upgrade HDP 1.3 Components
Symlink Directories with hdp-select
Configure and Start Apache ZooKeeper
Configure and Start Hadoop
Migrate the HDP Configurations
Create Local Directories
Start Hadoop Core
Verify HDFS Filesystem Health
Configure YARN and MapReduce
Start YARN/MapReduce Services
Run Hadoop Smoke Tests
Configure and Start Apache HBase
Configure and Start Apache Hive and Apache HCatalog
Configure and Start Apache Oozie
Configure and Start Apache WebHCat (Templeton)
Configure and Start Apache Pig
Configure and Start Apache Sqoop
Configure, Start, and Validate Apache Flume
Configure, Start, and Validate Apache Mahout
Configure and Start Hue
Finalize the Upgrade
Install New HDP 2.5 Services

List of Tables

1.1. Hive Metastore Database Backup and Restore
1.2. Oozie Metastore Database Backup and Restore
1.3. Hue Database Backup and Restore
1.4. Ranger_Admin install.properties names and values
1.5. Ranger_Usersync install.properties names and values
1.6. KMS install.properties names and values
1.7. Properties for the /etc/hadoo/conf/core-site.xml file
1.8. Ranger-admin install.properties
1.9.
1.10. Ranger-tagsync install.properties and values
1.11. Ranger-kms install.properties and values
1.12. hdfs-site.xml Property Names and Values
1.13. Ranger Tagsync_install.properties and values
2.1. Hive Metastore Database Backup and Restore
2.2. Oozie Metastore Database Backup and Restore
2.3. Hue Database Backup and Restore
2.4. Ranger_Admin install.properties names and values
2.5. Ranger_Usersync install.properties names and values
2.6. KMS install.properties names and values
2.7. Properties for the /etc/hadoop/conf/core-site.xml file
2.8. Ranger-admin install.properties
2.9.
2.10. Ranger-tagsync install.properties and values
2.11. Ranger-kms install.properties and values
2.12. hdfs-site.xml Property Names and Values
2.13. Ranger Tagsync_install.properties and values
3.1. Hive Metastore Database Backup and Restore
3.2. Oozie Metastore Database Backup and Restore
3.3. Hue Database Backup and Restore
3.4. Ranger_Admin install.properties names and values
3.5. Ranger_Usersync install.properties names and values
3.6. KMS install.properties names and values
3.7. Properties for the /etc/hadoo/conf/core-site.xml file
3.8. Ranger-admin install.properties
3.9.
3.10. Ranger-tagsync install.properties and values
3.11. Ranger-kms install.properties and values
3.12. hdfs-site.xml Property Names and Values
3.13. Ranger Tagsync_install.properties and values
4.1. Hive Metastore Database Backup and Restore
4.2. Oozie Metastore Database Backup and Restore
4.3. Hue Database Backup and Restore
5.1. Hive Metastore Database Backup and Restore
5.2. Oozie Metastore Database Backup and Restore
5.3. Hue Database Backup and Restore
6.1. Hive Metastore Database Backup and Restore
6.2. Oozie Metastore Database Backup and Restore
6.3. Hue Database Backup and Restore
6.4. HDP 1.3.2 Hadoop Core Site (core-site.xml)
6.5. HDP 1.3.2 Hadoop Core Site (hdfs-site.xml)
6.6. HDP 1.3.2 Configs now in Capacity Scheduler for HDP 2.x (mapred-site.xml)
6.7. HDP 1.3.2 Configs now in capacity scheduler for HDP 2.x (capacity-scheduler.xml)
6.8. HDP 1.3.2 Configs and HDP 2.x for hadoop-env.sh