Installing Hadoop Using Apache Ambari
Copyright © 2012, 2014 Hortonworks, Inc.
Except where otherwise noted, this document is licensed under Creative Commons Attribution ShareAlike 3.0 License |
Hortonworks Data Platform (HDP) and any of its components are not anticipated to be combined with any hardware, software or data, except as expressly recommended in this documentation.
2014-02-10
Abstract
The Hortonworks Data Platform, powered by Apache Hadoop, is a massively scalable and 100% open source platform for storing, processing and analyzing large volumes of data. It is designed to deal with data from many sources and formats in a very quick, easy and cost-effective manner. The Hortonworks Data Platform consists of the essential set of Apache Hadoop projects including MapReduce, Hadoop Distributed File System (HDFS), HCatalog, Pig, Hive, HBase, Zookeeper and Ambari. Hortonworks is the major contributor of code and patches to many of these projects. These projects have been integrated and tested as part of the Hortonworks Data Platform release process and installation and configuration tools have also been included.
Unlike other providers of platforms built using Apache Hadoop, Hortonworks contributes 100% of our code back to the Apache Software Foundation. The Hortonworks Data Platform is Apache-licensed and completely open source. We sell only expert technical support, training and partner-enablement services. All of our technology is, and will remain free and open source.
Please visit the Hortonworks Data Platform page for more information on Hortonworks technology. For more information on Hortonworks services, please visit either the Support or Training page. Feel free to Contact Us directly to discuss your specific needs.
Contents
- I. Setting Up Ambari
- I.1. Getting Ready
- I.2. Running the Ambari Installer
- 1. Set Up the Bits
- 2. Set Up the Server
- 3. Optional: Set Up LDAP or Active Directory Authentication
- 4. Optional: Set Up Security for Ambari
- 5. Optional: Set Up Two-Way SSL Between Ambari Server and Ambari Agents
- 6. Optional: Change the Ambari Server Port
- 7. Optional: Configure Ambari Server for Internet Proxy
- 8. Start the Ambari Server
- II. Hadoop 2.x - Deploying, Configuring, and Upgrading Ambari
- II.3. Hadoop 2.x - Installing, Configuring, and Deploying the Cluster
- II.4. Hadoop 2.x - Troubleshooting Ambari Deployments
- 1. Review Ambari Log Files
- 2. Quick Checks
- 3. Specific Issues
- 3.1. Problem: Browser crashed before Install Wizard completed
- 3.2. Problem: Install Wizard reports that the cluster install has failed
- 3.3. Problem: “Unable to create new native thread” exceptions in HDFS DataNode logs or those of any system daemon
- 3.4. Problem: The “yum install ambari-server” Command Fails
- 3.5. Problem: HDFS Smoke Test Fails
- 3.6. Problem: The HCatalog Daemon Metastore Smoke Test Fails
- 3.7. Problem: MySQL and Nagios fail to install on RightScale CentOS 5 images on EC2
- 3.8. Problem: Trouble starting Ambari on system reboot
- 3.9. Problem: Metrics and Host information display incorrectly in Ambari Web
- 3.10. Problem: On SUSE 11 Ambari Agent crashes within the first 24 hours
- 3.11. Problem: Attempting to Start HBase REST server causes either REST server or Ambari Web to fail
- 3.12. Problem: Multiple Ambari Agent processes are running, causing re-register
- 3.13. Problem: Some graphs do not show a complete hour of data until the cluster has been running for an hour
- 3.14. Problem: After performing a cluster install the Nagios server is not started
- 3.15. Problem: A service with a customized service user is not appearing properly in Ambari Web
- 3.16. Problem: Updated configuration changes are not pushed to client/gateway nodes
- II.5. Appendix: Upgrading Ambari Server to 1.4.4
- II.6. Appendix: Upgrading the HDP Stack from 1.3.2 or later to 2.0.6
- II.7. Appendix: Hadoop 2.x - Configuring Ports
- II.8. Appendix: NameNode High Availabilty
- III. Hadoop 1.x - Deploying, Configuring, and Upgrading Ambari
- III.9. Hadoop 1.x - Installing, Configuring, and Deploying the Cluster
- III.10. Troubleshooting Ambari Deployments
- 1. Review Ambari Log Files
- 2. Quick Checks
- 3. Specific Issues
- 3.1. Problem: Browser crashed before Install Wizard completed
- 3.2. Problem: Install Wizard reports that the cluster install has failed
- 3.3. Problem: “Unable to create new native thread” exceptions in HDFS DataNode logs or those of any system daemon
- 3.4. Problem: The “yum install ambari-server” Command Fails
- 3.5. Problem: HDFS Smoke Test Fails
- 3.6. Problem: The HCatalog Daemon Metastore Smoke Test Fails
- 3.7. Problem: MySQL and Nagios fail to install on RightScale CentOS 5 images on EC2
- 3.8. Problem: Trouble starting Ambari on system reboot
- 3.9. Problem: Metrics and Host information display incorrectly in Ambari Web
- 3.10. Problem: On SUSE 11 Ambari Agent crashes within the first 24 hours
- 3.11. Problem: Attempting to Start HBase REST server causes either REST server or Ambari Web to fail
- 3.12. Problem: Multiple Ambari Agent processes are running, causing re-register
- 3.13. Problem: Some graphs do not show a complete hour of data until the cluster has been running for an hour
- 3.14. Problem: After performing a cluster install the Nagios server is not started
- 3.15. Problem: A service with a customized service user is not appearing properly in Ambari Web
- 3.16. Problem: Updated configuration changes are not pushed to client/gateway nodes
- III.11. Appendix: Upgrading Ambari Server to 1.4.4
- III.12. Appendix: Upgrading Ambari Server to 1.2.5
- III.13. Appendix: Upgrading the HDP Stack to 1.3.3
- III.14. Appendix: Configuring Ports
- III.15. Configuring RHEL HA for Hadoop 1.x
- IV. Additional Tasks with Ambari
- IV.16. Appendix: Installing Ambari Agents Manually
- IV.17. Appendix: Using Custom Hostnames
- IV.18. Appendix: Upgrading Operating Systems on an Ambari-based Hadoop Installation
- IV.19. Appendix: Moving the Ambari Server
- IV.20. Appendix: Using Non-Default Databases
- IV.21. Setting Up Kerberos for Use with Ambari
List of Tables
- I.2.1. Download the repo
- I.2.2. Ambari Server LDAP Properties
- II.3.1. Service Users
- II.3.2. Service Group
- II.3.3. HDFS Settings: Advanced
- II.3.4. MapReduce Settings: Advanced
- II.6.1. Key Properties to Check
- II.6.2. Properties to Modify
- II.7.1. HDFS Ports
- II.7.2. MapReduce Ports
- II.7.3. YARN Ports
- II.7.4. Hive Ports
- II.7.5. HBase Ports
- II.7.6. HBase Ports
- II.7.7. WebHCat Port
- II.7.8. Ganglia Ports
- II.7.9. MySQL Port
- II.7.10. Ambari Web
- II.7.11. Nagios
- II.8.1. Core-site.xml properties and values for NameNode HA on a cluster using Hue
- II.8.2. Set Environment Variables
- III.9.1. Service Users
- III.9.2. Service Group
- III.9.3. HDFS Settings: Advanced
- III.9.4. MapReduce Settings: Advanced
- III.14.1. HDFS Ports
- III.14.2. MapReduce Ports
- III.14.3. Hive Ports
- III.14.4. HBase Ports
- III.14.5. HBase Ports
- III.14.6. WebHCat Port
- III.14.7. Ganglia Ports
- III.14.8. MySQL Port
- III.14.9. Ambari Web
- III.14.10. Ambari Web
- III.15.1. Parameter Options for relocate_host_components.py
- IV.20.1. Hive Security Authorization Settings
- IV.21.1. Kerberos terminology
- IV.21.2. Service Principals
- IV.21.3. Ambari Principals
- IV.21.4. Service Keytab File Names
- IV.21.5. Kerberos terminology
- IV.21.6. Service Principals
- IV.21.7. Ambari Principals
- IV.21.8. Service Keytab File Names