![]() | Except where otherwise noted, this document is licensed under Creative Commons Attribution ShareAlike 3.0 License |
Hortonworks Data Platform (HDP) and any of its components are not anticipated to be combined with any hardware, software or data, except as expressly recommended in this documentation.
Unlike other providers of platforms built using Apache Hadoop, Hortonworks contributes 100% of our code back to the Apache Software Foundation. The Hortonworks Data Platform is Apache-licensed and completely open source. We sell only expert technical support, training and partner-enablement services. All of our technology is, and will remain free and open source.
Please visit the Hortonworks Data Platform page for more information on Hortonworks technology. For more information on Hortonworks services, please visit either the Support or Training page. Feel free to Contact Us directly to discuss your specific needs.
Contents
- 1. Getting Ready to Install
- 2. Installing HDFS and MapReduce
- 3. Setting Up the Hadoop Configuration
- 4. Validating the Core Hadoop Installation
- 5. Installing Apache Pig
- 6. Installing Apache Hive and Apache HCatalog
- 7. Installing Apache WebHCat
- 8. Installing Apache HBase and Apache ZooKeeper
- 9. Installing Hue
- 10. Installing Apache Oozie
- 11. Installing Apache Sqoop
- 12. Installing and Configuring Apache Flume in HDP
- 13. Installing Ganglia
- 14. Installing Nagios
- 15. Setting Up Security for Manual Installs
- 16. Uninstalling HDP
- 17. Appendix: Tarballs
List of Tables
- 1.1. Typical Service Users and Groups
- 1.2. Define Users and Groups for Systems
- 1.3. Define Directories for Core Hadoop
- 1.4. Define Directories for Ecosystem Components
- 9.1. Dependencies on HDP components
- 14.1. Host Group Parameters
- 14.2. Core and Monitoring Hosts
- 14.3. Ecosystem Hosts
- 15.1. Kerberos terminology
- 15.2. Service Principal Names
- 15.3. Service Keytab File Names
- 17.1. RHEL/CentOS 5
- 17.2. RHEL/CentOS 6
- 17.3. SLES 11