Installing HDP Using Shell Scripts (gsInstaller)
Copyright © 2012, 2013 Hortonworks, Inc.
Except where otherwise noted, this document is licensed under Creative Commons Attribution ShareAlike 3.0 License |
2013-04-02
Abstract
The Hortonworks Data Platform, powered by Apache Hadoop, is a massively scalable and 100% open source platform for storing, processing and analyzing large volumes of data. It is designed to deal with data from many sources and formats in a very quick, easy and cost-effective manner. The Hortonworks Data Platform consists of the essential set of Apache Hadoop projects including MapReduce, Hadoop Distributed File System (HDFS), HCatalog, Pig, Hive, HBase, Zookeeper and Ambari. Hortonworks is the major contributor of code and patches to many of these projects. These projects have been integrated and tested as part of the Hortonworks Data Platform release process and installation and configuration tools have also been included.
Unlike other providers of platforms built using Apache Hadoop, Hortonworks contributes 100% of our code back to the Apache Software Foundation. The Hortonworks Data Platform is Apache-licensed and completely open source. We sell only expert technical support, training and partner-enablement services. All of our technology is, and will remain free and open source.
Please visit the Hortonworks Data Platform page for more information on Hortonworks technology. For more information on Hortonworks services, please visit either the Support or Training page. Feel free to Contact Us directly to discuss your specific needs.
Contents
- 1. Getting Ready to Install
- 2. Deploying Non Secure Hadoop Cluster
- 3. Deploying Secure Hadoop Cluster
- 4. Troubleshooting gsInstaller Deployments
- 5. Reference
List of Tables
- 1.1. Properties to create database users
- 2.1. gsInstaller Configuration Properties
- 2.2. gsInstaller Monitoring Properties
- 3.1. gsInstaller Configuration Properties
- 3.2. gsInstaller Monitoring Properties
- 3.3. Secure deployment - Mandatory naming conventions for principals
- 3.4. Secure deployment - Mandatory naming conventions for keytab files
- 3.5. Secure deployment - Mandatory naming conventions for HDFS service and Smoke test user's keytabs and host principals
- 3.6. Secure deployment - Configuring HDFS user keytab file for secure Hadoop cluster using Option II
- 5.1. Generic Properties
- 5.2. Hadoop Core Properties
- 5.3. Service User Properties
- 5.4. Data and Log Directory Configurations
- 5.5. HDP Stack Components Properties
- 5.6. Secure Hadoop Deployment Properties
- 5.7. Monitoring components (Ganglia and Nagios) Properties
- 5.8. Properties for Apache Oozie
- 5.9. Properties for Sqoop
- 5.10. Properties for Flume
- 5.11. Properties for Mahout
- 5.12. Hadoop-HDFS Properties
- 5.13. Hadoop-MapReduce Properties
- 5.14. Hadoop-ZooKeeper Properties
- 5.15. Secure deployment - Mandatory naming conventions for principals
- 5.16. Secure deployment - Mandatory naming conventions for principals
- 5.17. Secure deployment - Mandatory naming conventions for keytab files
- 5.18. Secure deployment - Mandatory naming conventions for HDFS and Smoke test users' keytab files