HDP Data Services
Copyright © 2012-2015 Hortonworks, Inc.
Except where otherwise noted, this document is licensed under Creative Commons Attribution ShareAlike 3.0 License |
Hortonworks Data Platform (HDP) and any of its components are not anticipated to be combined with any hardware, software or data, except as expressly recommended in this documentation.
2014-12-02
Abstract
The Hortonworks Data Platform, powered by Apache Hadoop, is a massively scalable and 100% open source platform for storing, processing and analyzing large volumes of data. It is designed to deal with data from many sources and formats in a very quick, easy and cost-effective manner. The Hortonworks Data Platform consists of the essential set of Apache Hadoop projects including YARN, Hadoop Distributed File System (HDFS), HCatalog, Pig, Hive, HBase, ZooKeeper and Ambari. Hortonworks is the major contributor of code and patches to many of these projects. These projects have been integrated and tested as part of the Hortonworks Data Platform release process and installation and configuration tools have also been included.
Unlike other providers of platforms built using Apache Hadoop, Hortonworks contributes 100% of our code back to the Apache Software Foundation. The Hortonworks Data Platform is Apache-licensed and completely open source. We sell only expert technical support, training and partner-enablement services. All of our technology is, and will remain free and open source.
Please visit the Hortonworks Data Platform page for more information on Hortonworks technology. For more information on Hortonworks services, please visit either the Support or Training page. Feel free to Contact Us directly to discuss your specific needs.
Contents
- 1. Using Apache Hive
- 1. Hive Documentation
- 2. New Feature: Temporary Tables
- 3. New Feature: Cost-based SQL Optimization
- 4. New Feature: ORC Format Improvement
- 5. Streaming Data Ingestion
- 6. Query Vectorization
- 7. Comparing Beeline to the Hive CLI
- 8. Hive JDBC and ODBC Drivers
- 9. Configuring HiveServer2 for LDAP and for LDAP over SSL
- 10. Troubleshooting Hive
- 11. Hive JIRAs
- 2. SQL Compliance
- 1. New Feature: INSERT ... VALUES, UPDATE, and DELETE SQL Statements
- 2. Hive 0.13 Feature: SQL Standard-based Authorization with
GRANT
AndREVOKE
SQL Statements - 3. Hive 0.13 Feature: Transactions
- 4. Hive 0.13 Feature: Subqueries in
WHERE
Clauses - 5. Hive 0.13 Feature: Common Table Expressions
- 6. Hive 0.13 Feature: Quoted Identifiers in Column Names
- 7. Hive 0.13 Feature:
CHAR
Data Type Support
- 3. Running Pig with the Tez Execution Engine
- 4. Using HDP for Metadata Services (HCatalog)
- 5. Using Apache HBase
- 6. Using HDP for Workflow and Scheduling (Oozie)
- 7. Using Apache Sqoop
- 1. Apache Sqoop Connectors
- 2. Sqoop Import Table Commands
- 3. Netezza Connector
- 4. Sqoop-HCatalog Integration
- 5. Controlling Transaction Isolation
- 6. Automatic Table Creation
- 7. Delimited Text Formats and Field and Line Delimiter Characters
- 8. HCatalog Table Requirements
- 9. Support for Partitioning
- 10. Schema Mapping
- 11. Support for HCatalog Data Types
- 12. Providing Hive and HCatalog Libraries for the Sqoop Job
- 13. Examples
List of Tables
- 1.1. CBO Configuration Parameters
- 1.2. Beeline Modes of Operation
- 1.3. HiveServer2 Transport Modes
- 1.4. Authentication Schemes with TCP Transport Mode
- 2.1. Configuration Parameters for Standard SQL Authorization
- 2.2. HiveServer2 Command-Line Options
- 2.3. Hive Compaction Types
- 2.4. Hive Transaction Configuration Parameters
- 2.5. Trailing Whitespace Characters on Various Databases