Hortonworks Docs
»
»
Also available as:
Developing Apache Spark Applications
Introduction
Using the Spark DataFrame API
Using Spark SQL
Access Spark SQL through the Spark shell
Access Spark SQL through JDBC or ODBC: prerequisites
Access Spark SQL through JDBC
Accessing Spark SQL through ODBC
Using the Hive Warehouse Connector with Spark
Calling Hive User-Defined Functions
Using Spark Streaming
Building and Running a Secure Spark Streaming Job
Running Spark Streaming Jobs on a Kerberos-Enabled Cluster
Sample pom.xml File for Spark Streaming with Kafka
HBase Data on Spark with Connectors
Selecting a Connector
Using the Connector with Apache Phoenix
Accessing HDFS Files from Spark
Accessing ORC Data in Hive Tables
Access ORC files from Spark
Predicate Push-Down Optimization
Load ORC Data into DataFrames Using Predicate Push-Down
Optimize Queries Using Partition Pruning
Enable Vectorized Query Execution
Read Hive ORC Tables
Additional Resources
Using Custom Libraries with Spark
Using Spark from R: SparkR
© 2012-2019, Hortonworks, Inc.
Document licensed under the
Creative Commons Attribution ShareAlike 4.0 License
.
Hortonworks.com
|
Documentation
|
Support
|
Community