Running a Python-based job
You can run a Python script to execute a spark-submit or pyspark command.
/* spark-demo.py */ from pyspark import SparkContext sc = SparkContext("local", "first app") from pyspark.sql import HiveContext hive_context = HiveContext(sc) hive_context.sql("drop table default.sales_spark_2_copy") hive_context.sql("CREATE TABLE IF NOT EXISTS default.sales_spark_2_copy as select * from default.sales_spark_2") hive_context.sql("show tables").show() hive_context.sql("select * from default.sales_spark_2_copy limit 10").show() hive_context.sql("select count(*) from default.sales_spark_2_copy").show()