Using the Connector with Apache Phoenix
If you use a Spark-HBase connector in an environment that uses Apache Phoenix as a SQL
skin, be aware that both connectors use only HBase .jar files by default. If you want to
submit jobs on an HBase cluster with Phoenix enabled, you must include --jars
phoenix-server.jar
in your spark-submit
command. For
example:
./bin/spark-submit --class your.application.class \ --master yarn-client \ --num-executors 2 \ --driver-memory 512m \ --executor-memory 512m --executor-cores 1 \ --packages com.hortonworks:shc:1.0.0-1.6-s_2.10 \ --repositories http://repo.hortonworks.com/content/groups/ public/ \ --jars /usr/hdp/current/phoenix-client/phoenix-server.jar \ --files /etc/hbase/conf/hbase-site.xml /To/your/application/jar