Run sample spark-submit command inside the docker container

After you activate the docker image, you can run your spark-submit commands on CDE without completely rewriting your existing spark-on-yarn command lines inside the docker container.

  • Sample spark-submit commands you can run on the CDE workloads.
    $ spark-submit --name pt_rpt_streams5 --master=yarn --deploy-mode=cluster --driver-memory 4G --executor-memory 4G --executor-cores 3 --num-executors 4 --conf spark.yarn.queue=hr --conf "spark.executor.extraJavaOptions=-Djava.security.auth.login.config=/home/hdpsparkprd/spark-hdpsparkprd-keytab-jaas.conf -Djava.security.krb5.conf=/etc/krb5.conf -Djavax.security.auth.useSubjectCredsOnly=true" --conf "spark.driver.extraJavaOptions=-Djava.security.auth.login.config=/home/hdpsparkprd/spark-hdpsparkprd-keytab-jaas.conf -Djava.security.krb5.conf=/etc/krb5.conf -Djavax.security.auth.useSubjectCredsOnly=true" --conf "spark.io.compression.codec=org.apache.spark.io.LZ4CompressionCodec" --class org.apache.spark.examples.SparkPi http://qe-repo.s3.amazonaws.com/dex/app-jar/spark-examples_2.11-2.4.4.jar 22