Start Spark
=======
sbin/start-master.sh
bin/start-master.sh --> bin/spark-daemon.sh --> bin/spark-class org.apache.spark.deploy.master.Master
bin/spark-shell
bin/spark-shell --> bin/spark-submit --class org.apache.spark.repl.Main --name "Spark shell" --> bin/spark-class org.apache.spark.deploy.SparkSubmit --class org.apache.spark.repl.Main --name "Spark shell" --> java org.apache.spark.launcher.Main org.apache.spark.deploy.SparkSubmit --class org.apache.spark.repl.Main --name "Spark shell"
bin/spark-submit
bin/spark-submit --> bin/spark-class org.apache.spark.deploy.SparkSubmit --> java org.apache.spark.launcher.Main
SparkClassCommandBuilder
SparkSubmitCommandBuilder
org.apache.spark.launcher.Main build the commands to run and then exit the process. spark-class will use the commands to launch another process using the commands
Tips for how to get the launcher details
export SPARK_PRINT_LAUNCH_COMMAND=true
bin/spark-shell --verbose
Last updated
Was this helpful?