I have a spark standalone cluster with no other job scheduler installed. I wonder if spark-submit
can be used as a job scheduler for both spark and non-spark jobs (e.g. a scala jar not written for Spark and not using RDD)?
Based on my testing, spark-submit
be used to submit non-Spark jobs and the jobs run successfully. But here are my questions:
- Are the the following options still meaningful?
--driver-cores --driver-memory --executor-memory --total-executor-cores
- If No to 1, does it mean
spark-submit
can maintain a queue of spark and non-spark jobs using FIFO but it does not manage the resource of the non-spark job? - If 2 is true, should I use another scheduler, e.g. SGE or LSF, to submit non-spark jobs?
Thanks!
I figured out after many testings. Yes, spark standalone can be a job scheduler for both spark and non-spark jobs.
spark-submit
only creates drivers, no executors.spark-submit
command, are met.