About LSF with Apache Spark
The IBM® Spectrum LSF integration with Apache Spark provides connector scripts that allow users to submit Spark applications as regular LSF jobs.
Apache Spark ("Spark") is an in-memory cluster computing system for large-scale data processing. It is an evolution of Apache Hadoop ("Hadoop") that provides high-level APIs in Java, Scala and Python, and an optimized engine that supports general execution graphs. It also provides various high-level tools, including Spark SQL for structured data processing, Spark Streaming for stream processing, and Mlib for machine learning.
Spark applications require distributed computed nodes, large memory, a high speed network, and no file system dependencies, so Spark applications can run in a traditional HPC environment. Use the IBM Spectrum LSF integration with Apache Spark to take advantage of the comprehensive LSF scheduling policies to allocate resources for Spark applications. LSF tracks, monitors and controls the job execution.
To run your Spark application through LSF, submit it as an LSF job, and the scheduler allocates resources according to the job's resource requirements, while the blaunch command starts a standalone Spark cluster. After the job is complete, LSF shuts down the Spark cluster.