Setting Docker/cgroup container definitions for a Spark instance group
Optionally set up the Spark instance group to run within containers, either Docker or control groups (cgroups).
- You can only set Docker/cgroup container definitions for a Spark instance group with certain Spark versions. Spark versions not supported: 1.5.2.
- Based on your requirements, ensure that you meet the requirements to create a Spark instance group. See Prerequisites for a Spark instance group.
You can enable Spark drivers and executors in a Spark instance group to run within Docker or cgroup
containers. You can also run Spark services in Docker containers.
- With Docker, you enable the Spark drivers, executors, and services of a Spark instance group to run in separate Docker containers. A Docker container holds everything that a Spark instance group needs to run, including an operating system, user-added files, metadata, and other dependencies.
- With cgroups, you enable per-resource limits (such as CPU shares and memory) for Spark drivers and executors in a Spark instance group. You can also set resource limits for Docker containers through Spark parameters so that Docker containers run within those CPU and memory bounds.
The Spark instance group is set up to
run in Docker or cgroup containers.
- Optionally, to add any extra packages that the Spark instance group requires, see Adding dependent packages. To add data connectors to the Spark instance group, see Adding data connectors.
- Create and deploy the Spark instance group.
- Click Create and Deploy Instance Group to create the Spark instance group and deploy its packages simultaneously. In this case, the new instance group appears on the Spark Instance Groups page in the Ready state. Verify your deployment and then start the Spark instance group.
- Click Create Only to create the Spark instance group but manually deploy its packages later. In this case, the new instance group appears on the Spark Instance Groups page in the Registered state. When are you ready to deploy packages, deploy the Spark instance group and verify the deployment. Then, start the Spark instance group.