Compute requirements for batch deployment jobs in Watson Machine Learning

The compute configuration for a batch deployment refers to the CPU and memory size that is allocated for a job. Learn about the predefined hardware specifications for your batch deployments.

Programmatically, this information must be specified in the hardware_spec API parameter of the deployment payload or deployment jobs payload.

Note:

For an AutoAI model, the compute configuration must be specified in hybrid_pipeline_hardware_specs instead of hardware_spec.

Compute configurations must be a reference to a predefined hardware specification. You can specify a hardware specification by name or ID with hardware_spec or hybrid_pipeline_hardware_specs (for AutoAI). You can access the list and details about the predefined hardware specifications through the watsonx.ai Python client or the Data and AI Common Core API.

Predefined hardware specifications

The following tables provide the predefined hardware specifications available by model type.

Watson Machine Learning models

Predefined hardware specifications for Watson Machine Learning models
Size Hardware definition
XS 1 CPU and 4 GB RAM
S 2 CPU and 8 GB RAM
M 4 CPU and 16 GB RAM
ML 4 CPU and 32 GB RAM
L 8 CPU and 32 GB RAM
XL 16 CPU and 64 GB RAM

Decision Optimization

Predefined hardware specifications for Decision Optimization models
Size Hardware definition
S 2 CPU and 8 GB RAM
M 4 CPU and 16 GB RAM
L 8 CPU and 32 GB RAM
XL 16 CPU and 64 GB RAM

Parent topic: Creating a batch deployment