Compute requirements for batch deployment jobs in Watson Machine Learning
The compute configuration for a batch deployment refers to the CPU and memory size that is allocated for a job. Learn about the predefined hardware specifications for your batch deployments.
Programmatically, this information must be specified in the hardware_spec
API parameter of the deployment payload or deployment jobs payload.
For an AutoAI model, the compute configuration must be specified in hybrid_pipeline_hardware_specs
instead of hardware_spec
.
Compute configurations must be a reference to a predefined hardware specification. You can specify a hardware specification by name or ID with hardware_spec
or hybrid_pipeline_hardware_specs
(for AutoAI). You can access
the list and details about the predefined hardware specifications through the watsonx.ai Python client or the Data and AI Common Core API.
Predefined hardware specifications
The following tables provide the predefined hardware specifications available by model type.
Watson Machine Learning models
Size | Hardware definition |
---|---|
XS | 1 CPU and 4 GB RAM |
S | 2 CPU and 8 GB RAM |
M | 4 CPU and 16 GB RAM |
ML | 4 CPU and 32 GB RAM |
L | 8 CPU and 32 GB RAM |
XL | 16 CPU and 64 GB RAM |
Decision Optimization
Size | Hardware definition |
---|---|
S | 2 CPU and 8 GB RAM |
M | 4 CPU and 16 GB RAM |
L | 8 CPU and 32 GB RAM |
XL | 16 CPU and 64 GB RAM |
Parent topic: Creating a batch deployment