Creating batch deployments in Watson Machine Learning

A batch deployment processes input data from a file, data connection, or connected data in a storage bucket, and writes the output to a selected destination.

Before you begin

  1. Save a model to a deployment space.
  2. Promote or add the input file for the batch deployment to the space. For details on promoting an asset to a space, refer to Deployment spaces.

Supported frameworks

Batch deployment is supported for these frameworks and asset types:

  • Decision Optimization
  • PMML
  • Python functions
  • PyTorch-Onnx
  • Tensorflow
  • Scikit-learn
  • Scripts (Python and R)
  • Spark MLlib
  • SPSS
  • XGBoost

Notes:

  • Batch deployments of Python functions and models based on the PMML framework can only be done programmatically.
  • Your list of deployment jobs can contain two types of jobs: WML deployment job and WML batch deployment.
  • When you create a batch deployment (through the UI or programmatically), an additional default deployment job is created. The type of this job is WML deployment job. This is a parent job that stores all deployment runs generated for that batch deployment that were triggered by the Watson Machine Learning API.
  • The standard WML batch deployment type job is created only when you create a deployment from the UI or by using the cpdctl tool. Is is not possible to create a WML batch deployment type job by using the API.
  • This is the list of limitations of a WML deployment job:
    • it cannot be edited
    • it cannot be deleted unless the associated batch deployment has been deleted
    • it doesn't allow for scheduling
    • it doesn't allow for customizing notifications
    • it doesn't allow for changing retention settings

For information on data sources that are used for scoring batch deployments, refer to Data sources for scoring batch deployments. For information on required input for scoring batch deployments, depending on model type, refer to Batch deployment input details by framework

Creating a batch deployment

To create a batch deployment:

  1. From the deployment space, click the name of the saved model that you want to deploy. The model detail page opens.

  2. Click New deployment.

  3. Choose Batch as the deployment type and enter a name and description for your deployment.

  4. Select hardware specification.

    Restriction:

    If you want to use a custom hardware specification, you must create and select the custom hardware specification by using the hardware_specifications.store function from the Python client or Environments API. You cannot create or select a custom hardware specification from the user interface in a deployment space.

  5. Click Create. When status changes to Deployed, deployment creation is complete.

Note: Additionally, you can create a batch deployment by using any of these interfaces:
  • Watson Studio user interface, from an Analytics deployment space
  • Watson Machine Learning Python Client
  • Watson Machine Learning REST APIs

Viewing deployment details

Click the name of a deployment to view the details.

View deployment details

You can view the configuration details such as hardware and software specifications. You can also get the deployment ID, which you can use in API calls from an endpoint. For details, refer to Looking up a deployment endpoint.

Learn more

Parent topic: Managing predictive deployments