Installing watsonx Orchestrate

An instance administrator can install watsonx Orchestrate on IBM® Software Hub Version 5.1.

Who needs to complete this task?

Instance administrator To install watsonx Orchestrate, you must be an instance administrator. An instance administrator has permission to install software in the following projects:

The operators project for the instance

The operators for this instance of watsonx Orchestrate are installed in the operators project.

In the installation commands, the ${PROJECT_CPD_INST_OPERATORS} environment variable refers to the operators project.

The operands project for the instance

The custom resources for the control plane and watsonx Orchestrate are installed in the operands project.

In the installation commands, the ${PROJECT_CPD_INST_OPERANDS} environment variable refers to the operands project.

When do you need to complete this task?

Review the following options to determine whether you need to complete this task:

  • If you want to install multiple services at the same time, follow the process in Running a batch installation of solutions and services instead.
  • If you didn't install watsonx Orchestrate as part of a batch installation, complete this task to add watsonx Orchestrate to your environment.

    Repeat as needed If you are responsible for multiple instances of IBM Software Hub, you can repeat this task to install more instances of watsonx Orchestrate on the cluster.

Information you need to complete this task

Review the following information before you install watsonx Orchestrate:

Version requirements

All of the components that are associated with an instance of IBM Software Hub must be installed at the same release. For example, if the IBM Software Hub control plane is installed at Version 5.1.3, you must install watsonx Orchestrate at Version 5.1.3.

Environment variables

The commands in this task use environment variables so that you can run the commands exactly as written.

  • If you don't have the script that defines the environment variables, see Setting up installation environment variables.
  • To use the environment variables from the script, you must source the environment variables before you run the commands in this task. For example, run:
    source ./cpd_vars.sh
Security context constraint

watsonx Orchestrate works with the default Red Hat® OpenShift® Container Platform security context constraint, restricted-v2.

Storage requirements
You must specify storage classes when you install watsonx Orchestrate. The following storage classes are recommended. However, if you don't use these storage classes on your cluster, ensure that you specify a storage class with an equivalent definition.
Storage Notes Storage classes
OpenShift Data Foundation When you install the service, specify file storage and block storage.
  • File storage: ocs-storagecluster-cephfs
  • Block storage: ocs-storagecluster-ceph-rbd
IBM Fusion Data Foundation When you install the service, specify file storage and block storage.
  • File storage: ocs-storagecluster-cephfs
  • Block storage: ocs-storagecluster-ceph-rbd
IBM Fusion Global Data Platform When you install the service, specify the same storage class for both file storage and block storage.
  • File storage:

    Either of the following storage classes:

    • ibm-spectrum-scale-sc
    • ibm-storage-fusion-cp-sc
  • Block storage:

    Either of the following storage classes:

    • ibm-spectrum-scale-sc
    • ibm-storage-fusion-cp-sc
IBM Storage Scale Container Native Not supported. Not applicable.
Portworx When you install the service, the --storage_vendor=portworx option ensures that the service uses the correct storage classes.
  • File storage: portworx-shared-gp3
  • Block storage: portworx-db-gp3-sc
NFS Not supported. Not applicable.
Amazon Elastic storage Not supported. Not applicable.
NetApp Trident When you install the service, specify the same storage class for both file storage and block storage.
  • File storage: ontap-nas
  • Block storage: ontap-nas
Nutanix Not supported. Not applicable.

Before you begin

This task assumes that the following prerequisites are met:

Prerequisite Where to find more information
The cluster meets the minimum requirements for installing watsonx Orchestrate. If this task is not complete, see System requirements.
The workstation from which you will run the installation is set up as a client workstation and includes the following command-line interfaces:
  • IBM Software Hub CLI: cpd-cli
  • OpenShift CLI: oc
If this task is not complete, see Setting up a client workstation.
The IBM Software Hub control plane is installed. If this task is not complete, see Installing an instance of IBM Software Hub.
For environments that use a private container registry, such as air-gapped environments, the watsonx Orchestrate software images are mirrored to the private container registry. If this task is not complete, see Mirroring images to a private container registry.
For environments that use a private container registry, such as air-gapped environments, the cpd-cli is configured to pull the olm-utils-v3 image from the private container registry. If this task is not complete, see Pulling the olm-utils-v3 image from the private container registry.
The operators that are required to use GPUs are installed. If this task is not complete, see Installing operators for services that require GPUs.
Red Hat OpenShift AI is installed. If this task is not complete, see Installing Red Hat OpenShift AI.
Multicloud Object Gateway is installed and configured. If this task is not complete, see Installing Multicloud Object Gateway.
The secrets that enable watsonx Orchestrate to connect to Multicloud Object Gateway exist. If this task is not complete, see Creating secrets for services that use Multicloud Object Gateway.
Red Hat OpenShift Serverless Knative Eventing is installed and configured. If this task is not complete, see Installing Red Hat OpenShift Serverless Knative Eventing.
IBM App Connect in containers is installed and configured.

5.1.0 5.1.1 5.1.2 This software is required only if you are installing IBM Software Hub Version 5.1.0, Version 5.1.1, or Version 5.1.2.

If this task is not complete, see Installing App Connect in containers.

Preparing to install

Before you install watsonx Orchestrate, do the following:
  • Verify that Identity and Access Management (IAM) Service is enabled. By default, the Identity and Access Management (IAM) Service is enabled with the IBM Software Hub Version 5.1.x. For more information, see Integrating with the Identity Management Service.
  • watsonx Orchestrate and standalone watsonx Assistant cannot be installed on the same namespace. If you want to install both separately, then you can install on different namespaces.

Procedure

Complete the following tasks to install watsonx Orchestrate:

  1. Specifying installation options
  2. Installing the service
  3. Validating the installation
  4. Optional: Enabling the models and GPU features
  5. What to do next

Specifying installation options

If you plan to install watsonx™ Orchestrate, you can specify the following installation option in a file named install-options.yml in the cpd-cli work directory (For example: cpd-cli-workspace/olm-utils-workspace/work).

The parameter is optional. If you do not set the installation parameter, the default values are used. Uncomment the parameter if you want to override the default behavior.

The sample YAML content uses the default values.

################################################################################
# watsonx Orchestrate parameters
################################################################################
#watson_orchestrate_watsonx_ai_type: false
#watson_orchestrate_syom_models: [] 
#watson_orchestrate_ootb_models: []
Property Description
watson_orchestrate_watsonx_ai_type 5.1.0 5.1.1 5.1.2 This parameter applies only to IBM Software Hub Version 5.1.0, Version 5.1.1, and Version 5.1.2.

Specify whether you want to install Inference foundation models (watsonx_ai_ifm) to enable the following features, which require GPUs:

For more information about supported GPUs, see the Hardware requirements for watsonx Orchestrate.

Default value
false

If you omit this option, the default value is used.

Valid values
false
Do not install Inference foundation models (watsonx_ai_ifm). The features that require GPUs will not be available
true
Install Inference foundation models (watsonx_ai_ifm) to enable features that require GPUs.
watson_orchestrate_syom_models 5.1.3 and later This option is available starting in IBM Software Hub Version 5.1.3.
Specify whether you want to use a specialized model that is specifically tuned for use with watsonx Orchestrate for:
Important: The following model will be automatically installed if you do not specify a value for watson_orchestrate_syom_models or watson_orchestrate_ootb_models:
  • granite-3-8b-instruct
Default value
[]
Valid values
[]
Do not install a specialized model.
ibm-granite-8b-unified-api-model-v2
Install the ibm-granite-8b-unified-api-model-v2 specialized model. This model enables assistants to:
  • Rewrite user questions to an understood format for conversational search
  • Gather information to fill in variables in a conversational skill
Including this parameter
Ensure that you uncomment the following line and specify the model name as a list item on a new line:
watson_orchestrate_syom_models:
  - ibm-granite-8b-unified-api-model-v2
watson_orchestrate_ootb_models 5.1.3 and later This option is available starting in IBM Software Hub Version 5.1.3.
Specify whether you want to use a general model for:
Important: The following model will be automatically installed if you do not specify a value for watson_orchestrate_syom_models or watson_orchestrate_ootb_models:
  • granite-3-8b-instruct
Default value
[]
Valid values
[]
Do not install a general model.
granite-3-8b-instruct
Install the granite-3-8b-instruct general model. This model enables assistants to:
  • Answer conversational search questions
Important: If you use a private container registry, you must explicitly mirror the granite-3-8b-instruct image to the private container registry.
llama-3-1-70b-instruct
Install the llama-3-1-70b-instruct general model. This model enables assistants to:
  • Rewrite user questions to an understood format for conversational search
  • Answer conversational search questions.
  • Gather information to fill in variables in a conversational skill
Important: If you use a private container registry, you must explicitly mirror the llama-3-1-70b-instruct image to the private container registry.
Including this parameter
Ensure that you uncomment the following line and specify the model name as a list item on a new line.
Install the granite-3-8b-instruct model
watson_orchestrate_ootb_models:
  - granite-3-8b-instruct
Install the llama-3-1-70b-instruct model
watson_orchestrate_ootb_models:
  - llama-3-1-70b-instruct
Install both models
watson_orchestrate_ootb_models:
  - granite-3-8b-instruct
  - llama-3-1-70b-instruct

Installing the service

To install watsonx Orchestrate:

  1. Log the cpd-cli in to the Red Hat OpenShift Container Platform cluster:
    ${CPDM_OC_LOGIN}
    Remember: CPDM_OC_LOGIN is an alias for the cpd-cli manage login-to-ocp command.
  2. Important:
    • Do this step only if you are in Version 5.1.0, 5.1.1, or 5.1.2.
    • Before you proceed to deploy the App Connect instance and create App Connect secret, you must create dedicated namespaces for Cloud Pak for Data operators and Cloud Pak for Data instance.
    Create an instance of App Connect for watsonx Orchestrate:
    cpd-cli manage setup-appconnect \
    --appconnect_ns=${PROJECT_IBM_APP_CONNECT} \
    --cpd_instance_ns=${PROJECT_CPD_INST_OPERANDS} \
    --release=${VERSION} \
    --components=watsonx_orchestrate \
    --file_storage_class=${STG_CLASS_FILE}
    For more information, see Manage setup App Connect.
  3. Run the following command to create the required OLM objects for watsonx Orchestrate in the operators project for the instance:
    cpd-cli manage apply-olm \
    --release=${VERSION} \
    --cpd_operator_ns=${PROJECT_CPD_INST_OPERATORS} \
    --components=watsonx_orchestrate
    Wait for the cpd-cli to return the following message before you proceed to the next step:
    [SUCCESS]... The apply-olm command ran successfully

    If the apply-olm fails, see Troubleshooting the apply-olm command during installation or upgrade.

  4. Create the custom resource for watsonx Orchestrate.

    The command that you run depends on the storage on your cluster.


    Red Hat OpenShift Data Foundation storage

    Run the appropriate command to create the custom resource.

    Default installation (without installation options)
    cpd-cli manage apply-cr \
    --components=watsonx_orchestrate \
    --release=${VERSION} \
    --cpd_instance_ns=${PROJECT_CPD_INST_OPERANDS} \
    --block_storage_class=${STG_CLASS_BLOCK} \
    --file_storage_class=${STG_CLASS_FILE} \
    --license_acceptance=true
    Custom installation (with installation options)
    cpd-cli manage apply-cr \
    --components=watsonx_orchestrate \
    --release=${VERSION} \
    --cpd_instance_ns=${PROJECT_CPD_INST_OPERANDS} \
    --block_storage_class=${STG_CLASS_BLOCK} \
    --file_storage_class=${STG_CLASS_FILE} \
    --param-file=/tmp/work/install-options.yml \
    --license_acceptance=true

    IBM Fusion Data Foundation storage

    Run the appropriate command to create the custom resource.

    Default installation (without installation options)
    cpd-cli manage apply-cr \
    --components=watsonx_orchestrate \
    --release=${VERSION} \
    --cpd_instance_ns=${PROJECT_CPD_INST_OPERANDS} \
    --block_storage_class=${STG_CLASS_BLOCK} \
    --file_storage_class=${STG_CLASS_FILE} \
    --license_acceptance=true
    Custom installation (with installation options)
    cpd-cli manage apply-cr \
    --components=watsonx_orchestrate \
    --release=${VERSION} \
    --cpd_instance_ns=${PROJECT_CPD_INST_OPERANDS} \
    --block_storage_class=${STG_CLASS_BLOCK} \
    --file_storage_class=${STG_CLASS_FILE} \
    --param-file=/tmp/work/install-options.yml \
    --license_acceptance=true

    Portworx storage

    Run the appropriate command to create the custom resource.

    Default installation (without installation options)
    cpd-cli manage apply-cr \
    --components=watsonx_orchestrate \
    --release=${VERSION} \
    --cpd_instance_ns=${PROJECT_CPD_INST_OPERANDS} \
    --storage_vendor=portworx \
    --license_acceptance=true
    Custom installation (with installation options)
    cpd-cli manage apply-cr \
    --components=watsonx_orchestrate \
    --release=${VERSION} \
    --cpd_instance_ns=${PROJECT_CPD_INST_OPERANDS} \
    --storage_vendor=portworx \
    --param-file=/tmp/work/install-options.yml \
    --license_acceptance=true

    NetApp Trident
    Remember: When you use NetApp Trident storage, both ${STG_CLASS_BLOCK} and ${STG_CLASS_FILE} point to the same storage class, typically ontap-nas.

    Run the appropriate command to create the custom resource.

    Default installation (without installation options)
    cpd-cli manage apply-cr \
    --components=watsonx_orchestrate \
    --release=${VERSION} \
    --cpd_instance_ns=${PROJECT_CPD_INST_OPERANDS} \
    --block_storage_class=${STG_CLASS_BLOCK} \
    --file_storage_class=${STG_CLASS_FILE} \
    --license_acceptance=true
    Custom installation (with installation options)
    cpd-cli manage apply-cr \
    --components=watsonx_orchestrate \
    --release=${VERSION} \
    --cpd_instance_ns=${PROJECT_CPD_INST_OPERANDS} \
    --block_storage_class=${STG_CLASS_BLOCK} \
    --file_storage_class=${STG_CLASS_FILE} \
    --param-file=/tmp/work/install-options.yml \
    --license_acceptance=true

Validating the installation

watsonx Orchestrate is installed when the apply-cr command returns:
[SUCCESS]... The apply-cr command ran successfully

If you want to confirm that the custom resource status is Completed, you can run the cpd-cli manage get-cr-status command:

cpd-cli manage get-cr-status \
--cpd_instance_ns=${PROJECT_CPD_INST_OPERANDS} \
--components=watsonx_orchestrate

Optional: Enabling the models and GPU features

Prerequisites for enabling the models in Version 5.1.2
The foundation models must be enabled to use the AI agent and other features in watsonx Orchestrate.
Supported foundation models for GPU features
GPU features support the following foundation models:
Specialized model in watsonx Orchestrate Out of the Box models
ibm-granite-8b-unified-api-model-v2
  • granite-3-8b-instruct
  • llama-3-1-70b-instruct
System requirements for models
The following table lists the recommended GPU requirements that you must configure on a single OpenShift worker node for using the specialized model:
For Specialized model in watsonx Orchestrate
Model name Description System requirements Supported GPU
Model name
ibm-granite-8b-unified-api-model-v2
Model ID
ibm-granite-8b-unified-api-model-v2
Granite models are used for a wide range of generative and nongenerative tasks with appropriate prompt engineering. They employ a GPT-style decoder-only architecture, with more innovations from IBM Research and the open community.
CPUs
10
Memory
64 GB RAM
Storage
45 GB
Configuration
You can use any of the following GPU types:
  • 1 NVIDIA A100
  • 1 NVIDIA H100
  • 1 NVIDIA L40S
For details about the Out of the Box models and their system requirements, see Foundation models.
Enabling the default models for using AI agent feature in Version 5.1.2
For using the AI agent feature in watsonx Orchestrate, you must enable the default models.
Choose one of the following options:
  • You can enable llama-3-1-70b-instruct as default to use AI agent features and other features in watsonx Orchestrate. This model requires a minimum of 4 GPUs.
  • You can enable two Granite models: ibm-granite-8b-unified-api-model-v2 and granite-3-8b-instruct. Enabling both the models require a minimum of 2 GPUs. In addition, you must update the models through an API to use the AI agent and other features. For more information, see Updating the Granite models.

To enable llama-3-1-70b-instruct as default, use the following command:

oc patch wo wo \
--namespace=${PROJECT_CPD_INST_OPERANDS} \
--type=merge \
--patch='{
  "spec": {
    "watsonAssistants": {
      "config": {
        "configOverrides": {
          "enabled_components": {
            "store": {
              "ifm": true
            }
          },
          "watsonx_enabled": true,
          "ifm": {
            "model_config": {
              "ootb": {
                "llama-3-1-70b-instruct": {}
              }
            }
          }
        }
      }
    }
  }
}'
To enable ibm-granite-8b-unified-api-model-v2 and granite-3-8b-instruct, use the following command:
oc patch wo wo \
--namespace=${PROJECT_CPD_INST_OPERANDS} \
--type=merge \
--patch='{
  "spec": {
    "watsonAssistants": {
      "config": {
        "configOverrides": {
          "enabled_components": {
            "store": {
              "ifm": true
            }
          },
          "watsonx_enabled": true,
          "ifm": {
            "model_config": {
              "ootb": {
                "granite-3-8b-instruct": {}
              },
              "syom": {
                "ibm-granite-8b-unified-api-model-v2": {}
              }
            }
          }
        }
      }
    }
  }
}'
Updating the Granite models in Version 5.1.2
Note: Before you start to update, you must create a Zen API key using Generating a Zen API key.
To use AI agent and other features, update the Granite models per the following steps:
  • Update the Base LLM model to granite-3-8b-instruct.
  • Update the model for slot-filling and routing features to granite-8b-unified-api-model-v2.
To update the models, run the following API request as a cURL command:
curl --request PATCH \
  --url https://{root_uri}/orchestrate/{namespace}/instances/{instanceid}/v1/orchestrate/default-agent-settings \
  --header 'Authorization: ZenApiKey {REPLACE_THIS_VALUE}' \
  --header 'Content-Type: application/json' \
  --header 'accept: application/json' \
  --data '{
	"settings": {
    "model_settings": {
        "base_llm": {
            "model_name": "ibm/granite-3-8b-instruct",
            "model_params": {
                "random_seed": 2024,
                "temperature": 0.7,
                "max_new_tokens": 2048,
                "min_new_tokens": 1,
                "decoding_method": "sample"
            },
            "guardrails": {
                "hap": {
                    "input": {
                        "enabled": true,
                        "threshold": 0.5
                    },
                    "output": {
                        "enabled": true,
                        "threshold": 0.5
                    }
                },
                "pii": {
                    "input": {
                        "enabled": true,
                        "threshold": 0.5
                    },
                    "output": {
                        "enabled": true,
                        "threshold": 0.5
                    },
                    "mask": {
                        "remove_entity_value": true
                    }
                }
            },
            "system_prompt": "You are Granite Chat, an AI language model developed by IBM. You are a cautious assistant. You carefully follow instructions. You are helpful and harmless and you follow ethical guidelines and promote positive behavior. You always respond to greetings (for example, hi, hello, g'\''day, morning, afternoon, evening, night, what'\''s up, nice to meet you, sup, etc) with \"Hello! I am watsonx Orchestrate, created by IBM. How can I help you today?\". Please do not say anything else and do not start a conversation.\n"
        },
        "embeddings": {
            "model_name": "ibm/slate-125m-english-rtrvr"
        }
    },
    "router_settings": {
        "router_type": "unified",
        "model_name": "ibm/granite-8b-unified-api-model-v2",
        "model_params": {
            "top_k": 50,
            "top_p": 1,
            "random_seed": 2024,
            "temperature": 0.7,
            "max_new_tokens": 200,
            "min_new_tokens": 0,
            "return_options": {
                "input_text": false,
                "token_ranks": false,
                "input_tokens": true,
                "top_n_tokens": 0,
                "token_logprobs": true,
                "generated_tokens": true
            },
            "decoding_method": "sample",
            "repetition_penalty": 1
        },
        "router_config": {
            "continue_journey_on_none": true,
            "confidence_threshold": 0.8,
            "confidence_method": "perplexity"
        }
    },
    "user_settings": {
        "confirm_routing": true,
        "clientside_shortlisting": false
    },
    "slot_filling_settings": {
        "slot_filler_type": "unified",
        "model_name": "ibm/granite-8b-unified-api-model-v2",
        "model_params": {
            "top_k": 50,
            "top_p": 1,
            "random_seed": 2024,
            "temperature": 0.7,
            "max_new_tokens": 200,
            "min_new_tokens": 0,
            "return_options": {
                "input_text": false,
                "token_ranks": false,
                "input_tokens": true,
                "top_n_tokens": 0,
                "token_logprobs": true,
                "generated_tokens": true
            },
            "decoding_method": "sample",
            "repetition_penalty": 1
        }
    }
	}
}'
Where
  • <root_uri> is the deployment URL.
  • <namespace> is the namespace of your instance. To retrieve the <namespace>, use the Get service instance details API.
  • <instanceid> is the unique identifier of your instance. It is the number sequence after orchestrate- in the instance details URL.
  • {REPLACE_THIS_VALUE} is the generated Zen API key.
Enabling the GPU feature for Version 5.1.0 and 5.1.1
watsonx Orchestrate uses the watsonx Assistant's GPU feature, which is disabled by default. GPU configuration is a prerequisite to use features such as Skill-based actions and Conversational search. To enable GPU features on watsonx Assistant through watsonx Orchestrate, run the following command.
oc patch wo wo \
--namespace=${PROJECT_CPD_INST_OPERANDS} \
--type=merge \
--patch='{"spec": {"watsonAssistants": {"config": {"configOverrides": {"enabled_components": {"store": {"ifm": true}}, "watsonx_enabled": true}}}}}'
For more information about supported GPUs, see Hardware requirements for watsonx Orchestrate.

What to do next

Important: To use the Assistant Builder features, users must have access to both the watsonx Orchestrate service instance and the wxo-assistant-ab watsonx Assistant service instance. For more information, see Giving users access to a watsonx Orchestrate instance.