Evaluating AI models

If you are using the Watson OpenScale or watsonx.governance service, you can track and measures outcomes from your AI models to help ensure that they are compliant with business processes no matter where your models are built or running.

Service This service is not available by default. An administrator must install this service on the IBM Cloud Pak for Data platform, and you must be given access to the service. To determine whether the service is installed, open the Services catalog and check whether the service is enabled.

Enterprises use model evaluations as part of AI governance strategies to ensure that models in deployment environments meet established compliance standards regardless of the tools and frameworks that are used to build and run the models. This approach ensures that AI models are free from bias, can be easily explained and understood by business users, and are auditable in business transactions. You can evaluate machine learning models or evaluate prompt templates for foundation models, depending on which service that you use.

If you use Watsonx.governance, you can evaluate machine learning models and prompt templates in projects and spaces for foundation models.

Learn more

Parent topic: Governing AI assets