Creating a detached deployment for an external prompt
Create a detached deployment for evaluating a prompt template for an externally-hosted large language model (LLM) without inferencing the model.
As part of your governance process, you can evaluate prompt templates for generative AI models that are not created or hosted by IBM. For example, you can evaluate a prompt template for a Google Vertex AI model, an Azure OpenAI model, or an AWS Bedrock model. For this type of evaluation, all inferencing is done on the remote model. Evaluations are run on the generated prompt output.
This type of asset is called a detached prompt template. It must be created programmatically. However, you can evaluate a detached prompt template in a deployment space by creating a detached deployment. Creating a detached deployment provides you with these benefits and capabilities:
- Evaluating a prompt template in a project or a space provides a richer experience for reviewing the results of evaluations.
- Use access control for projects and spaces to invite collaborators or restrict access, as needed.
- Track the results of your evaluations in factsheets in AI use cases as part of your governance solution.
Creating and deploying a detached prompt template
Refer to the following topics for detailed instructions:
- To learn how to create a detached prompt template for an externally-hosted LLM and save it as a project asset, see Evaluating detached prompt templates in projects.
- To learn how to deploy a detached prompt template and evaluate the deployment in a space, see Evaluating detached prompt templates in spaces.
- For sample notebooks that demonstrate how to create a prompt template for a remotely-hosted external LLMs, see Evaluating prompt templates for non-IBM foundation models.
Learn more
If you are tracking the detached deployment in an AI use case, details about the model and evaluation results are recorded in a factsheet. For more information, see Tracking prompt templates.
Parent topic: Deploying foundation model assets