Skip to main content
Retrieval-Augmented Generation (RAG) enhances the performance of foundation models by grounding their calculations in external knowledge sources, resulting in more accurate and up-to-date responses. RAG is a versatile pattern that combines the power of retrieval systems with generative models to unlock use cases that are less accurate when using standalone foundation models. This tutorial demonstrates RAG leveraging LangChain and Granite. LangChain is a framework for building LLM-powered applications by chaining interoperable components. LangChain provides a standard interface for models, embeddings, vector stores, and more. You will need a Replicate API token and a Hugging Face token to run this recipe in Colab. Instructions for obtaining these credentials can be found here.

Get started

Explore sample code in a GitHub repo
https://mintcdn.com/ibmgranite/m3dncz2KrKeb3pcV/granite/docs/images/icons8-google-colab.svg?fit=max&auto=format&n=m3dncz2KrKeb3pcV&q=85&s=fb39ef667c012d0fcef53599b6c5c0fd

Try it out

Execute sample code in Colab