Learn how to use Granite Code models on Replicate for inference calls.
This guide demonstrates using inference calls against a model hosted on
Replicate. This guide will demonstrate a basic inference call using the
replicate package as well as via LangChain. In both cases, you will provide a
Replicate API Token.To see how you can use Ollama to host models locally instead, see how to build a
VS Code Assistant with Granite.
Granite Code on Replicate
Jupyter notebook showing how to use Granite Code on Replicate
⌘I
Assistant
Responses are generated using AI and may contain mistakes.