Skip to main content


Ollama allows you to run open-source large language models, such as LLaMA2, locally.

Ollama bundles model weights, configuration, and data into a single package, defined by a Modelfile. It optimizes setup and configuration details, including GPU usage. For a complete list of supported models and model variants, see the Ollama model library.

See this guide for more details on how to use Ollama with LangChain.

Installation and Setup​

Follow these instructions to set up and run a local Ollama instance.


from langchain_community.llms import Ollama
API Reference:Ollama

See the notebook example here.

Chat Models​

Chat Ollama​

from langchain_community.chat_models import ChatOllama
API Reference:ChatOllama

See the notebook example here.

Ollama functions​

from langchain_experimental.llms.ollama_functions import OllamaFunctions
API Reference:OllamaFunctions

See the notebook example here.

Embedding models​

from langchain_community.embeddings import OllamaEmbeddings
API Reference:OllamaEmbeddings

See the notebook example here.

Was this page helpful?

You can also leave detailed feedback on GitHub.