Skip to main content

Chat Models

Chat Models are a core component of LangChain.

A chat model is a language model that uses chat messages as inputs and returns chat messages as outputs (as opposed to using plain text).

LangChain has integrations with many model providers (OpenAI, Cohere, Hugging Face, etc.) and exposes a standard interface to interact with all of these models.

LangChain allows you to use models in sync, async, batching and streaming modes and provides other features (e.g., caching) and more.

Quick Start

Check out this quick start to get an overview of working with ChatModels, including all the different methods they expose


For a full list of all LLM integrations that LangChain provides, please go to the Integrations page

How-To Guides

We have several how-to guides for more advanced usage of LLMs. This includes:

Help us out by providing feedback on this documentation page: