Chat Models are a core component of LangChain.
A chat model is a language model that uses chat messages as inputs and returns chat messages as outputs (as opposed to using plain text).
LangChain has integrations with many model providers (OpenAI, Cohere, Hugging Face, etc.) and exposes a standard interface to interact with all of these models.
LangChain allows you to use models in sync, async, batching and streaming modes and provides other features (e.g., caching) and more.
Check out this quick start to get an overview of working with ChatModels, including all the different methods they expose
For a full list of all LLM integrations that LangChain provides, please go to the Integrations page
We have several how-to guides for more advanced usage of LLMs. This includes: