Chat Models are a core component of LangChain.
A chat model is a language model that uses chat messages as inputs and returns chat messages as outputs (as opposed to using plain text).
LangChain has integrations with many model providers (OpenAI, Cohere, Hugging Face, etc.) and exposes a standard interface to interact with all of these models.
LangChain allows you to use models in sync, async, batching and streaming modes and provides other features (e.g., caching) and more.
Quick StartCheck out this quick start to get an overview of working with ChatModels, including all the different methods they expose
IntegrationsFor a full list of all LLM integrations that LangChain provides, please go to the Integrations page
How-To GuidesWe have several how-to guides for more advanced usage of LLMs. This includes:
RetroSearch is an open source project built by @garambo | Open a GitHub Issue
Search and Browse the WWW like it's 1997 | Search results from DuckDuckGo
HTML:
3.2
| Encoding:
UTF-8
| Version:
0.7.4