langchain-community: 0.3.7#
Main entrypoint into package.
adapters#
Classes
Chat. |
|
Chat completion. |
|
Chat completion chunk. |
|
Chat completions. |
|
Choice. |
|
Choice chunk. |
|
Completions. |
|
Allows a BaseModel to return its fields by string variable indexing. |
Functions
|
Async version of enumerate function. |
Convert a dictionary to a LangChain message. |
|
Convert a LangChain message to a dictionary. |
|
Convert messages to a list of lists of dictionaries for fine-tuning. |
|
|
Convert dictionaries representing OpenAI messages to LangChain format. |
agent_toolkits#
Classes
Toolkit for interacting with AINetwork Blockchain. |
|
Toolkit for interacting with Amadeus which offers APIs for travel. |
|
Toolkit for Azure AI Services. |
|
|
Toolkit for Azure Cognitive Services. |
|
Toolkit for interacting with an Apache Cassandra database. |
Clickup Toolkit. |
|
Toolkit for CogniSwitch. |
|
Toolkit with a list of Connery Actions as tools. |
|
|
Toolkit for interacting with local files. |
|
Toolkit for interacting with financialdatasets.ai. |
Schema for operations that require a branch name as input. |
|
Schema for operations that require a comment as input. |
|
Schema for operations that require a file path and content as input. |
|
Schema for operations that require a PR title and body as input. |
|
Schema for operations that require a username as input. |
|
Schema for operations that require a file path as input. |
|
Schema for operations that require a directory path as input. |
|
Schema for operations that require an issue number as input. |
|
Schema for operations that require a PR number as input. |
|
GitHub Toolkit. |
|
Schema for operations that do not require any input. |
|
Schema for operations that require a file path as input. |
|
Schema for operations that require a search query as input. |
|
Schema for operations that require a search query as input. |
|
Schema for operations that require a file path and content as input. |
|
GitLab Toolkit. |
|
Toolkit for interacting with Gmail. |
|
Jira Toolkit. |
|
Toolkit for interacting with a JSON spec. |
|
Toolkit for interacting with the Browser Agent. |
|
Nasa Toolkit. |
|
Natural Language API Tool. |
|
Natural Language API Toolkit. |
|
Toolkit for interacting with Office 365. |
|
|
Tool that sends a DELETE request and parses the response. |
Requests GET tool with LLM-instructed extraction of truncated responses. |
|
Requests PATCH tool with LLM-instructed extraction of truncated responses. |
|
Requests POST tool with LLM-instructed extraction of truncated responses. |
|
Requests PUT tool with LLM-instructed extraction of truncated responses. |
|
A reduced OpenAPI spec. |
|
Toolkit for interacting with an OpenAPI API. |
|
Toolkit for making REST requests. |
|
Toolkit for PlayWright browser tools. |
|
Polygon Toolkit. |
|
Toolkit for interacting with Power BI dataset. |
|
Toolkit for interacting with Slack. |
|
Toolkit for interacting with Spark SQL. |
|
SQLDatabaseToolkit for interacting with SQL databases. |
|
Steam Toolkit. |
|
Zapier Toolkit. |
Functions
Construct a json agent from an LLM and tools. |
|
Get a list of all possible tool names. |
|
Loads a tool from the HuggingFace Hub. |
|
|
Load tools based on their name. |
|
|
Construct an OpenAPI agent from an LLM and tools. |
|
Construct an OpenAI API planner and controller for a given spec. |
|
Simplify/distill/minify a spec somehow. |
|
Construct a Power BI agent from an LLM and tools. |
|
Construct a Power BI agent from a Chat LLM and tools. |
|
Construct a Spark SQL agent from an LLM and tools. |
|
Construct a SQL agent from an LLM and toolkit or database. |
agents#
Classes
cache#
Classes
|
Cache that uses Redis as a backend. |
|
Cache that uses Cosmos DB Mongo vCore vector-store backend |
|
Cache that uses Cassandra / Astra DB as a backend. |
|
Cache that uses Cassandra as a vector-store backend for semantic (i.e. similarity-based) lookup. |
|
SQLite table for full LLM Cache (all generations). |
|
SQLite table for full LLM Cache (all generations). |
|
Cache that uses GPTCache as a backend. |
Cache that stores things in memory. |
|
|
Cache that uses Memcached backend through pymemcache client lib |
|
Cache that uses Momento as a backend. |
|
Cache that uses OpenSearch vector store backend |
|
Cache that uses Redis as a backend. |
|
Cache that uses Redis as a vector-store backend. |
|
Cache that uses SQAlchemy as a backend. |
|
Cache that uses SQAlchemy as a backend. |
|
Cache that uses SQLite as a backend. |
|
Cache that uses SingleStore DB as a backend |
|
Cache that uses Upstash Redis as a backend. |
Deprecated classes
|
|
|
callbacks#
Classes
Callback Handler that logs to Aim. |
|
Callback handler for the metadata and associated function states for callbacks. |
|
Callback Handler that logs into Argilla. |
|
Callback Handler that logs to Arize. |
|
Callback Handler that logs to Arthur platform. |
|
|
Callback Handler that tracks bedrock anthropic info. |
Callback Handler that logs to ClearML. |
|
Callback Handler that logs to Comet. |
|
|
Callback Handler that logs into deepeval. |
Callback Handler that records transcripts to the Context service. |
|
Initialize Fiddler callback handler. |
|
Callback handler that is used within a Flyte task. |
|
Asynchronous callback for manually validating values. |
|
Callback for manually validating values. |
|
Exception to raise when a person manually review and rejects a value. |
|
Callback Handler that logs to Infino. |
|
|
Label Studio callback handler. |
Label Studio mode enumerator. |
|
|
Callback Handler for LLMonitor`. |
Context manager for LLMonitor user context. |
|
Callback Handler that logs metrics and artifacts to mlflow server. |
|
|
Callback Handler that logs metrics and artifacts to mlflow server. |
Callback Handler that tracks OpenAI info. |
|
|
Callback handler for promptlayer. |
Callback Handler that logs prompt artifacts and metrics to SageMaker Experiments. |
|
Child record as a NamedTuple. |
|
Enumerator of the child type. |
|
Streamlit expander that can be renamed and dynamically expanded/collapsed. |
|
|
A thought in the LLM's thought stream. |
|
Generates markdown labels for LLMThought containers. |
|
Enumerator of the LLMThought state. |
|
Callback handler that writes to a Streamlit app. |
|
Tool record as a NamedTuple. |
|
Comet Tracer. |
Arguments for the WandbTracer. |
|
Callback Handler that logs to Weights and Biases. |
|
Callback handler for Trubrics. |
|
|
Upstash Ratelimit Error |
|
Callback to handle rate limiting based on the number of requests or the number of tokens in the input. |
Callback Handler that logs evaluation results to uptrain and the console. |
|
The UpTrain data schema for tracking evaluation results. |
|
Handle the metadata and associated function states for callbacks. |
|
Callback Handler that logs to Weights and Biases. |
|
Callback Handler for logging to WhyLabs. |
Functions
Import the aim python package and raise an error if it is not installed. |
|
Import the clearml python package and raise an error if it is not installed. |
|
Import comet_ml and raise an error if it is not installed. |
|
Import the getcontext package. |
|
Import the fiddler python package and raise an error if it is not installed. |
|
Analyze text using textstat and spacy. |
|
Import flytekit and flytekitplugins-deck-standard. |
|
Calculate num tokens for OpenAI with tiktoken package. |
|
Import the infino client. |
|
Import tiktoken for counting tokens for OpenAI models. |
|
|
Get default Label Studio configs for the given mode. |
Builds an LLMonitor UserContextManager |
|
Get the Bedrock anthropic callback handler in a context manager. |
|
Get the OpenAI callback handler in a context manager. |
|
Get the WandbTracer in a context manager. |
|
Analyze text using textstat and spacy. |
|
|
Construct an html element from a prompt and a generation. |
Get the text complexity metrics from textstat. |
|
Import the mlflow python package and raise an error if it is not installed. |
|
Get the metrics to log to MLFlow. |
|
Get the cost in USD for a given model and number of tokens. |
|
Standardize the model name to a format that can be used in the OpenAI API. |
|
|
Save dict to local file path. |
Import comet_llm api and raise an error if it is not installed. |
|
Builds a nested dictionary from a list of runs. :param runs: The list of runs to build the tree from. :return: The nested dictionary representing the langchain Run in a tree structure compatible with WBTraceTree. |
|
Utility to flatten a nest run object into a list of runs. |
|
Utility to modify the serialized field of a list of runs dictionaries. removes any keys that match the exact_keys and any keys that contain any of the partial_keys. recursively moves the dictionaries under the kwargs key to the top level. changes the "id" field to a string "_kind" field that tells WBTraceTree how to visualize the run. promotes the "serialized" field to the top level. :param runs: The list of runs to modify. :param exact_keys: A tuple of keys to remove from the serialized field. :param partial_keys: A tuple of partial keys to remove from the serialized field. :return: The modified list of runs. |
|
Utility to truncate a list of runs dictionaries to only keep the specified |
|
Import the uptrain package. |
|
|
Flatten a nested dictionary into a flat dictionary. |
Hash a string using sha1. |
|
Import the pandas python package and raise an error if it is not installed. |
|
Import the spacy python package and raise an error if it is not installed. |
|
Import the textstat python package and raise an error if it is not installed. |
|
|
Load json file to a string. |
Analyze text using textstat and spacy. |
|
|
Construct an html element from a prompt and a generation. |
Import the wandb python package and raise an error if it is not installed. |
|
Load json file to a dictionary. |
|
Import the langkit python package and raise an error if it is not installed. |
chains#
Classes
Chain for question-answering against a graph by generating AQL statements. |
|
Chain for question-answering against a graph. |
|
Chain for question-answering against a graph by generating Cypher statements. |
|
Used to correct relationship direction in generated Cypher statements. |
|
Create new instance of Schema(left_node, relation, right_node) |
|
Chain for question-answering against a graph by generating Cypher statements. |
|
Chain for question-answering against a graph by generating gremlin statements. |
|
Chain for question-answering against a graph by generating gremlin statements. |
|
Question-answering against a graph by generating Cypher statements for Kùzu. |
|
Chain for question-answering against a graph by generating nGQL statements. |
|
Chain for question-answering against a Neptune graph by generating openCypher statements. |
|
Chain for question-answering against a Neptune graph by generating SPARQL statements. |
|
Question-answering against Ontotext GraphDB |
|
Question-answering against an RDF or OWL graph by generating SPARQL statements. |
|
Chain that requests a URL and then uses an LLM to parse results. |
|
Chain interacts with an OpenAPI endpoint using natural language. |
|
Get the request parser. |
|
Parse the request and error tags. |
|
Get the response parser. |
|
Parse the response and error tags. |
|
Retrieval Chain with Identity & Semantic Enforcement for question-answering against a vector database. |
|
Create a new model by parsing and validating input data from keyword arguments. |
|
Class for an authorization context. |
|
Create a new model by parsing and validating input data from keyword arguments. |
|
Input for PebbloRetrievalQA chain. |
|
Create a new model by parsing and validating input data from keyword arguments. |
|
Langchain framework details |
|
Create a new model by parsing and validating input data from keyword arguments. |
|
Create a new model by parsing and validating input data from keyword arguments. |
|
Create a new model by parsing and validating input data from keyword arguments. |
|
Create a new model by parsing and validating input data from keyword arguments. |
|
OS, language details |
|
Class for a semantic context. |
|
Class for a semantic entity filter. |
|
Class for a semantic topic filter. |
|
Create a new model by parsing and validating input data from keyword arguments. |
|
Wrapper for Pebblo Retrieval API. |
|
Routes available for the Pebblo API as enumerator. |
Functions
|
Convert a Python function to an Ernie function-calling API compatible dict. |
Convert a raw function/class to an Ernie function. |
|
[Legacy] Create an LLM chain that uses Ernie functions. |
|
Create a runnable sequence that uses Ernie functions. |
|
|
[Legacy] Create an LLMChain that uses an Ernie function to get a structured output. |
|
Create a runnable that uses an Ernie function to get a structured output. |
Get the appropriate function output parser given the user functions. |
|
Filter the schema based on included or excluded types |
|
Extract Cypher code from a text. |
|
Extract Cypher code from a text. |
|
Extract Gremlin code from a text. |
|
Extract Cypher code from a text. |
|
|
Remove a prefix from a text. |
Extract Cypher code from text using Regex. |
|
Trim the query to only include Cypher keywords. |
|
Decides whether to use the simple prompt |
|
Extract SPARQL code from a text. |
|
|
Clear the identity and semantic enforcement filters in the retriever search_kwargs. |
|
Set identity and semantic enforcement filters in the retriever. |
Fetch local runtime ip address. |
|
Fetch the current Framework and Runtime details. |
chat_loaders#
Classes
|
Load Facebook Messenger chat data from a folder. |
|
Load Facebook Messenger chat data from a single file. |
Load chat sessions from the iMessage chat.db SQLite file. |
|
Load chat sessions from a LangSmith dataset with the "chat" data type. |
|
Load chat sessions from a list of LangSmith "llm" runs. |
|
Load Slack conversations from a dump zip file. |
|
Load telegram conversations to LangChain chat messages. |
|
Load WhatsApp conversations from a dump zip file or directory. |
Functions
|
Convert nanoseconds since 2001 to a datetime object. |
Convert messages from the specified 'sender' to AI messages. |
|
Convert messages from the specified 'sender' to AI messages. |
|
|
Merge chat runs together. |
Merge chat runs together in a chat session. |
Deprecated classes
|
chat_message_histories#
Classes
|
Chat message history that is backed by Cassandra. |
|
Chat message history backed by Azure CosmosDB. |
|
Chat message history that stores history in AWS DynamoDB. |
Chat message history that stores history in a local file. |
|
|
Chat message history backed by Google Firestore. |
Consume start position for Kafka consumer to get chat history messages. |
|
Chat message history stored in Kafka. |
|
|
Chat message history cache that uses Momento as a backend. |
Chat message history stored in a Neo4j database. |
|
Chat message history stored in a Redis database. |
|
|
Uses Rockset to store chat messages. |
|
Chat message history stored in a SingleStoreDB database. |
Convert BaseMessage to the SQLAlchemy model. |
|
The default message converter for SQLChatMessageHistory. |
|
Chat message history stored in an SQL database. |
|
|
Chat message history that stores messages in Streamlit session state. |
Represents a chat message history stored in a TiDB database. |
|
|
Chat message history stored in an Upstash Redis database. |
Chat message history stored in a Xata database. |
|
Scope for the document search. |
|
Enumerator of the types of search to perform. |
|
Chat message history that uses Zep as a backend. |
|
|
Chat message history that uses Zep Cloud as a backend. |
Functions
Create topic if it doesn't exist, and return the number of partitions. |
|
Create a message model for a given table name. |
|
|
Condense Zep memory into a human message. |
|
Get the Zep role type from the role string. |
Deprecated classes
chat_models#
Classes
Anyscale Chat large language models. |
|
Azure ML Online Endpoint chat models. |
|
|
Chat Content formatter for models with OpenAI like API scheme. |
Deprecated: Kept for backwards compatibility |
|
Content formatter for LLaMA. |
|
Content formatter for Mistral. |
|
Baichuan chat model integration. |
|
Baidu Qianfan chat model integration. |
|
Adapter class to prepare the inputs from Langchain to prompt format that Chat model expects. |
|
Custom chat model for Cloudflare Workers AI |
|
ChatCoze chat models API by coze.com |
|
Dappier chat large language models. |
|
A chat model that uses the DeepInfra API. |
|
Exception raised when the DeepInfra API returns an error. |
|
EdenAI chat large language models. |
|
EverlyAI Chat large language models. |
|
Fake ChatModel for testing purposes. |
|
Fake ChatModel for testing purposes. |
|
Friendli LLM for chat. |
|
Google PaLM Chat models API. |
|
Error with the Google PaLM API. |
|
GPTRouter by Writesonic Inc. |
|
Error with the GPTRouter APIs |
|
GPTRouter model. |
|
ChatModel which returns user input as the response. |
|
Tencent Hunyuan chat models API by Tencent. |
|
Javelin AI Gateway chat models API. |
|
Parameters for the Javelin AI Gateway LLM. |
|
Jina AI Chat models API. |
|
Kinetica LLM Chat Model API. |
|
Fetch and return data from the Kinetica LLM. |
|
Response containing SQL and the fetched data. |
|
Kinetica utility functions. |
|
ChatKonko Chat large language models API. |
|
Chat model that uses the LiteLLM API. |
|
Error with the LiteLLM I/O library |
|
LiteLLM Router as LangChain Model. |
|
Chat with LLMs via llama-api-server |
|
llama.cpp model. |
|
MariTalk Chat models API. |
|
Initialize RequestException with request and response objects. |
|
MiniMax chat model integration. |
|
MLflow chat models API. |
|
MLflow AI Gateway chat models API. |
|
Parameters for the MLflow AI Gateway LLM. |
|
MLX chat models. |
|
Moonshot large language models. |
|
NCP ClovaStudio Chat Completion API. |
|
OCI Data Science Model Deployment chat model integration. |
|
OCI large language chat models deployed with Text Generation Inference. |
|
OCI large language chat models deployed with vLLM. |
|
ChatOCIGenAI chat model integration. |
|
OctoAI Chat large language models. |
|
Outlines chat model integration. |
|
Alibaba Cloud PAI-EAS LLM Service chat model API. |
|
Perplexity AI Chat models API. |
|
PremAI Chat models. |
|
Error with the PremAI API. |
|
PromptLayer and OpenAI Chat large language models API. |
|
Reka chat large language models. |
|
SambaNova Cloud chat model. |
|
SambaStudio chat model. |
|
Snowflake Cortex based Chat model |
|
Error with Snowpark client. |
|
IFlyTek Spark chat model integration. |
|
Nebula chat large language model - https://docs.symbl.ai/docs/nebula-llm |
|
Alibaba Tongyi Qwen chat model integration. |
|
Volc Engine Maas hosts a plethora of models. |
|
Writer chat model. |
|
YandexGPT large language models. |
|
Yi chat models API. |
|
Yuan2.0 Chat models API. |
|
ZhipuAI chat model integration. |
Functions
|
Format a list of messages into a full prompt for the Anthropic model |
Async context manager for connecting to an SSE stream. |
|
|
Convert a message to a dictionary that can be passed to the API. |
Convert a list of messages to a prompt for mistral. |
|
Get the request for the Cohere chat API. |
|
|
Get the role of the message. |
Use tenacity to retry the async completion call. |
|
Use tenacity to retry the completion call for streaming. |
|
Use tenacity to retry the completion call. |
|
Define conditional decorator. |
|
Convert a dict response to a message. |
|
|
Get a request of the Friendli chat API. |
|
Get role of the message. |
Use tenacity to retry the async completion call. |
|
|
Use tenacity to retry the completion call. |
Use tenacity to retry the async completion call. |
|
Use tenacity to retry the completion call. |
|
Return the body for the model router input. |
|
Use tenacity to retry the async completion call. |
|
Use tenacity to retry the async completion call. |
|
Get llm output from usage and params. |
|
Convert a list of messages to a prompt for llama. |
|
Async context manager for connecting to an SSE stream. |
|
Context manager for connecting to an SSE stream. |
|
Use tenacity to retry the async completion call. |
|
|
Using tenacity for retry in completion call |
Create a retry decorator for PremAI API errors. |
|
Convert LangChain messages to Reka message format. |
|
|
Process content to handle both text and media inputs, returning a list of content items. |
Process a single content item. |
|
Convert a dict to a message. |
|
Convert a message chunk to a message. |
|
Convert a message to a dict. |
|
Convert a dict to a message. |
|
Use tenacity to retry the async completion call. |
|
Use tenacity to retry the completion call. |
|
|
|
Use tenacity to retry the async completion call. |
|
|
Async context manager for connecting to an SSE stream. |
|
Context manager for connecting to an SSE stream. |
Deprecated classes
cross_encoders#
Classes
Fake cross encoder model. |
|
HuggingFace cross encoder models. |
|
|
Content handler for CrossEncoder class. |
|
SageMaker Inference CrossEncoder endpoint. |
docstore#
Classes
|
Docstore via arbitrary lookup function. |
Mixin class that supports adding texts. |
|
Interface to access to place that stores documents. |
|
|
Simple in memory docstore in the form of a dict. |
Wikipedia API. |
document_compressors#
Classes
Document compressor that uses DashScope Rerank API. |
|
Document compressor using Flashrank interface. |
|
Document compressor that uses Infinity Rerank API. |
|
Document compressor that uses Jina Rerank API. |
|
Compress using LLMLingua Project. |
|
OpenVINO rerank models. |
|
Request for reranking. |
|
Document compressor using Flashrank interface. |
|
Document compressor that uses Volcengine Rerank API. |
document_loaders#
Classes
|
Load acreom vault from a directory. |
Load with an Airbyte source connector implemented using the CDK. |
|
Load from Gong using an Airbyte source connector. |
|
Load from Hubspot using an Airbyte source connector. |
|
Load from Salesforce using an Airbyte source connector. |
|
Load from Shopify using an Airbyte source connector. |
|
Load from Stripe using an Airbyte source connector. |
|
Load from Typeform using an Airbyte source connector. |
|
Load from Zendesk Support using an Airbyte source connector. |
|
Load local Airbyte json files. |
|
Load the Airtable tables. |
|
Load datasets from Apify web scraping, crawling, and data extraction platform. |
|
Load records from an ArcGIS FeatureLayer. |
|
|
Load a query result from Arxiv. |
Load AssemblyAI audio transcripts. |
|
|
Load AssemblyAI audio transcripts. |
Transcript format to use for the document loader. |
|
Load HTML asynchronously. |
|
|
Load documents from AWS Athena. |
Load AZLyrics webpages. |
|
Load from Azure AI Data. |
|
|
Load from Azure Blob Storage container. |
|
Load from Azure Blob Storage files. |
|
Load from Baidu BOS directory. |
|
Load from Baidu Cloud BOS file. |
Base class for all loaders that uses O365 Package |
|
|
Load a bibtex file. |
Load fetching transcripts from BiliBili videos. |
|
Load a Blackboard course. |
|
|
Load blobs from cloud URL or file:. |
|
Load blobs in the local file system. |
|
Load YouTube urls as audio file(s). |
Load elements from a blockchain smart contract. |
|
Enumerator of the supported blockchains. |
|
Load with Brave Search engine. |
|
Load pre-rendered web pages using a headless browser hosted on Browserbase. |
|
Load webpages with Browserless /content endpoint. |
|
Document Loader for Apache Cassandra. |
|
|
Load conversations from exported ChatGPT data. |
Microsoft Compiled HTML Help (CHM) Parser. |
|
Load CHM files using Unstructured. |
|
Scrape HTML pages from URLs using a headless instance of the Chromium. |
|
|
Load College Confidential webpages. |
Load and pars Documents concurrently. |
|
Load Confluence pages. |
|
Enumerator of the content formats of Confluence page. |
|
|
Load CoNLL-U files. |
Load documents from Couchbase. |
|
|
Load a CSV file into a list of Documents. |
Load CSV files using Unstructured. |
|
Load Cube semantic layer metadata. |
|
Load Datadog logs. |
|
Initialize with dataframe object. |
|
Load Pandas DataFrame. |
|
Load files using dedoc API. The file loader automatically detects the file type (even with the wrong extension). By default, the loader makes a call to the locally hosted dedoc API. More information about dedoc API can be found in dedoc documentation: https://dedoc.readthedocs.io/en/latest/dedoc_api_usage/api.html. |
|
Base Loader that uses dedoc (https://dedoc.readthedocs.io). |
|
DedocFileLoader document loader integration to load files using dedoc. |
|
Load Diffbot json file. |
|
Load from a directory. |
|
Load Discord chat logs. |
|
|
Load a PDF with Azure Document Intelligence. |
Load from Docusaurus Documentation. |
|
Load files from Dropbox. |
|
Load from DuckDB. |
|
Loads Outlook Message files using extract_msg. |
|
Load email files using Unstructured. |
|
Load EPub files using Unstructured. |
|
Load transactions from Ethereum mainnet. |
|
Load from EverNote. |
|
Load Microsoft Excel files using Unstructured. |
|
Load Facebook Chat messages directory dump. |
|
|
Load from FaunaDB. |
Load Figma file. |
|
FireCrawlLoader document loader integration |
|
Generic Document Loader. |
|
Load geopandas Dataframe. |
|
|
Load Git repository files. |
|
Load GitBook data. |
Load GitHub repository Issues. |
|
Load issues of a GitHub repository. |
|
Load GitHub File |
|
Load table schemas from AWS Glue. |
|
Load from Gutenberg.org. |
|
File encoding as the NamedTuple. |
|
|
Load Hacker News data. |
Load HTML files using Unstructured. |
|
|
__ModuleName__ document loader integration |
|
Load from Hugging Face Hub datasets. |
|
Load model information from Hugging Face Hub, including README content. |
|
Load iFixit repair guides, device wikis and answers. |
Load PNG and JPG files using Unstructured. |
|
Load image captions. |
|
Load IMSDb webpages. |
|
|
Load from IUGU. |
Load notes from Joplin. |
|
Load a JSON file using a jq schema. |
|
Load from Kinetica API. |
|
Client for lakeFS. |
|
|
Load from lakeFS. |
Load from lakeFS as unstructured data. |
|
Load from LarkSuite (FeiShu). |
|
Load from LarkSuite (FeiShu) wiki. |
|
Load Documents using LLMSherpa. |
|
Load Markdown files using Unstructured. |
|
Load the Mastodon 'toots'. |
|
Load from Alibaba Cloud MaxCompute table. |
|
Load MediaWiki dump from an XML file. |
|
Merge documents from a list of loaders |
|
|
Parse MHTML files with BeautifulSoup. |
Load elements from a blockchain smart contract. |
|
Load from Modern Treasury. |
|
Load MongoDB documents. |
|
|
Load news articles from URLs using Unstructured. |
Load Jupyter notebook (.ipynb) files. |
|
Load Notion directory dump. |
|
Load from Notion DB. |
|
|
Load from any file type using Nuclia Understanding API. |
Load from Huawei OBS directory. |
|
Load from the Huawei OBS file. |
|
Load Obsidian files from directory. |
|
Load OpenOffice ODT files using Unstructured. |
|
Load documents from Microsoft OneDrive. |
|
Load a file from Microsoft OneDrive. |
|
Load pages from OneNote notebooks. |
|
Load from Open City. |
|
|
Load from oracle adb |
Read documents using OracleDocLoader :param conn: Oracle Connection, :param params: Loader parameters. |
|
Read a file |
|
Splitting text using Oracle chunker. |
|
Parse Oracle doc metadata... |
|
Load Org-Mode files using Unstructured. |
|
Transcribe and parse audio files using Azure OpenAI Whisper. |
|
Transcribe and parse audio files with faster-whisper. |
|
Transcribe and parse audio files. |
|
|
Transcribe and parse audio files with OpenAI Whisper model. |
Transcribe and parse audio files. |
|
|
Loads a PDF with Azure Document Intelligence (formerly Forms Recognizer). |
Dataclass to store Document AI parsing results. |
|
Parser that uses mime-types to parse a blob. |
|
Load article PDF files using Grobid. |
|
Exception raised when the Grobid server is unavailable. |
|
Parse HTML files using Beautiful Soup. |
|
Code segmenter for C. |
|
|
Code segmenter for COBOL. |
|
Abstract class for the code segmenter. |
Code segmenter for C++. |
|
|
Code segmenter for C#. |
|
Code segmenter for Elixir. |
Code segmenter for Go. |
|
Code segmenter for Java. |
|
|
Code segmenter for JavaScript. |
|
Code segmenter for Kotlin. |
|
Parse using the respective programming language syntax. |
Code segmenter for Lua. |
|
Code segmenter for Perl. |
|
Code segmenter for PHP. |
|
|
Code segmenter for Python. |
Code segmenter for Ruby. |
|
Code segmenter for Rust. |
|
|
Code segmenter for Scala. |
|
Abstract class for `CodeSegmenter`s that use the tree-sitter library. |
|
Code segmenter for TypeScript. |
Parse the Microsoft Word documents from a blob. |
|
Send PDF files to Amazon Textract and parse them. |
|
|
Loads a PDF with Azure Document Intelligence (formerly Form Recognizer) and chunks at character level. |
Parse PDF using PDFMiner. |
|
Parse PDF with PDFPlumber. |
|
Parse PDF using PyMuPDF. |
|
Load PDF using pypdf |
|
Parse PDF with PyPDFium2. |
|
Parser for text blobs. |
|
Parser for vsdx files. |
|
Load PDF files from a local file system, HTTP or S3. |
|
|
Base Loader class for PDF files. |
|
DedocPDFLoader document loader integration to load PDF files using dedoc. The file loader can automatically detect the correctness of a textual layer in the PDF document. Note that __init__ method supports parameters that differ from ones of DedocBaseLoader. |
Load a PDF with Azure Document Intelligence |
|
|
Load PDF files using Mathpix service. |
|
Load online PDF. |
|
Load PDF files using PDFMiner. |
Load PDF files as HTML content using PDFMiner. |
|
|
Load PDF files using pdfplumber. |
alias of |
|
|
Load PDF files using PyMuPDF. |
Load a directory with PDF files using pypdf and chunks at character level. |
|
|
PyPDFLoader document loader integration |
|
Load PDF using pypdfium2 and chunks at character level. |
Load PDF files using Unstructured. |
|
|
Document loader utilizing Zerox library: getomni-ai/zerox |
Pebblo Safe Loader class is a wrapper around document loaders enabling the data to be scrutinized. |
|
Loader for text data. |
|
|
Load Polars DataFrame. |
|
Load Microsoft PowerPoint files using Unstructured. |
Load from Psychic.dev. |
|
Load from the PubMed biomedical library. |
|
|
Load PySpark DataFrames. |
|
Load Python files, respecting any non-default encoding if specified. |
|
Load Quip pages. |
Load ReadTheDocs documentation directory. |
|
|
Recursively load all child links from a root URL. |
Load Reddit posts. |
|
Load Roam files from a directory. |
|
Column not found error. |
|
Load from a Rockset database. |
|
|
Load content from RSpace notebooks, folders, documents or PDF Gallery files. |
|
Load news articles from RSS feeds using Unstructured. |
Load RST files using Unstructured. |
|
Load RTF files using Unstructured. |
|
Load from Amazon AWS S3 directory. |
|
|
Load from Amazon AWS S3 file. |
Turn a url to llm accessible markdown with Scrapfly.io. |
|
Turn an url to LLM accessible markdown with ScrapingAnt. |
|
Load from SharePoint. |
|
|
Load a sitemap and its URLs. |
Load from a Slack directory dump. |
|
Load from Snowflake API. |
|
Load web pages as Documents using Spider AI. |
|
Load from Spreedly API. |
|
Load documents by querying database tables supported by SQLAlchemy. |
|
|
Load .srt (subtitle) files. |
|
Load from Stripe API. |
Load SurrealDB documents. |
|
Load Telegram chat json directory dump. |
|
Load from Telegram chat dump. |
|
alias of |
|
|
Load from Tencent Cloud COS directory. |
Load from Tencent Cloud COS file. |
|
|
Load from TensorFlow Dataset. |
|
Load text file. |
|
Load documents from TiDB. |
Load HTML using 2markdown API. |
|
|
Load TOML files. |
|
Load cards from a Trello board. |
Load TSV files using Unstructured. |
|
Load Twitter tweets. |
|
Base Loader that uses Unstructured. |
|
Load files from remote URLs using Unstructured. |
|
Abstract base class for all evaluators. |
|
Load HTML pages with Playwright and parse with Unstructured. |
|
|
Evaluate the page HTML content using the unstructured library. |
Load HTML pages with Selenium and parse with Unstructured. |
|
|
Initialize with file path. |
Load weather data with Open Weather Map API. |
|
WebBaseLoader document loader integration |
|
Load WhatsApp messages text file. |
|
Load from Wikipedia. |
|
Load DOCX file using docx2txt and chunks at character level. |
|
|
Load Microsoft Word file using Unstructured. |
Load XML file using Unstructured. |
|
Load Xorbits DataFrame. |
|
Generic Google API Client. |
|
Load all Videos from a YouTube Channel. |
|
Output formats of transcripts from YoutubeLoader. |
|
|
Load YouTube video transcripts. |
|
Load documents from Yuque. |
Functions
Fetch the mime types for the specified file types. |
|
Fetch the mime types for the specified file types. |
|
Combine message information in a readable format ready to be used. |
|
Combine message information in a readable format ready to be used. |
|
Try to detect the file encoding. |
|
Combine cells information in a readable format ready to be used. |
|
Recursively remove newlines, no matter the data structure they are stored in. |
|
|
Extract text from images with RapidOCR. |
Get a parser by parser name. |
|
Default joiner for content columns. |
|
Combine message information in a readable format ready to be used. |
|
Convert a string or list of strings to a list of Documents with metadata. |
|
Retrieve a list of elements from the Unstructured API. |
|
|
Check if the installed Unstructured version exceeds the minimum version for the feature in question. |
|
Raise an error if the Unstructured version does not exceed the specified minimum. |
Combine message information in a readable format ready to be used. |
Deprecated classes
document_transformers#
Classes
|
Transform HTML content by extracting specific tags and removing unwanted ones. |
|
Extract properties from text documents using doctran. |
|
Extract QA from text documents using doctran. |
|
Translate text documents using doctran. |
|
Perform K-means clustering on document vectors. |
|
Filter that drops redundant documents by comparing their embeddings. |
Replace occurrences of a particular search pattern with a replacement string |
|
|
Reorder long context. |
|
Converts HTML documents to Markdown format with customizable options for handling links, images, other tags and heading styles using the markdownify library. |
|
Nuclia Text Transformer. |
Extract metadata tags from document contents using OpenAI functions. |
Functions
|
Get all navigable strings from a BeautifulSoup element. |
|
Convert a list of documents to a list of documents with state. |
|
Create a DocumentTransformer that uses an OpenAI function chain to automatically |
Deprecated classes
embeddings#
Classes
|
Aleph Alpha's asymmetric semantic embedding. |
Symmetric version of the Aleph Alpha's semantic embeddings. |
|
Anyscale Embeddings API. |
|
Ascend NPU accelerate Embedding model |
|
Embedding documents and queries with Awa DB. |
|
Baichuan Text Embedding models. |
|
Baidu Qianfan Embeddings embedding models. |
|
Bookend AI sentence_transformers embedding models. |
|
Clarifai embedding models. |
|
|
Cloudflare Workers AI embedding model. |
DashScope embedding models. |
|
Deep Infra's embedding inference service. |
|
EdenAI embedding. |
|
Embaas's embedding service. |
|
Payload for the Embaas embeddings API. |
|
Fake embedding model that always returns the same embedding vector for the same text. |
|
Fake embedding model. |
|
Qdrant FastEmbedding models. |
|
Google's PaLM Embeddings APIs. |
|
GPT4All embedding models. |
|
Gradient.ai Embedding models. |
|
|
Deprecated, TinyAsyncGradientEmbeddingClient was removed. |
HuggingFace sentence_transformers embedding models. |
|
Embed texts using the HuggingFace API. |
|
Wrapper around sentence_transformers embedding models. |
|
Self-hosted embedding models for infinity package. |
|
|
Helper tool to embed Infinity. |
Optimized Infinity embedding models. |
|
Wrapper around the BGE embedding model with IPEX-LLM optimizations on Intel CPUs and GPUs. |
|
Leverage Itrex runtime to unlock the performance of compressed NLP models. |
|
Javelin AI Gateway embeddings. |
|
Jina embedding models. |
|
JohnSnowLabs embedding models |
|
LASER Language-Agnostic SEntence Representations. |
|
llama.cpp embedding models. |
|
Llamafile lets you distribute and run large language models with a single file. |
|
LLMRails embedding models. |
|
LocalAI embedding models. |
|
MiniMax embedding model integration. |
|
Cohere embedding LLMs in MLflow. |
|
Embedding LLMs in MLflow. |
|
MLflow AI Gateway embeddings. |
|
ModelScopeHub embedding models. |
|
MosaicML embedding service. |
|
NCP ClovaStudio Embedding API. |
|
NLP Cloud embedding models. |
|
OCI authentication types as enumerator. |
|
OCI embedding models. |
|
OctoAI Compute Service embedding models. |
|
OpenVNO BGE embedding models. |
|
OpenVINO embedding models. |
|
Quantized bi-encoders embedding models. |
|
Get Embeddings |
|
OVHcloud AI Endpoints Embeddings. |
|
Prem's Embedding APIs |
|
Content handler for LLM class. |
|
Custom Sagemaker Inference Endpoints. |
|
SambaNova embedding models. |
|
Custom embedding models on self-hosted remote hardware. |
|
|
HuggingFace embedding models on self-hosted remote hardware. |
|
HuggingFace InstructEmbedding models on self-hosted remote hardware. |
Embeddings by spaCy models. |
|
Exception raised for errors in the header assembly. |
|
SparkLLM embedding model integration. |
|
|
URL class for parsing the URL. |
TensorflowHub embedding models. |
|
text2vec embedding models. |
|
|
A client to handle synchronous and asynchronous requests to the TextEmbed API. |
A class to handle embedding requests to the TextEmbed API. |
|
|
Device to use for inference, cuda or cpu. |
Exception raised when no consumer group is provided on initialization of TitanTakeoffEmbed or in embed request. |
|
Configuration for the reader to be deployed in Takeoff. |
|
Custom exception for interfacing with Takeoff Embedding class. |
|
Interface with Takeoff Inference API for embedding models. |
|
Volcengine Embeddings embedding models. |
|
Xinference embedding models. |
|
YandexGPT Embeddings models. |
|
ZhipuAI embedding model integration. |
Functions
Use tenacity to retry the embedding call. |
|
Use tenacity to retry the completion call. |
|
|
Get the bytes string of a file. |
Check if a URL is a local file. |
|
Use tenacity to retry the embedding call. |
|
Use tenacity to retry the embedding call. |
|
Use tenacity to retry the completion call. |
|
|
Check if an endpoint is live by sending a GET request to the specified URL. |
Use tenacity to retry the embedding call. |
|
Use tenacity to retry the embedding call. |
|
Create a retry decorator for PremAIEmbeddings. |
|
|
Using tenacity for retry in embedding calls |
|
Load the embedding model. |
Use tenacity to retry the completion call. |
|
Use tenacity to retry the embedding call. |
Deprecated classes
example_selectors#
Classes
Select and order examples based on ngram overlap score (sentence_bleu score from NLTK package). |
Functions
Compute ngram overlap score of source and example as sentence_bleu score from NLTK package. |
graph_vectorstores#
Classes
Functions
|
|
Return the networkx directed graph corresponding to the documents. |
graphs#
Classes
|
Apache AGE wrapper for graph operations. |
|
Exception for the AGE queries. |
ArangoDB wrapper for graph operations. |
|
|
FalkorDB wrapper for graph operations. |
Represents a graph document consisting of nodes and relationships. |
|
Represents a node in a graph with associated properties. |
|
Represents a directed relationship between two nodes in a graph. |
|
Abstract class for graph operations. |
|
|
Gremlin wrapper for graph operations. |
|
HugeGraph wrapper for graph operations. |
Functionality to create graph index. |
|
|
Kùzu wrapper for graph operations. |
|
Memgraph wrapper for graph operations. |
|
NebulaGraph wrapper for graph operations. |
|
Neo4j database wrapper for various graph operations. |
Abstract base class for Neptune. |
|
Neptune Analytics wrapper for graph operations. |
|
|
Neptune wrapper for graph operations. |
Exception for the Neptune queries. |
|
Neptune wrapper for RDF graph operations. |
|
Knowledge triple in the graph. |
|
Networkx wrapper for entity graph operations. |
|
Ontotext GraphDB https://graphdb.ontotext.com/ wrapper for graph operations. |
|
|
RDFlib wrapper for graph operations. |
TigerGraph wrapper for graph operations. |
Functions
Get the Arango DB client from credentials. |
|
Clean string values for schema. |
|
Sanitize the input dictionary or list. |
|
|
Extract entities from entity string. |
Parse knowledge triples from the knowledge string. |
indexes#
Classes
|
Abstract base class for a record manager. |
llms#
Classes
AI21 large language models. |
|
Parameters for AI21 penalty data. |
|
Aleph Alpha large language models. |
|
Amazon API Gateway to access LLM models hosted on AWS. |
|
Adapter to prepare the inputs from Langchain to a format that LLM model expects. |
|
Anyscale large language models. |
|
Aphrodite language model. |
|
Arcee's Domain Adapted Language Models (DALMs). |
|
Aviary hosted models. |
|
|
Aviary backend. |
Azure ML Online Endpoint models. |
|
Azure ML endpoints API types. |
|
AzureML Managed Endpoint client. |
|
Azure ML Online Endpoint models. |
|
Transform request and response of AzureML endpoint to match with required schema. |
|
Content formatter for models that use the OpenAI like API scheme. |
|
Content handler for the Dolly-v2-12b model |
|
Content handler for GPT2 |
|
Content handler for LLMs from the HuggingFace catalog. |
|
Deprecated: Kept for backwards compatibility |
|
Deprecated: Kept for backwards compatibility |
|
Baichuan large language models. |
|
Baidu Qianfan completion model integration. |
|
Banana large language models. |
|
Baseten model |
|
Beam API for gpt2 large language model. |
|
Base class for Bedrock models. |
|
Adapter class to prepare the inputs from Langchain to a format that LLM model expects. |
|
Wrapper around the BigdlLLM model |
|
NIBittensor LLMs |
|
CerebriumAI large language models. |
|
ChatGLM LLM service. |
|
ChatGLM3 LLM service. |
|
Clarifai large language models. |
|
Cloudflare Workers AI service. |
|
C Transformers LLM models. |
|
CTranslate2 language model. |
|
DeepInfra models. |
|
Neural Magic DeepSparse LLM interface. |
|
EdenAI models. |
|
ExllamaV2 API. |
|
Fake LLM for testing purposes. |
|
Fake streaming list LLM for testing purposes. |
|
ForefrontAI large language models. |
|
Base class of Friendli. |
|
Friendli LLM. |
|
GigaChat large language models API. |
|
GooseAI large language models. |
|
GPT4All language models. |
|
Gradient.ai LLM Endpoints. |
|
Train result. |
|
User input as the response. |
|
IpexLLM model. |
|
Javelin AI Gateway LLMs. |
|
Parameters for the Javelin AI Gateway LLM. |
|
Kobold API language model. |
|
Konko AI models. |
|
Layerup Security LLM service. |
|
llama.cpp model. |
|
Llamafile lets you distribute and run large language models with a single file. |
|
HazyResearch's Manifest library. |
|
Minimax large language models. |
|
Common parameters for Minimax large language models. |
|
MLflow LLM service. |
|
MLflow AI Gateway LLMs. |
|
Parameters for the MLflow AI Gateway LLM. |
|
MLX Pipeline API. |
|
Modal large language models. |
|
Moonshot large language models. |
|
Common parameters for Moonshot LLMs. |
|
MosaicML LLM service. |
|
NLPCloud large language models. |
|
|
Base class for LLM deployed on OCI Data Science Model Deployment. |
|
LLM deployed on OCI Data Science Model Deployment. |
|
OCI Data Science Model Deployment TGI Endpoint. |
|
VLLM deployed on OCI Data Science Model Deployment |
Raises when encounter server error when making inference. |
|
|
Raises when token expired. |
|
OCI authentication types as enumerator. |
OCI large language models. |
|
Base class for OCI GenAI models |
|
OctoAI LLM Endpoints - OpenAI compatible. |
|
Raised when the Ollama endpoint is not found. |
|
LLM that uses OpaquePrompts to sanitize prompts. |
|
Base OpenAI large language model class. |
|
Parameters for identifying a model as a typed dict. |
|
OpenLLM, supporting both in-process model instance and remote OpenLLM servers. |
|
OpenLM models. |
|
LLM wrapper for the Outlines library. |
|
Langchain LLM class to help to access eass llm service. |
|
Petals Bloom models. |
|
PipelineAI large language models. |
|
Use your Predibase models with Langchain. |
|
Prediction Guard large language models. |
|
PromptLayer OpenAI large language models. |
|
PromptLayer OpenAI large language models. |
|
Replicate models. |
|
RWKV language models. |
|
Handler class to transform input from LLM to a format that SageMaker endpoint expects. |
|
Content handler for LLM class. |
|
Parse the byte stream input. |
|
Sagemaker Inference Endpoint models. |
|
SambaNova Cloud large language models. |
|
SambaStudio large language models. |
|
Model inference on self-hosted remote hardware. |
|
HuggingFace Pipeline API to run on self-hosted remote hardware. |
|
Solar large language models. |
|
Common configuration for Solar LLMs. |
|
iFlyTek Spark completion model integration. |
|
StochasticAI large language models. |
|
Nebula Service models. |
|
Text generation models from WebUI. |
|
|
The device to use for inference, cuda or cpu |
Configuration for the reader to be deployed in Titan Takeoff API. |
|
Titan Takeoff API LLMs. |
|
Tongyi completion model integration. |
|
VLLM language model. |
|
vLLM OpenAI-compatible API client |
|
Base class for VolcEngineMaas models. |
|
volc engine maas hosts a plethora of models. |
|
Weight only quantized model. |
|
Writer large language models. |
|
Xinference large-scale model inference service. |
|
Yandex large language models. |
|
Yi large language models. |
|
Wrapper around You.com's conversational Smart and Research APIs. |
|
Yuan2.0 language models. |
Functions
|
Create the LLMResult from the choices and prompts. |
|
Update token usage. |
|
Get completions from Aviary models. |
List available models |
|
|
Use tenacity to retry the completion call. |
|
Use tenacity to retry the completion call. |
Get the default Databricks personal access token. |
|
Get the default Databricks workspace hostname. |
|
Get the notebook REPL context if running inside a Databricks notebook. |
|
|
Use tenacity to retry the completion call. |
Use tenacity to retry the completion call. |
|
Use tenacity to retry the completion call for streaming. |
|
|
Use tenacity to retry the completion call. |
Use tenacity to retry the completion call. |
|
Conditionally apply a decorator. |
|
|
Use tenacity to retry the completion call. |
Remove trailing slash and /api from url if present. |
|
|
Default guardrail violation handler. |
|
Load LLM from a file. |
|
Load LLM from Config Dict. |
|
Use tenacity to retry the async completion call. |
|
Use tenacity to retry the completion call. |
|
Update token usage. |
Use tenacity to retry the completion call. |
|
|
Generate text from the model. |
Generate elements from an async iterable, and a boolean indicating if it is the last element. |
|
|
Async version of stream_generate_with_retry. |
Check the response from the completion call. |
|
Generate elements from an iterable, and a boolean indicating if it is the last element. |
|
|
Use tenacity to retry the completion call. |
|
Use tenacity to retry the completion call. |
|
Cut off the text as soon as any stop words occur. |
|
Use tenacity to retry the completion call. |
|
Use tenacity to retry the completion call. |
|
Return True if the model name is a Codey model. |
|
Return True if the model name is a Gemini model. |
|
Use tenacity to retry the async completion call. |
|
Use tenacity to retry the completion call. |
Deprecated classes
memory#
Classes
Knowledge graph conversation memory. |
|
Chat message memory backed by Motorhead service. |
|
Persist your chain history to the Zep MemoryStore. |
output_parsers#
Classes
Parse an output as the element of the Json object. |
|
Parse an output as the Json object. |
|
Parse an output that is one of sets of values. |
|
|
Parse an output as an attribute of a pydantic object. |
|
Parse an output as a pydantic object. |
Parse the output of an LLM call using Guardrails. |
query_constructors#
Classes
Translate AstraDB internal query language elements to valid filters. |
|
Translate Chroma internal query language elements to valid filters. |
|
Logic for converting internal query language elements to valid filters. |
|
|
Translate Databricks vector search internal query language elements to valid filters. |
Translate DeepLake internal query language elements to valid filters. |
|
Translate DingoDB internal query language elements to valid filters. |
|
Translate Elasticsearch internal query language elements to valid filters. |
|
Translate internal query language elements to valid filters params for HANA vectorstore. |
|
Translate Milvus internal query language elements to valid filters. |
|
Translate Mongo internal query language elements to valid filters. |
|
Translate MyScale internal query language elements to valid filters. |
|
Translate Neo4j internal query language elements to valid filters. |
|
Translate OpenSearch internal query domain-specific language elements to valid filters. |
|
Translate PGVector internal query language elements to valid filters. |
|
Translate Pinecone internal query language elements to valid filters. |
|
Translate Qdrant internal query language elements to valid filters. |
|
Visitor for translating structured queries to Redis filter expressions. |
|
Translate Langchain filters to Supabase PostgREST filters. |
|
|
Translate StructuredQuery to Tencent VectorDB query. |
|
Translate the internal query language elements to valid filters. |
Translate Vectara internal query language elements to valid filters. |
|
Translate Weaviate internal query language elements to valid filters. |
Functions
Check if a string can be cast to a float. |
|
Convert a value to a string and add double quotes if it is a string. |
|
Convert a value to a string and add single quotes if it is a string. |
retrievers#
Classes
Arcee Domain Adapted Language Models (DALMs) retriever. |
|
Arxiv retriever. |
|
AskNews retriever. |
|
Azure AI Search service retriever. |
|
Azure Cognitive Search service retriever. |
|
Amazon Bedrock Knowledge Bases retriever. |
|
Configuration for retrieval. |
|
Configuration for vector search. |
|
BM25 retriever without Elasticsearch. |
|
A retriever class for Breebs. |
|
Chaindesk API retriever. |
|
ChatGPT plugin retriever. |
|
Databerry API retriever. |
|
DocArray Document Indices retriever. |
|
|
Enumerator of the types of search to perform. |
Dria retriever using the DriaAPIWrapper. |
|
Elasticsearch retriever that uses BM25. |
|
Embedchain retriever. |
|
|
Google Vertex Search API retriever alias for backwards compatibility. |
Retriever for Kay.ai datasets. |
|
Additional result attribute. |
|
Value of an additional result attribute. |
|
Amazon Kendra Index retriever. |
|
Document attribute. |
|
Value of a document attribute. |
|
Information that highlights the keywords in the excerpt. |
|
Amazon Kendra Query API search result. |
|
Query API result item. |
|
Base class of a result item. |
|
Amazon Kendra Retrieve API search result. |
|
Retrieve API result item. |
|
Text with highlights. |
|
KNN retriever. |
|
LlamaIndex graph data structure retriever. |
|
LlamaIndex retriever. |
|
Metal API retriever. |
|
Milvus API retriever. |
|
`NanoPQ retriever. |
|
Retriever for Outline API. |
|
|
Pinecone Hybrid Search retriever. |
PubMed API retriever. |
|
Rememberizer retriever. |
|
LangChain API retriever. |
|
SVM retriever. |
|
Search depth as enumerator. |
|
Tavily Search API retriever. |
|
TF-IDF retriever. |
|
Document retriever that uses ThirdAI's NeuralDB. |
|
Vespa retriever. |
|
|
Weaviate hybrid search retriever. |
Output parser for a list of numbered questions. |
|
Search queries to research for the user's goal. |
|
Google Search API retriever. |
|
Wikipedia API retriever. |
|
You.com Search API retriever. |
|
|
Which documents to search. |
|
Enumerator of the types of search to perform. |
Zep MemoryStore Retriever. |
|
Zep Cloud MemoryStore Retriever. |
|
Zilliz API retriever. |
Functions
|
Clean an excerpt from Kendra. |
Combine a ResultItem title and excerpt into a single string. |
|
|
Create an index of embeddings for a list of contexts. |
|
Deprecated MilvusRetreiver. |
|
Create an index of embeddings for a list of contexts. |
Create an index from a list of contexts. |
|
Hash a text using SHA256. |
|
|
Create an index of embeddings for a list of contexts. |
|
Deprecated ZillizRetreiver. |
Deprecated classes
storage#
Classes
|
Base class for the DataStax AstraDB data store. |
|
A ByteStore implementation using Cassandra as the backend. |
|
BaseStore implementation using MongoDB as the underlying store. |
|
BaseStore implementation using MongoDB as the underlying store. |
|
BaseStore implementation using Redis as the underlying store. |
|
Table used to save values. |
|
BaseStore interface that works on an SQL database. |
BaseStore implementation using Upstash Redis as the underlying store to store raw bytes. |
Functions
|
Deprecated classes
|
|
|
|
|
tools#
Classes
Tool for app operations. |
|
Type of app operation as enumerator. |
|
Schema for app operations. |
|
Base class for the AINetwork tools. |
|
|
Type of operation as enumerator. |
Tool for owner operations. |
|
Schema for owner operations. |
|
Tool for owner operations. |
|
Schema for owner operations. |
|
Tool for transfer operations. |
|
Schema for transfer operations. |
|
Tool for value operations. |
|
Schema for value operations. |
|
Base Tool for Amadeus. |
|
Tool for finding the closest airport to a particular location. |
|
Schema for the AmadeusClosestAirport tool. |
|
Tool for searching for a single flight between two airports. |
|
Schema for the AmadeusFlightSearch tool. |
|
Input for the Arxiv tool. |
|
Tool that searches the Arxiv API. |
|
Tool that searches the AskNews API. |
|
Input for the AskNews Search tool. |
|
|
HuggingFace Text-to-Speech Model Inference. |
|
Tool that queries the Azure AI Services Document Intelligence API. |
|
Tool that queries the Azure AI Services Image Analysis API. |
|
Tool that queries the Azure AI Services Speech to Text API. |
|
Tool that queries the Azure AI Services Text Analytics for Health API. |
|
Tool that queries the Azure AI Services Text to Speech API. |
|
Tool that queries the Azure Cognitive Services Form Recognizer API. |
|
Tool that queries the Azure Cognitive Services Image Analysis API. |
|
Tool that queries the Azure Cognitive Services Speech2Text API. |
|
Tool that queries the Azure Cognitive Services Text2Speech API. |
|
Tool that queries the Azure Cognitive Services Text Analytics for Health API. |
Tool for evaluating python code in a sandbox environment. |
|
Arguments for the BearlyInterpreterTool. |
|
Information about a file to be uploaded. |
|
Bing Search tool. |
|
Tool that queries the Bing search API. |
|
Tool that queries the BraveSearch. |
|
Base tool for interacting with an Apache Cassandra database. |
|
|
Tool for getting the schema of a keyspace in an Apache Cassandra database. |
|
Tool for getting data from a table in an Apache Cassandra database. |
Tool for querying an Apache Cassandra database with provided CQL. |
|
Tool that queries the Clickup API. |
|
Tool that uses the Cogniswitch service to answer questions. |
|
Tool that uses the Cogniswitch services to store data from file. |
|
Tool that uses the Cogniswitch services to store data from a URL. |
|
Tool that uses the Cogniswitch services to get the |
|
Connery Action model. |
|
Connery Action parameter model. |
|
Connery Action parameter validation model. |
|
Service for interacting with the Connery Runner API. |
|
Connery Action tool. |
|
Create a new model by parsing and validating input data from keyword arguments. |
|
Tool that queries the DataForSeo Google Search API and get back json. |
|
Tool that queries the DataForSeo Google search API. |
|
Tool that queries using the Dataherald SDK. |
|
Create a new model by parsing and validating input data from keyword arguments. |
|
Input for the DuckDuckGo search tool. |
|
Tool that queries the DuckDuckGo search API and returns the results in output_format. |
|
DuckDuckGo tool. |
|
Tool for running python code in a sandboxed environment for data analysis. |
|
Arguments for the E2BDataAnalysisTool. |
|
Description of the uploaded path with its remote path. |
|
Traverse an AST and output source code for the abstract syntax; original formatting is disregarded. |
|
Tool that queries the Eden AI Speech To Text API. |
|
Create a new model by parsing and validating input data from keyword arguments. |
|
Tool that queries the Eden AI Text to speech API. |
|
Create a new model by parsing and validating input data from keyword arguments. |
|
the base tool for all the EdenAI Tools . |
|
Tool that queries the Eden AI Explicit image detection. |
|
Create a new model by parsing and validating input data from keyword arguments. |
|
|
Tool that queries the Eden AI Object detection API. |
Create a new model by parsing and validating input data from keyword arguments. |
|
Tool that queries the Eden AI Identity parsing API. |
|
Create a new model by parsing and validating input data from keyword arguments. |
|
Tool that queries the Eden AI Invoice parsing API. |
|
Create a new model by parsing and validating input data from keyword arguments. |
|
Tool that queries the Eden AI Explicit text detection. |
|
Create a new model by parsing and validating input data from keyword arguments. |
|
Models available for Eleven Labs Text2Speech. |
|
Models available for Eleven Labs Text2Speech. |
|
Tool that queries the Eleven Labs Text2Speech API. |
|
Tool that copies a file. |
|
Input for CopyFileTool. |
|
Tool that deletes a file. |
|
Input for DeleteFileTool. |
|
Input for FileSearchTool. |
|
Tool that searches for files in a subdirectory that match a regex pattern. |
|
Input for ListDirectoryTool. |
|
Tool that lists files and directories in a specified folder. |
|
Input for MoveFileTool. |
|
Tool that moves a file. |
|
Input for ReadFileTool. |
|
Tool that reads a file. |
|
Mixin for file system tools. |
|
Error for paths outside the root directory. |
|
Input for WriteFileTool. |
|
Tool that writes a file to disk. |
|
Tool that gets balance sheets for a given ticker over a given period. |
|
Input for BalanceSheets. |
|
|
Tool that gets cash flow statements for a given ticker over a given period. |
|
Input for CashFlowStatements. |
Tool that gets income statements for a given ticker over a given period. |
|
|
Input for IncomeStatements. |
Tool for interacting with the GitHub API. |
|
Tool for interacting with the GitLab API. |
|
Base class for Gmail tools. |
|
Input for CreateDraftTool. |
|
Tool that creates a draft email for Gmail. |
|
Tool that gets a message by ID from Gmail. |
|
Input for GetMessageTool. |
|
Input for GetMessageTool. |
|
Tool that gets a thread by ID from Gmail. |
|
Tool that searches for messages or threads in Gmail. |
|
|
Enumerator of Resources to search. |
Input for SearchGmailTool. |
|
Tool that sends a message to Gmail. |
|
Input for SendMessageTool. |
|
Tool that adds the capability to query using the Golden API and get back JSON. |
|
Input for the GoogleBooksQuery tool. |
|
Tool that searches the Google Books API. |
|
Tool that queries the Google Finance API. |
|
Tool that queries the Google Jobs API. |
|
Tool that queries the Google Lens API. |
|
Input for GooglePlacesTool. |
|
Tool that queries the Google search API. |
|
Tool that queries the Serper.dev Google Search API and get back json. |
|
Tool that queries the Serper.dev Google search API. |
|
Tool that queries the Google trends API. |
|
Base tool for querying a GraphQL API. |
|
Tool that asks user for input. |
|
IFTTT Webhook. |
|
Input for the Jina search tool. |
|
Tool that queries the JinaSearch. |
|
Tool that queries the Atlassian Jira API. |
|
Tool for getting a value in a JSON spec. |
|
Tool for listing keys in a JSON spec. |
|
Base class for JSON spec. |
|
Tool that trains a language model. |
|
|
Protocol for trainable language models. |
Tool that searches the Merriam-Webster API. |
|
Initialize the tool. |
|
Input for UpdateSessionTool. |
|
Tool that closes an existing Multion Browser Window with provided fields. |
|
Input for CreateSessionTool. |
|
Tool that creates a new Multion Browser Window with provided fields. |
|
Tool that updates an existing Multion Browser Window with provided fields. |
|
Input for UpdateSessionTool. |
|
Tool that queries the Atlassian Jira API. |
|
Input for Nuclia Understanding API. |
|
Tool to process files with the Nuclia Understanding API. |
|
Base class for the Office 365 tools. |
|
|
Input for SendMessageTool. |
Tool for creating a draft email in Office 365. |
|
Search calendar events in Office 365. |
|
Input for SearchEmails Tool. |
|
Search email messages in Office 365. |
|
Input for SearchEmails Tool. |
|
Tool for sending calendar events in Office 365. |
|
Input for CreateEvent Tool. |
|
Send an email in Office 365. |
|
Input for SendMessageTool. |
|
|
Tool that generates an image using OpenAI DALLE. |
A model for a single API operation. |
|
A model for a property in the query, path, header, or cookie params. |
|
Base model for an API property. |
|
The location of the property. |
|
A model for a request body. |
|
A model for a request body property. |
|
Tool that queries the OpenWeatherMap API. |
|
Tool that queries the Passio Nutrition AI API. |
|
Inputs to the Passio Nutrition AI tool. |
|
Base class for browser tools. |
|
Tool for clicking on an element with the given CSS selector. |
|
Input for ClickTool. |
|
Tool for getting the URL of the current webpage. |
|
Explicit no-args input for CurrentWebPageTool. |
|
Extract all hyperlinks on the page. |
|
|
Input for ExtractHyperlinksTool. |
Tool for extracting all the text on the current webpage. |
|
Explicit no-args input for ExtractTextTool. |
|
Tool for getting elements in the current web page matching a CSS selector. |
|
Input for GetElementsTool. |
|
Tool for navigating a browser to a URL. |
|
Input for NavigateToolInput. |
|
Navigate back to the previous page in the browser history. |
|
Explicit no-args input for NavigateBackTool. |
|
AI Plugin Definition. |
|
Tool for getting the OpenAPI spec for an AI Plugin. |
|
Schema for AIPluginTool. |
|
API Configuration. |
|
Tool that gets aggregate bars (stock prices) over a given date range for a given ticker from Polygon. |
|
Input for PolygonAggregates. |
|
Inputs for Polygon's Financials API |
|
Tool that gets the financials of a ticker from Polygon |
|
Inputs for Polygon's Last Quote API |
|
Tool that gets the last quote of a ticker from Polygon |
|
Inputs for Polygon's Ticker News API |
|
Tool that gets the latest news for a given ticker from Polygon |
|
Tool for getting metadata about a PowerBI Dataset. |
|
Tool for getting tables names. |
|
Tool for querying a Power BI Dataset. |
|
Tool that searches the PubMed API. |
|
Tool that queries for posts on a subreddit. |
|
Input for Reddit search. |
|
Base class for requests tools. |
|
Tool for making a DELETE request to an API endpoint. |
|
Tool for making a GET request to an API endpoint. |
|
Tool for making a PATCH request to an API endpoint. |
|
Tool for making a POST request to an API endpoint. |
|
Tool for making a PUT request to an API endpoint. |
|
A tool implementation to execute JavaScript via Riza's Code Interpreter API. |
|
Create a new model by parsing and validating input data from keyword arguments. |
|
Riza Code tool. |
|
Create a new model by parsing and validating input data from keyword arguments. |
|
Input for SceneXplain. |
|
Tool that explains images. |
|
Tool that queries the SearchApi.io search API and returns JSON. |
|
Tool that queries the SearchApi.io search API. |
|
Input for the SearxSearch tool. |
|
Tool that queries a Searx instance and gets back json. |
|
Tool that queries a Searx instance. |
|
Tool that searches the semanticscholar API. |
|
Input for the SemanticScholar tool. |
|
Commands for the Bash Shell tool. |
|
Tool to run shell commands. |
|
Base class for Slack tools. |
|
Tool that gets Slack channel information. |
|
Tool that gets Slack messages. |
|
Input schema for SlackGetMessages. |
|
Input for ScheduleMessageTool. |
|
Tool for scheduling a message in Slack. |
|
Input for SendMessageTool. |
|
Tool for sending a message in Slack. |
|
Input for CopyFileTool. |
|
Tool that adds the capability to sleep. |
|
Base tool for interacting with Spark SQL. |
|
Tool for getting metadata about a Spark SQL. |
|
Tool for getting tables names. |
|
Use an LLM to check if a query is correct. |
|
Tool for querying a Spark SQL. |
|
Base tool for interacting with a SQL database. |
|
Tool for getting metadata about a SQL database. |
|
Tool for getting tables names. |
|
Use an LLM to check if a query is correct. |
|
Tool for querying a SQL database. |
|
Tool that uses StackExchange |
|
Tool that searches the Steam Web API. |
|
Supported Image Models for generation. |
|
|
Tool used to generate images from a text-prompt. |
Tool that queries the Tavily Search API and gets back an answer. |
|
Input for the Tavily tool. |
|
Tool that queries the Tavily Search API and gets back json. |
|
Base class for tools that use a VectorStore. |
|
Tool for the VectorDBQA chain. |
|
Tool for the VectorDBQAWithSources chain. |
|
Tool that searches the Wikidata API. |
|
Input for the WikipediaQuery tool. |
|
Tool that searches the Wikipedia API. |
|
Tool that queries using the Wolfram Alpha SDK. |
|
Input for the YahooFinanceNews tool. |
|
Tool that searches financial news on Yahoo Finance. |
|
Input schema for the you.com tool. |
|
Tool that searches the you.com API. |
|
Tool that queries YouTube. |
|
Returns a list of all exposed (enabled) actions associated with current user (associated with the set api_key). |
|
Executes an action that is identified by action_id, must be exposed |
|
|
|
|
|
Create a new model by parsing and validating input data from keyword arguments. |
|
Initialize the tool. |
Functions
|
Authenticate using the AIN Blockchain |
Authenticate using the Amadeus API |
|
Detect if the file is local or remote. |
|
Download audio from url to local. |
|
|
Detect if the file is local or remote. |
|
Download audio from url to local. |
Convert a file to base64. |
|
|
Get the first n lines of a file. |
|
Strip markdown code from a string. |
Deprecated. |
|
Add print statement to the last line if it's missing. |
|
Call f on each item in seq, calling inter() in between. |
|
Parse a file and pretty-print it to output. |
|
|
Resolve a relative path, raising an error if not within the root directory. |
Check if path is relative to root. |
|
Build a Gmail service. |
|
Clean email body. |
|
Get credentials. |
|
Import google libraries. |
|
Import googleapiclient.discovery.build function. |
|
Import InstalledAppFlow class. |
|
Tool for asking the user for input. |
|
Authenticate using the Microsoft Graph API |
|
Clean body of a message or event. |
|
Lazy import playwright browsers. |
|
Asynchronously get the current page of the browser. |
|
|
Create an async playwright browser. |
|
Create a playwright browser. |
Get the current page of the browser. |
|
Run an async coroutine. |
|
Convert the yaml or json serialized spec to a dict. |
|
Authenticate using the Slack API. |
|
|
Upload a block to a signed URL and return the public URL. |
Deprecated classes
utilities#
Classes
Wrapper for AlphaVantage API for Currency Exchange Rate. |
|
Wrapper around Apify. |
|
Arcee document. |
|
Adapter for Arcee documents |
|
Source of an Arcee document. |
|
|
Routes available for the Arcee API as enumerator. |
|
Wrapper for Arcee API. |
Filters available for a DALM retrieval and generation. |
|
|
Filter types available for a DALM retrieval as enumerator. |
Wrapper around ArxivAPI. |
|
Wrapper for AskNews API. |
|
|
Setup mode for AstraDBEnvironment as enumerator. |
Wrapper for AWS Lambda SDK. |
|
Wrapper around bibtexparser. |
|
Wrapper for Bing Web Search API. |
|
Wrapper around the Brave search engine. |
|
|
|
Apache Cassandra® database wrapper. |
|
Exception raised for errors in the database schema. |
|
Create a new model by parsing and validating input data from keyword arguments. |
|
|
Component class for a list. |
Wrapper for Clickup API. |
|
Base class for all components. |
|
|
Component class for a member. |
|
Component class for a space. |
|
Class for a task. |
|
Component class for a team. |
Wrapper for OpenAI's DALL-E Image Generator. |
|
Wrapper around the DataForSeo API. |
|
Wrapper for Dataherald. |
|
|
Wrapper around Dria API. |
Wrapper for DuckDuckGo Search API. |
|
Wrapper for financial datasets API. |
|
Wrapper for GitHub API. |
|
Wrapper for GitLab API. |
|
Wrapper for Golden. |
|
Wrapper around Google Books API. |
|
Wrapper for SerpApi's Google Finance API |
|
Wrapper for SerpApi's Google Scholar API |
|
Wrapper for SerpApi's Google Lens API |
|
Wrapper for Google Scholar API |
|
Wrapper around the Serper.dev Google Search API. |
|
Wrapper for SerpApi's Google Scholar API |
|
Wrapper around GraphQL API. |
|
Wrapper for Infobip API for messaging. |
|
Wrapper around the Jina search engine. |
|
Wrapper for Jira API. |
|
Interface for querying Alibaba Cloud MaxCompute tables. |
|
Wrapper for Merriam-Webster. |
|
Wrapper for Metaphor Search API. |
|
Create a new model by parsing and validating input data from keyword arguments. |
|
Wrapper for NASA API. |
|
alias of |
|
|
A message containing streaming audio. |
alias of |
|
alias of |
|
alias of |
|
A runnable that performs Automatic Speech Recognition (ASR) using NVIDIA Riva. |
|
An enum of the possible choices for Riva audio encoding. |
|
Configuration for the authentication to a Riva service connection. |
|
A collection of common Riva settings. |
|
A runnable that performs Text-to-Speech (TTS) with NVIDIA Riva. |
|
An empty Sentinel type. |
|
|
Enumerator of the HTTP verbs. |
OpenAPI Model that removes mis-formatted parts of the spec. |
|
Wrapper for OpenWeatherMap API using PyOWM. |
|
|
Get Summary :param conn: Oracle Connection, :param params: Summary parameters, :param proxy: Proxy |
Wrapper around OutlineAPI. |
|
Manage the token for the NutritionAI API. |
|
Mixin to prevent storing on disk. |
|
Wrapper for the Passio Nutrition AI API. |
|
Pebblo AI application. |
|
Pebblo document. |
|
Pebblo Framework instance. |
|
Pebblo Indexed Document. |
|
Wrapper for Pebblo Loader API. |
|
|
Routes available for the Pebblo API as enumerator. |
Pebblo Runtime. |
|
Wrapper for Polygon API. |
|
Portkey configuration. |
|
Create PowerBI engine from dataset ID and credential or token. |
|
Wrapper around PubMed API. |
|
Wrapper for Reddit API |
|
|
Escape punctuation within an input string. |
Wrapper for Rememberizer APIs. |
|
Lightweight wrapper around requests library. |
|
Lightweight wrapper around requests library, with async support. |
|
Wrapper around requests to handle auth and async. |
|
alias of |
|
Lightweight wrapper around requests library, with async support. |
|
Wrapper for SceneXplain API. |
|
Wrapper around SearchApi API. |
|
Dict like wrapper around search api results. |
|
Wrapper for Searx API. |
|
Wrapper around semanticscholar.org API. |
|
Context manager to hide prints. |
|
Wrapper around SerpAPI. |
|
|
SparkSQL is a utility class for interacting with Spark SQL. |
|
SQLAlchemy wrapper around a database. |
Wrapper for Stack Exchange API. |
|
Wrapper for Steam API. |
|
Wrapper for Tavily Search API. |
|
Access to the TensorFlow Datasets. |
|
Messaging Client using Twilio. |
|
Wrapper around the Wikidata API. |
|
Wrapper around WikipediaAPI. |
|
Wrapper for Wolfram Alpha. |
|
Output from you.com API. |
|
Output of parsing one snippet. |
|
A single hit from you.com, which may contain multiple snippets |
|
Metadata on a single hit from you.com |
|
Wrapper for you.com Search and News API. |
|
Wrapper for Zapier NLA. |
Functions
Get the number of tokens in a string of text. |
|
Get the token ids for a string of text. |
|
|
Execute a CQL query asynchronously. |
Wrap a Cassandra response future in an asyncio future. |
|
|
Extract elements from a dictionary. |
|
Fetch data from a URL. |
|
Fetch the first id from a dictionary. |
|
Fetch the folder id. |
|
Fetch the list id. |
|
Fetch the space id. |
|
Fetch the team id. |
|
Parse a JSON string and return the parsed object. |
Parse a dictionary by creating a component and then turning it back into a dictionary. |
|
Restore the original sensitive data from the sanitized text. |
|
Sanitize input string or dict of strings by replacing sensitive data with placeholders. |
|
Check if a HTTP response is retryable. |
|
Calculate the content size in bytes: - Encode the string to bytes using a specific encoding (e.g., UTF-8) - Get the length of the encoded bytes. |
|
Generate batches of documents based on page_content size. |
|
Fetch owner of local file path. |
|
Return an absolute local path for a local file/directory, for a network related path, return as is. |
|
Fetch local runtime ip address. |
|
Return an absolute source path of source of loader based on the keys present in Document. |
|
|
Return loader type among, file, dir or in-memory. |
Fetch the current Framework and Runtime details. |
|
|
Fetch size of source path. |
Add single quotes around table names that contain spaces. |
|
|
Convert a JSON object to a markdown table. |
Check if the correct Redis modules are installed. |
|
|
Get a redis client from the connection url given. |
|
Truncate a string to a certain number of words, based on the max string length. |
Create a retry decorator for Vertex / Palm LLMs. |
|
|
Return a custom user agent header. |
|
Init Vertex AI. |
Load an image from Google Cloud Storage. |
|
Raise ImportError related to Vertex SDK being not available. |
Deprecated classes
utils#
Classes
Representation of a callable function to the Ernie API. |
|
Representation of a callable function to the Ernie API. |
Functions
|
Convert a Pydantic model to a function description for the Ernie API. |
Convert a Pydantic model to a function description for the Ernie API. |
|
|
Return a custom user agent header. |
Row-wise cosine similarity between two equal-width matrices. |
|
|
Row-wise cosine similarity with optional top-k and score threshold filtering. |
Return whether OpenAI API is v1 or more. |
|
Get user agent from environment variable. |
vectorstores#
Classes
|
Aerospike vector store. |
|
Alibaba Cloud OpenSearch vector store. |
|
Alibaba Cloud Opensearch` client configuration. |
|
AnalyticDB (distributed PostgreSQL) vector store. |
|
Annoy vector store. |
|
Apache Doris vector store. |
Apache Doris client configuration. |
|
|
Create a vectorstore backed by ApertureDB |
|
Atlas vector store. |
|
AwaDB vector store. |
Azure Cosmos DB for MongoDB vCore vector store. |
|
Cosmos DB Similarity Type as enumerator. |
|
|
Cosmos DB Vector Search Type as enumerator. |
|
Azure Cosmos DB for NoSQL vector store. |
|
Azure Cognitive Search vector store. |
Retriever that uses Azure Cognitive Search. |
|
|
|
Baidu Elasticsearch vector store. |
|
Baidu VectorDB as a vector store. |
|
Baidu VectorDB Connection params. |
|
|
Baidu VectorDB table params. |
|
Apache Cassandra(R) for vector-store workloads. |
|
Clarifai AI vector store. |
|
ClickHouse vector store integration. |
ClickHouse client configuration. |
|
DashVector vector store. |
|
Activeloop Deep Lake vector store. |
|
|
Dingo vector store. |
Base class for DocArray based vector stores. |
|
HnswLib storage using DocArray package. |
|
In-memory DocArray storage for exact search. |
|
DocumentDB Similarity Type as enumerator. |
|
Amazon DocumentDB (with MongoDB compatibility) vector store. |
|
|
DuckDB vector store. |
ecloud Elasticsearch vector store. |
|
Base class for Elasticsearch retrieval strategies. |
|
|
Wrapper around Epsilla vector database. |
|
FAISS vector store integration. |
|
SAP HANA Cloud Vector Engine |
|
Hippo vector store. |
|
Hologres API vector store. |
Helper class for Infinispan REST interface. |
|
Infinispan VectorStore interface. |
|
|
Jaguar API vector store. |
|
KDB.AI vector store. |
|
Some default dimensions for known embeddings. |
Enumerator of the Distance strategies. |
|
|
Kinetica vector store. |
Kinetica client configuration. |
|
|
LanceDB vector store. |
Base class for the Lantern embedding store. |
|
Enumerator of the Distance strategies. |
|
|
Postgres with the lantern extension as a vector store. |
Result from a query. |
|
Implementation of Vector Store using LLMRails. |
|
Retriever for LLMRails. |
|
ManticoreSearch Engine vector store. |
|
Create a new model by parsing and validating input data from keyword arguments. |
|
|
Marqo vector store. |
|
Meilisearch vector store. |
Momento Vector Index (MVI) vector store. |
|
|
MyScale vector store. |
MyScale client configuration. |
|
MyScale vector store without metadata column |
|
|
Enumerator of the index types. |
|
Neo4j vector index. |
Enumerator of the Distance strategies. |
|
|
NucliaDB vector store. |
|
Amazon OpenSearch Vector Engine vector store. |
|
OracleVS vector store. |
VectorStore connecting to Pathway Vector Store. |
|
|
Base model for all SQL stores. |
Collection store. |
|
|
Embedding store. |
|
Postgres with the pg_embedding extension as a vector store. |
Result from a query. |
|
|
VectorStore backed by pgvecto_rs. |
|
Base model for the SQL stores. |
Enumerator of the Distance strategies. |
|
Qdrant related exceptions. |
|
|
Redis vector database. |
Retriever for Redis VectorStore. |
|
Collection of RedisFilterFields. |
|
Logical expression of RedisFilterFields. |
|
Base class for RedisFilterFields. |
|
RedisFilterOperator enumerator is used to create RedisFilterExpressions. |
|
RedisFilterField representing a numeric field in a Redis index. |
|
RedisFilterField representing a tag in a Redis index. |
|
RedisFilterField representing a text field in a Redis index. |
|
Schema for flat vector fields in Redis. |
|
Schema for HNSW vector fields in Redis. |
|
Schema for numeric fields in Redis. |
|
Distance metrics for Redis vector fields. |
|
Base class for Redis fields. |
|
Schema for Redis index. |
|
Base class for Redis vector fields. |
|
Schema for tag fields in Redis. |
|
Schema for text fields in Redis. |
|
|
Relyt (distributed PostgreSQL) vector store. |
|
Rockset vector store. |
|
ScaNN vector store. |
|
SemaDB vector store. |
SingleStore DB vector store. |
|
|
Base class for serializing data. |
|
Serialize data in Binary JSON using the bson python package. |
|
Serialize data in JSON using the json package from python standard library. |
Serialize data in Apache Parquet format using the pyarrow package. |
|
Simple in-memory vector store based on the scikit-learn library NearestNeighbors. |
|
Exception raised by SKLearnVectorStore. |
|
|
SQLite with Vec extension as a vector database. |
|
SQLite with VSS extension as a vector database. |
|
StarRocks vector store. |
StarRocks client configuration. |
|
Supabase Postgres vector store. |
|
SurrealDB as Vector Store. |
|
|
Tair vector store. |
Tencent vector DB Connection params. |
|
Tencent vector DB Index params. |
|
MetaData Field for Tencent vector DB. |
|
Tencent VectorDB as a vector store. |
|
Vectorstore that uses ThirdAI's NeuralDB Enterprise Python Client for NeuralDBs. |
|
Vectorstore that uses ThirdAI's NeuralDB. |
|
TiDB Vector Store. |
|
|
Tigris vector store. |
|
TileDB vector store. |
Timescale Postgres vector store |
|
|
Typesense vector store. |
Upstash Vector vector store |
|
|
USearch vector store. |
|
Enumerator of the Distance strategies for calculating distances between vectors. |
|
Vald vector database. |
|
Intel Lab's VDMS for vector-store workloads. |
|
Initialize vearch vector store flag 1 for cluster,0 for standalone |
|
Configuration for Maximal Marginal Relevance (MMR) search. |
Configuration for Reranker. |
|
Configuration for summary generation. |
|
|
Vectara API vector store. |
|
Configuration for Vectara query. |
|
Vectara RAG runnable. |
Vectara Retriever class. |
|
|
Vespa vector store. |
|
vikingdb as a vector store |
|
vikingdb connection config |
|
VLite is a simple and fast vector database for semantic search. |
|
Weaviate vector store. |
|
Xata vector store. |
|
Yellowbrick as a vector database. |
|
Configuration for a Zep Collection. |
|
Zep vector store. |
Zep vector store. |
|
|
Zilliz vector store. |
Functions
|
Create metadata from fields. |
Import annoy if available, otherwise raise error. |
|
|
Check if a string contains multiple substrings. |
Import faiss if available, otherwise raise error. |
|
Import lancedb package. |
|
Converts a dict filter to a LanceDB filter string. |
|
Get the embedding store class. |
|
|
Check if a string contains multiple substrings. |
Check if the values are not None or empty string |
|
Transform the input data into the desired format. |
|
Combine multiple queries with an operator. |
|
Construct a metadata filter. |
|
Convert a dictionary to a YAML-like string without using external libraries. |
|
Remove Lucene special characters |
|
Sort first element to match the index_name if exists |
|
|
Create an index on the vector store. |
Drop an index if it exists. |
|
Drop a table and purge it from the database. |
|
Decorator to call the synchronous method of the class if the async method is not implemented. |
|
Check if Redis index exists. |
|
Decorator to check for misuse of equality operators. |
|
Read in the index schema from a dict or yaml file. |
|
Import scann if available, otherwise raise error. |
|
Normalize vectors to unit length. |
|
Serializes a list of floats into a compact "raw bytes" format |
|
Print a debug message if DEBUG is True. |
|
Get a named result from a query. |
|
|
Check if a string has multiple substrings. |
Translate LangChain filter to Tencent VectorDB filter. |
|
Import tiledb-vector-search if available, otherwise raise error. |
|
Get the URI of the documents array. |
|
|
Get the URI of the documents array from group. |
Get the URI of the vector index. |
|
Get the URI of the vector index. |
|
Import usearch if available, otherwise raise error. |
|
Filter out metadata types that are not supported for a vector store. |
|
Calculate maximal marginal relevance. |
|
|
VDMS client for the VDMS server. |
|
Convert embedding to bytes. |
Deprecated classes
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|