OpenAIWhisperParser#
- class langchain_community.document_loaders.parsers.audio.OpenAIWhisperParser(api_key: str | None = None, *, chunk_duration_threshold: float = 0.1, base_url: str | None = None, language: str | None = None, prompt: str | None = None, response_format: Literal['json', 'text', 'srt', 'verbose_json', 'vtt'] | None = None, temperature: float | None = None)[source]#
Transcribe and parse audio files.
Audio transcription is with OpenAI Whisper model.
- Parameters:
api_key (str | None) – OpenAI API key
chunk_duration_threshold (float) – minimum duration of a chunk in seconds NOTE: According to the OpenAI API, the chunk duration should be at least 0.1 seconds. If the chunk duration is less or equal than the threshold, it will be skipped.
base_url (str | None) –
language (str | None) –
prompt (str | None) –
response_format (Literal['json', 'text', 'srt', 'verbose_json', 'vtt'] | None) –
temperature (float | None) –
Methods
__init__
([api_key, ...])lazy_parse
(blob)Lazily parse the blob.
parse
(blob)Eagerly parse the blob into a document or documents.
- __init__(api_key: str | None = None, *, chunk_duration_threshold: float = 0.1, base_url: str | None = None, language: str | None = None, prompt: str | None = None, response_format: Literal['json', 'text', 'srt', 'verbose_json', 'vtt'] | None = None, temperature: float | None = None)[source]#
- Parameters:
api_key (str | None) –
chunk_duration_threshold (float) –
base_url (str | None) –
language (str | None) –
prompt (str | None) –
response_format (Literal['json', 'text', 'srt', 'verbose_json', 'vtt'] | None) –
temperature (float | None) –
Examples using OpenAIWhisperParser