LLMInputOutputAdapter#
- class langchain_community.llms.bedrock.LLMInputOutputAdapter[source]#
Adapter class to prepare the inputs from Langchain to a format that LLM model expects.
It also provides helper function to extract the generated text from the model response.
Attributes
provider_to_output_key_map
Methods
aprepare_output_stream
(provider, response[, ...])prepare_input
(provider, model_kwargs[, ...])prepare_output
(provider, response)prepare_output_stream
(provider, response[, ...])- classmethod aprepare_output_stream(
- provider: str,
- response: Any,
- stop: List[str] | None = None,
- Parameters:
provider (str)
response (Any)
stop (List[str] | None)
- Return type:
AsyncIterator[GenerationChunk]
- classmethod prepare_input(
- provider: str,
- model_kwargs: Dict[str, Any],
- prompt: str | None = None,
- system: str | None = None,
- messages: List[Dict] | None = None,
- Parameters:
provider (str)
model_kwargs (Dict[str, Any])
prompt (str | None)
system (str | None)
messages (List[Dict] | None)
- Return type:
Dict[str, Any]
- classmethod prepare_output(
- provider: str,
- response: Any,
- Parameters:
provider (str)
response (Any)
- Return type:
dict
- classmethod prepare_output_stream(
- provider: str,
- response: Any,
- stop: List[str] | None = None,
- messages_api: bool = False,
- Parameters:
provider (str)
response (Any)
stop (List[str] | None)
messages_api (bool)
- Return type:
Iterator[GenerationChunk]