chat_with_retry#
- langchain_community.chat_models.premai.chat_with_retry(llm: ChatPremAI, project_id: int, messages: List[dict], stream: bool = False, run_manager: CallbackManagerForLLMRun | None = None, **kwargs: Any) Any [source]#
Using tenacity for retry in completion call
- Parameters:
llm (ChatPremAI) –
project_id (int) –
messages (List[dict]) –
stream (bool) –
run_manager (CallbackManagerForLLMRun | None) –
kwargs (Any) –
- Return type:
Any