chat_with_retry#

langchain_community.chat_models.premai.chat_with_retry(llm: ChatPremAI, project_id: int, messages: List[dict], stream: bool = False, run_manager: CallbackManagerForLLMRun | None = None, **kwargs: Any) Any[source]#

Using tenacity for retry in completion call

Parameters:
  • llm (ChatPremAI) –

  • project_id (int) –

  • messages (List[dict]) –

  • stream (bool) –

  • run_manager (CallbackManagerForLLMRun | None) –

  • kwargs (Any) –

Return type:

Any