Source code for langchain_core.outputs.chat_result

from typing import List, Optional

from langchain_core.outputs.chat_generation import ChatGeneration
from langchain_core.pydantic_v1 import BaseModel

[docs]class ChatResult(BaseModel): """Class that contains all results for a single chat model call.""" generations: List[ChatGeneration] """List of the chat generations. This is a List because an input can have multiple candidate generations. """ llm_output: Optional[dict] = None """For arbitrary LLM provider specific output."""