Home / Function/ on_llm_end() — langchain Function Reference

on_llm_end() — langchain Function Reference

Architecture documentation for the on_llm_end() function in streaming_stdout.py from the langchain codebase.

Entity Profile

Dependency Diagram

graph TD
  e6dcf34b_4146_9acb_b04d_df843dac031c["on_llm_end()"]
  6b65bf57_0fa9_b411_5886_294d6dbe5842["StreamingStdOutCallbackHandler"]
  e6dcf34b_4146_9acb_b04d_df843dac031c -->|defined in| 6b65bf57_0fa9_b411_5886_294d6dbe5842
  style e6dcf34b_4146_9acb_b04d_df843dac031c fill:#6366f1,stroke:#818cf8,color:#fff

Relationship Graph

Source Code

libs/core/langchain_core/callbacks/streaming_stdout.py lines 60–66

    def on_llm_end(self, response: LLMResult, **kwargs: Any) -> None:
        """Run when LLM ends running.

        Args:
            response: The response from the LLM.
            **kwargs: Additional keyword arguments.
        """

Domain

Subdomains

Frequently Asked Questions

What does on_llm_end() do?
on_llm_end() is a function in the langchain codebase, defined in libs/core/langchain_core/callbacks/streaming_stdout.py.
Where is on_llm_end() defined?
on_llm_end() is defined in libs/core/langchain_core/callbacks/streaming_stdout.py at line 60.

Analyze Your Own Codebase

Get architecture documentation, dependency graphs, and domain analysis for your codebase in minutes.

Try Supermodel Free