Home / Function/ llm_prefix() — langchain Function Reference

llm_prefix() — langchain Function Reference

Architecture documentation for the llm_prefix() function in base.py from the langchain codebase.

Entity Profile

Dependency Diagram

graph TD
  20920d2d_6d29_575b_75f2_c1a8f1026002["llm_prefix()"]
  d6865578_75e6_5985_062f_e7681a6f1acf["ConversationalChatAgent"]
  20920d2d_6d29_575b_75f2_c1a8f1026002 -->|defined in| d6865578_75e6_5985_062f_e7681a6f1acf
  style 20920d2d_6d29_575b_75f2_c1a8f1026002 fill:#6366f1,stroke:#818cf8,color:#fff

Relationship Graph

Source Code

libs/langchain/langchain_classic/agents/conversational_chat/base.py lines 64–70

    def llm_prefix(self) -> str:
        """Prefix to append the llm call with.

        Returns:
            "Thought: "
        """
        return "Thought:"

Subdomains

Frequently Asked Questions

What does llm_prefix() do?
llm_prefix() is a function in the langchain codebase, defined in libs/langchain/langchain_classic/agents/conversational_chat/base.py.
Where is llm_prefix() defined?
llm_prefix() is defined in libs/langchain/langchain_classic/agents/conversational_chat/base.py at line 64.

Analyze Your Own Codebase

Get architecture documentation, dependency graphs, and domain analysis for your codebase in minutes.

Try Supermodel Free