Home / File/ fake_llm.py — langchain Source File

fake_llm.py — langchain Source File

Architecture documentation for fake_llm.py, a python file in the langchain codebase. 6 imports, 0 dependents.

Entity Profile

Dependency Diagram

graph LR
  0ec936bc_a546_aac1_192c_e8d47c63bd34["fake_llm.py"]
  cfe2bde5_180e_e3b0_df2b_55b3ebaca8e7["collections.abc"]
  0ec936bc_a546_aac1_192c_e8d47c63bd34 --> cfe2bde5_180e_e3b0_df2b_55b3ebaca8e7
  8e2034b7_ceb8_963f_29fc_2ea6b50ef9b3["typing"]
  0ec936bc_a546_aac1_192c_e8d47c63bd34 --> 8e2034b7_ceb8_963f_29fc_2ea6b50ef9b3
  e8ec017e_6c91_4b34_675f_2a96c5aa9be6["langchain_core.callbacks.manager"]
  0ec936bc_a546_aac1_192c_e8d47c63bd34 --> e8ec017e_6c91_4b34_675f_2a96c5aa9be6
  89934eed_a823_2184_acf2_039f48eed5f9["langchain_core.language_models.llms"]
  0ec936bc_a546_aac1_192c_e8d47c63bd34 --> 89934eed_a823_2184_acf2_039f48eed5f9
  6e58aaea_f08e_c099_3cc7_f9567bfb1ae7["pydantic"]
  0ec936bc_a546_aac1_192c_e8d47c63bd34 --> 6e58aaea_f08e_c099_3cc7_f9567bfb1ae7
  91721f45_4909_e489_8c1f_084f8bd87145["typing_extensions"]
  0ec936bc_a546_aac1_192c_e8d47c63bd34 --> 91721f45_4909_e489_8c1f_084f8bd87145
  style 0ec936bc_a546_aac1_192c_e8d47c63bd34 fill:#6366f1,stroke:#818cf8,color:#fff

Relationship Graph

Source Code

"""Fake LLM wrapper for testing purposes."""

from collections.abc import Mapping
from typing import Any, cast

from langchain_core.callbacks.manager import CallbackManagerForLLMRun
from langchain_core.language_models.llms import LLM
from pydantic import model_validator
from typing_extensions import override


class FakeLLM(LLM):
    """Fake LLM wrapper for testing purposes."""

    queries: Mapping | None = None
    sequential_responses: bool | None = False
    response_index: int = 0

    @model_validator(mode="before")
    @classmethod
    def check_queries_required(cls, values: dict) -> dict:
        if values.get("sequential_response") and not values.get("queries"):
            msg = "queries is required when sequential_response is set to True"
            raise ValueError(msg)
        return values

    def get_num_tokens(self, text: str) -> int:
        """Return number of tokens."""
        return len(text.split())

    @property
    def _llm_type(self) -> str:
        """Return type of llm."""
        return "fake"

    @override
    def _call(
        self,
        prompt: str,
        stop: list[str] | None = None,
        run_manager: CallbackManagerForLLMRun | None = None,
        **kwargs: Any,
    ) -> str:
        if self.sequential_responses:
            return self._get_next_response_in_sequence
        if self.queries is not None:
            return self.queries[prompt]
        if stop is None:
            return "foo"
        return "bar"

    @property
    def _identifying_params(self) -> dict[str, Any]:
        return {}

    @property
    def _get_next_response_in_sequence(self) -> str:
        queries = cast("Mapping", self.queries)
        response = queries[list(queries.keys())[self.response_index]]
        self.response_index = self.response_index + 1
        return response

Subdomains

Classes

Dependencies

  • collections.abc
  • langchain_core.callbacks.manager
  • langchain_core.language_models.llms
  • pydantic
  • typing
  • typing_extensions

Frequently Asked Questions

What does fake_llm.py do?
fake_llm.py is a source file in the langchain codebase, written in python. It belongs to the CoreAbstractions domain, MessageSchema subdomain.
What does fake_llm.py depend on?
fake_llm.py imports 6 module(s): collections.abc, langchain_core.callbacks.manager, langchain_core.language_models.llms, pydantic, typing, typing_extensions.
Where is fake_llm.py in the architecture?
fake_llm.py is located at libs/langchain/tests/unit_tests/llms/fake_llm.py (domain: CoreAbstractions, subdomain: MessageSchema, directory: libs/langchain/tests/unit_tests/llms).

Analyze Your Own Codebase

Get architecture documentation, dependency graphs, and domain analysis for your codebase in minutes.

Try Supermodel Free