fake_llm.py — langchain Source File
Architecture documentation for fake_llm.py, a python file in the langchain codebase. 6 imports, 0 dependents.
Entity Profile
Dependency Diagram
graph LR 0ec936bc_a546_aac1_192c_e8d47c63bd34["fake_llm.py"] cfe2bde5_180e_e3b0_df2b_55b3ebaca8e7["collections.abc"] 0ec936bc_a546_aac1_192c_e8d47c63bd34 --> cfe2bde5_180e_e3b0_df2b_55b3ebaca8e7 8e2034b7_ceb8_963f_29fc_2ea6b50ef9b3["typing"] 0ec936bc_a546_aac1_192c_e8d47c63bd34 --> 8e2034b7_ceb8_963f_29fc_2ea6b50ef9b3 e8ec017e_6c91_4b34_675f_2a96c5aa9be6["langchain_core.callbacks.manager"] 0ec936bc_a546_aac1_192c_e8d47c63bd34 --> e8ec017e_6c91_4b34_675f_2a96c5aa9be6 89934eed_a823_2184_acf2_039f48eed5f9["langchain_core.language_models.llms"] 0ec936bc_a546_aac1_192c_e8d47c63bd34 --> 89934eed_a823_2184_acf2_039f48eed5f9 6e58aaea_f08e_c099_3cc7_f9567bfb1ae7["pydantic"] 0ec936bc_a546_aac1_192c_e8d47c63bd34 --> 6e58aaea_f08e_c099_3cc7_f9567bfb1ae7 91721f45_4909_e489_8c1f_084f8bd87145["typing_extensions"] 0ec936bc_a546_aac1_192c_e8d47c63bd34 --> 91721f45_4909_e489_8c1f_084f8bd87145 style 0ec936bc_a546_aac1_192c_e8d47c63bd34 fill:#6366f1,stroke:#818cf8,color:#fff
Relationship Graph
Source Code
"""Fake LLM wrapper for testing purposes."""
from collections.abc import Mapping
from typing import Any, cast
from langchain_core.callbacks.manager import CallbackManagerForLLMRun
from langchain_core.language_models.llms import LLM
from pydantic import model_validator
from typing_extensions import override
class FakeLLM(LLM):
"""Fake LLM wrapper for testing purposes."""
queries: Mapping | None = None
sequential_responses: bool | None = False
response_index: int = 0
@model_validator(mode="before")
@classmethod
def check_queries_required(cls, values: dict) -> dict:
if values.get("sequential_response") and not values.get("queries"):
msg = "queries is required when sequential_response is set to True"
raise ValueError(msg)
return values
def get_num_tokens(self, text: str) -> int:
"""Return number of tokens."""
return len(text.split())
@property
def _llm_type(self) -> str:
"""Return type of llm."""
return "fake"
@override
def _call(
self,
prompt: str,
stop: list[str] | None = None,
run_manager: CallbackManagerForLLMRun | None = None,
**kwargs: Any,
) -> str:
if self.sequential_responses:
return self._get_next_response_in_sequence
if self.queries is not None:
return self.queries[prompt]
if stop is None:
return "foo"
return "bar"
@property
def _identifying_params(self) -> dict[str, Any]:
return {}
@property
def _get_next_response_in_sequence(self) -> str:
queries = cast("Mapping", self.queries)
response = queries[list(queries.keys())[self.response_index]]
self.response_index = self.response_index + 1
return response
Domain
Subdomains
Classes
Dependencies
- collections.abc
- langchain_core.callbacks.manager
- langchain_core.language_models.llms
- pydantic
- typing
- typing_extensions
Source
Frequently Asked Questions
What does fake_llm.py do?
fake_llm.py is a source file in the langchain codebase, written in python. It belongs to the CoreAbstractions domain, MessageSchema subdomain.
What does fake_llm.py depend on?
fake_llm.py imports 6 module(s): collections.abc, langchain_core.callbacks.manager, langchain_core.language_models.llms, pydantic, typing, typing_extensions.
Where is fake_llm.py in the architecture?
fake_llm.py is located at libs/langchain/tests/unit_tests/llms/fake_llm.py (domain: CoreAbstractions, subdomain: MessageSchema, directory: libs/langchain/tests/unit_tests/llms).
Analyze Your Own Codebase
Get architecture documentation, dependency graphs, and domain analysis for your codebase in minutes.
Try Supermodel Free