id
stringlengths
14
16
text
stringlengths
13
2.7k
source
stringlengths
57
178
306f625d40ef-0
langchain.utils.strings.comma_list¶ langchain.utils.strings.comma_list(items: List[Any]) → str[source]¶ Convert a list to a comma-separated string.
lang/api.python.langchain.com/en/latest/utils/langchain.utils.strings.comma_list.html
a5734605922a-0
langchain.utils.openai_functions.convert_pydantic_to_openai_tool¶ langchain.utils.openai_functions.convert_pydantic_to_openai_tool(model: Type[BaseModel], *, name: Optional[str] = None, description: Optional[str] = None) → ToolDescription[source]¶ Converts a Pydantic model to a function description for the OpenAI API.
lang/api.python.langchain.com/en/latest/utils/langchain.utils.openai_functions.convert_pydantic_to_openai_tool.html
dab95cb0d62f-0
langchain.utils.utils.guard_import¶ langchain.utils.utils.guard_import(module_name: str, *, pip_name: Optional[str] = None, package: Optional[str] = None) → Any[source]¶ Dynamically imports a module and raises a helpful exception if the module is not installed.
lang/api.python.langchain.com/en/latest/utils/langchain.utils.utils.guard_import.html
f0da86dcdacf-0
langchain.utils.utils.build_extra_kwargs¶ langchain.utils.utils.build_extra_kwargs(extra_kwargs: Dict[str, Any], values: Dict[str, Any], all_required_field_names: Set[str]) → Dict[str, Any][source]¶ Build extra kwargs from values and extra_kwargs. Parameters extra_kwargs – Extra kwargs passed in by user. values – Value...
lang/api.python.langchain.com/en/latest/utils/langchain.utils.utils.build_extra_kwargs.html
5a87f39023f1-0
langchain.utils.openai_functions.FunctionDescription¶ class langchain.utils.openai_functions.FunctionDescription[source]¶ Representation of a callable function to the OpenAI API. name: str¶ The name of the function. description: str¶ A description of the function. parameters: dict¶ The parameters of the function.
lang/api.python.langchain.com/en/latest/utils/langchain.utils.openai_functions.FunctionDescription.html
9fde07c69f2f-0
langchain.utils.input.get_bolded_text¶ langchain.utils.input.get_bolded_text(text: str) → str[source]¶ Get bolded text.
lang/api.python.langchain.com/en/latest/utils/langchain.utils.input.get_bolded_text.html
3b462b59c626-0
langchain.utils.strings.stringify_value¶ langchain.utils.strings.stringify_value(val: Any) → str[source]¶ Stringify a value. Parameters val – The value to stringify. Returns The stringified value. Return type str
lang/api.python.langchain.com/en/latest/utils/langchain.utils.strings.stringify_value.html
cba182e29ffc-0
langchain.utils.utils.get_pydantic_field_names¶ langchain.utils.utils.get_pydantic_field_names(pydantic_cls: Any) → Set[str][source]¶ Get field names, including aliases, for a pydantic class. Parameters pydantic_cls – Pydantic class.
lang/api.python.langchain.com/en/latest/utils/langchain.utils.utils.get_pydantic_field_names.html
d40e8ea5d7e6-0
langchain.utils.utils.mock_now¶ langchain.utils.utils.mock_now(dt_value)[source]¶ Context manager for mocking out datetime.now() in unit tests. Example: with mock_now(datetime.datetime(2011, 2, 3, 10, 11)): assert datetime.datetime.now() == datetime.datetime(2011, 2, 3, 10, 11)
lang/api.python.langchain.com/en/latest/utils/langchain.utils.utils.mock_now.html
ccda1b495f38-0
langchain.utils.pydantic.get_pydantic_major_version¶ langchain.utils.pydantic.get_pydantic_major_version() → int[source]¶ Get the major version of Pydantic.
lang/api.python.langchain.com/en/latest/utils/langchain.utils.pydantic.get_pydantic_major_version.html
ce7f7f94e89b-0
langchain.utils.env.get_from_dict_or_env¶ langchain.utils.env.get_from_dict_or_env(data: Dict[str, Any], key: str, env_key: str, default: Optional[str] = None) → str[source]¶ Get a value from a dictionary or an environment variable.
lang/api.python.langchain.com/en/latest/utils/langchain.utils.env.get_from_dict_or_env.html
2be8f6118a20-0
langchain.utils.iter.Tee¶ class langchain.utils.iter.Tee(iterable: Iterator[T], n: int = 2, *, lock: Optional[ContextManager[Any]] = None)[source]¶ Create n separate asynchronous iterators over iterable This splits a single iterable into multiple iterators, each providing the same items in the same order. All child ite...
lang/api.python.langchain.com/en/latest/utils/langchain.utils.iter.Tee.html
2be8f6118a20-1
and access is automatically synchronised. Methods __init__(iterable[, n, lock]) close() __init__(iterable: Iterator[T], n: int = 2, *, lock: Optional[ContextManager[Any]] = None)[source]¶ close() → None[source]¶
lang/api.python.langchain.com/en/latest/utils/langchain.utils.iter.Tee.html
813eb5c38092-0
langchain.utils.json_schema.dereference_refs¶ langchain.utils.json_schema.dereference_refs(schema_obj: dict, *, full_schema: Optional[dict] = None, skip_keys: Optional[Sequence[str]] = None) → dict[source]¶ Try to substitute $refs in JSON Schema.
lang/api.python.langchain.com/en/latest/utils/langchain.utils.json_schema.dereference_refs.html
7ecc30160d32-0
langchain.utils.utils.convert_to_secret_str¶ langchain.utils.utils.convert_to_secret_str(value: Union[SecretStr, str]) → SecretStr[source]¶ Convert a string to a SecretStr if needed.
lang/api.python.langchain.com/en/latest/utils/langchain.utils.utils.convert_to_secret_str.html
0ab8db6e2340-0
langchain.utils.aiter.Tee¶ class langchain.utils.aiter.Tee(iterable: AsyncIterator[T], n: int = 2, *, lock: Optional[AsyncContextManager[Any]] = None)[source]¶ Create n separate asynchronous iterators over iterable This splits a single iterable into multiple iterators, each providing the same items in the same order. A...
lang/api.python.langchain.com/en/latest/utils/langchain.utils.aiter.Tee.html
0ab8db6e2340-1
and access is automatically synchronised. Methods __init__(iterable[, n, lock]) aclose() __init__(iterable: AsyncIterator[T], n: int = 2, *, lock: Optional[AsyncContextManager[Any]] = None)[source]¶ async aclose() → None[source]¶
lang/api.python.langchain.com/en/latest/utils/langchain.utils.aiter.Tee.html
1bdf2871d948-0
langchain.utils.utils.xor_args¶ langchain.utils.utils.xor_args(*arg_groups: Tuple[str, ...]) → Callable[source]¶ Validate specified keyword args are mutually exclusive.
lang/api.python.langchain.com/en/latest/utils/langchain.utils.utils.xor_args.html
ea6aee6442f2-0
langchain.utils.iter.tee_peer¶ langchain.utils.iter.tee_peer(iterator: Iterator[T], buffer: Deque[T], peers: List[Deque[T]], lock: ContextManager[Any]) → Generator[T, None, None][source]¶ An individual iterator of a tee()
lang/api.python.langchain.com/en/latest/utils/langchain.utils.iter.tee_peer.html
d34702f3c4de-0
langchain.utils.html.extract_sub_links¶ langchain.utils.html.extract_sub_links(raw_html: str, url: str, *, base_url: Optional[str] = None, pattern: Optional[Union[str, Pattern]] = None, prevent_outside: bool = True, exclude_prefixes: Sequence[str] = ()) → List[str][source]¶ Extract all links from a raw html string and ...
lang/api.python.langchain.com/en/latest/utils/langchain.utils.html.extract_sub_links.html
09400e3a2ad2-0
langchain.utils.aiter.NoLock¶ class langchain.utils.aiter.NoLock[source]¶ Dummy lock that provides the proper interface but no protection Methods __init__() __init__()¶
lang/api.python.langchain.com/en/latest/utils/langchain.utils.aiter.NoLock.html
138433fb33c6-0
langchain.utils.env.get_from_env¶ langchain.utils.env.get_from_env(key: str, env_key: str, default: Optional[str] = None) → str[source]¶ Get a value from a dictionary or an environment variable.
lang/api.python.langchain.com/en/latest/utils/langchain.utils.env.get_from_env.html
a677e3630830-0
langchain.utils.strings.stringify_dict¶ langchain.utils.strings.stringify_dict(data: dict) → str[source]¶ Stringify a dictionary. Parameters data – The dictionary to stringify. Returns The stringified dictionary. Return type str
lang/api.python.langchain.com/en/latest/utils/langchain.utils.strings.stringify_dict.html
75e498b90996-0
langchain.utils.html.find_all_links¶ langchain.utils.html.find_all_links(raw_html: str, *, pattern: Optional[Union[str, Pattern]] = None) → List[str][source]¶ Extract all links from a raw html string. Parameters raw_html – original html. pattern – Regex to use for extracting links from raw html. Returns all links Retur...
lang/api.python.langchain.com/en/latest/utils/langchain.utils.html.find_all_links.html
443e2b91a537-0
langchain.utils.openai_functions.convert_pydantic_to_openai_function¶ langchain.utils.openai_functions.convert_pydantic_to_openai_function(model: Type[BaseModel], *, name: Optional[str] = None, description: Optional[str] = None) → FunctionDescription[source]¶ Converts a Pydantic model to a function description for the ...
lang/api.python.langchain.com/en/latest/utils/langchain.utils.openai_functions.convert_pydantic_to_openai_function.html
1b16711d5af8-0
langchain.utils.iter.NoLock¶ class langchain.utils.iter.NoLock[source]¶ Dummy lock that provides the proper interface but no protection Methods __init__() __init__()¶
lang/api.python.langchain.com/en/latest/utils/langchain.utils.iter.NoLock.html
f2ae65cf1718-0
langchain.utils.loading.try_load_from_hub¶ langchain.utils.loading.try_load_from_hub(path: Union[str, Path], loader: Callable[[str], T], valid_prefix: str, valid_suffixes: Set[str], **kwargs: Any) → Optional[T][source]¶ Load configuration from hub. Returns None if path is not a hub path.
lang/api.python.langchain.com/en/latest/utils/langchain.utils.loading.try_load_from_hub.html
61ccf7c83c48-0
langchain.utils.openai.is_openai_v1¶ langchain.utils.openai.is_openai_v1() → bool[source]¶
lang/api.python.langchain.com/en/latest/utils/langchain.utils.openai.is_openai_v1.html
48d7096c5a72-0
langchain.utils.aiter.tee_peer¶ async langchain.utils.aiter.tee_peer(iterator: AsyncIterator[T], buffer: Deque[T], peers: List[Deque[T]], lock: AsyncContextManager[Any]) → AsyncGenerator[T, None][source]¶ An individual iterator of a tee()
lang/api.python.langchain.com/en/latest/utils/langchain.utils.aiter.tee_peer.html
abfdb7248744-0
langchain.utils.aiter.atee¶ langchain.utils.aiter.atee¶ alias of Tee
lang/api.python.langchain.com/en/latest/utils/langchain.utils.aiter.atee.html
84f57ab63f9a-0
langchain.utils.math.cosine_similarity¶ langchain.utils.math.cosine_similarity(X: Union[List[List[float]], List[ndarray], ndarray], Y: Union[List[List[float]], List[ndarray], ndarray]) → ndarray[source]¶ Row-wise cosine similarity between two equal-width matrices.
lang/api.python.langchain.com/en/latest/utils/langchain.utils.math.cosine_similarity.html
f0067e92cc31-0
langchain.utils.iter.safetee¶ langchain.utils.iter.safetee¶ alias of Tee
lang/api.python.langchain.com/en/latest/utils/langchain.utils.iter.safetee.html
c90d182290c9-0
langchain.utils.input.get_colored_text¶ langchain.utils.input.get_colored_text(text: str, color: str) → str[source]¶ Get colored text.
lang/api.python.langchain.com/en/latest/utils/langchain.utils.input.get_colored_text.html
98efc4794f9b-0
langchain.utils.input.print_text¶ langchain.utils.input.print_text(text: str, color: Optional[str] = None, end: str = '', file: Optional[TextIO] = None) → None[source]¶ Print text with highlighting and no end characters.
lang/api.python.langchain.com/en/latest/utils/langchain.utils.input.print_text.html
1366b40eeaec-0
langchain.utils.iter.batch_iterate¶ langchain.utils.iter.batch_iterate(size: int, iterable: Iterable[T]) → Iterator[List[T]][source]¶ Utility batching function.
lang/api.python.langchain.com/en/latest/utils/langchain.utils.iter.batch_iterate.html
89d6ab0dddc0-0
langchain.utils.utils.check_package_version¶ langchain.utils.utils.check_package_version(package: str, lt_version: Optional[str] = None, lte_version: Optional[str] = None, gt_version: Optional[str] = None, gte_version: Optional[str] = None) → None[source]¶ Check the version of a package.
lang/api.python.langchain.com/en/latest/utils/langchain.utils.utils.check_package_version.html
fa20f8b34455-0
langchain.utils.formatting.StrictFormatter¶ class langchain.utils.formatting.StrictFormatter[source]¶ A subclass of formatter that checks for extra keys. Methods __init__() check_unused_args(used_args, args, kwargs) Check to see if extra parameters are passed. convert_field(value, conversion) format(format_string, /, *...
lang/api.python.langchain.com/en/latest/utils/langchain.utils.formatting.StrictFormatter.html
087f4c446b85-0
langchain.utils.aiter.py_anext¶ langchain.utils.aiter.py_anext(iterator: ~typing.AsyncIterator[~langchain.utils.aiter.T], default: ~typing.Union[~langchain.utils.aiter.T, ~typing.Any] = <object object>) → Awaitable[Union[T, None, Any]][source]¶ Pure-Python implementation of anext() for testing purposes. Closely matches...
lang/api.python.langchain.com/en/latest/utils/langchain.utils.aiter.py_anext.html
8a99f3e0050f-0
langchain.utils.utils.raise_for_status_with_text¶ langchain.utils.utils.raise_for_status_with_text(response: Response) → None[source]¶ Raise an error with the response text.
lang/api.python.langchain.com/en/latest/utils/langchain.utils.utils.raise_for_status_with_text.html
0b8a206d35c0-0
langchain.utils.input.get_color_mapping¶ langchain.utils.input.get_color_mapping(items: List[str], excluded_colors: Optional[List] = None) → Dict[str, str][source]¶ Get mapping for items to a support color.
lang/api.python.langchain.com/en/latest/utils/langchain.utils.input.get_color_mapping.html
75531ebdd941-0
langchain.utils.openai_functions.ToolDescription¶ class langchain.utils.openai_functions.ToolDescription[source]¶ Representation of a callable function to the OpenAI API. type: Literal['function']¶ function: langchain.utils.openai_functions.FunctionDescription¶
lang/api.python.langchain.com/en/latest/utils/langchain.utils.openai_functions.ToolDescription.html
af6a19faef1b-0
langchain_experimental.tabular_synthetic_data.base.SyntheticDataGenerator¶ class langchain_experimental.tabular_synthetic_data.base.SyntheticDataGenerator[source]¶ Bases: BaseModel Generates synthetic data using the given LLM and few-shot template. Utilizes the provided LLM to produce synthetic data based on the few-sh...
lang/api.python.langchain.com/en/latest/tabular_synthetic_data/langchain_experimental.tabular_synthetic_data.base.SyntheticDataGenerator.html
af6a19faef1b-1
the “extra” keyword argument. Parameters subject (str) – The subject the synthetic data will be about. runs (int) – Number of times to generate the data asynchronously. extra (str) – Extra instructions for steerability in data generation. Returns List of generated synthetic data for the given subject. Return type List[...
lang/api.python.langchain.com/en/latest/tabular_synthetic_data/langchain_experimental.tabular_synthetic_data.base.SyntheticDataGenerator.html
af6a19faef1b-2
deep – set to True to make a deep copy of the model Returns new model instance dict(*, include: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, exclude: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, by_alias: bool = False, skip_defaults: Optional[bool] = None, exclude_unset: bool = False, ex...
lang/api.python.langchain.com/en/latest/tabular_synthetic_data/langchain_experimental.tabular_synthetic_data.base.SyntheticDataGenerator.html
af6a19faef1b-3
Generate a JSON representation of the model, include and exclude arguments as per dict(). encoder is an optional function to supply as default to json.dumps(), other arguments as per json.dumps(). classmethod parse_file(path: Union[str, Path], *, content_type: unicode = None, encoding: unicode = 'utf8', proto: Protocol...
lang/api.python.langchain.com/en/latest/tabular_synthetic_data/langchain_experimental.tabular_synthetic_data.base.SyntheticDataGenerator.html
8f0f6ccff58c-0
langchain_experimental.tabular_synthetic_data.openai.create_openai_data_generator¶ langchain_experimental.tabular_synthetic_data.openai.create_openai_data_generator(output_schema: Union[Dict[str, Any], Type[BaseModel]], llm: ChatOpenAI, prompt: BasePromptTemplate, output_parser: Optional[BaseLLMOutputParser] = None, **...
lang/api.python.langchain.com/en/latest/tabular_synthetic_data/langchain_experimental.tabular_synthetic_data.openai.create_openai_data_generator.html
0f45bcde4ab6-0
langchain.adapters.openai.ChatCompletion¶ class langchain.adapters.openai.ChatCompletion[source]¶ Chat completion. Methods __init__() acreate() create() __init__()¶ async static acreate(messages: Sequence[Dict[str, Any]], *, provider: str = "'ChatOpenAI'", stream: Literal[False] = 'False', **kwargs: Any) → dict[source]...
lang/api.python.langchain.com/en/latest/adapters/langchain.adapters.openai.ChatCompletion.html
36b675d44581-0
langchain.adapters.openai.convert_openai_messages¶ langchain.adapters.openai.convert_openai_messages(messages: Sequence[Dict[str, Any]]) → List[BaseMessage][source]¶ Convert dictionaries representing OpenAI messages to LangChain format. Parameters messages – List of dictionaries representing OpenAI messages Returns Lis...
lang/api.python.langchain.com/en/latest/adapters/langchain.adapters.openai.convert_openai_messages.html
09e9dc5a00d3-0
langchain.adapters.openai.aenumerate¶ async langchain.adapters.openai.aenumerate(iterable: AsyncIterator[Any], start: int = 0) → AsyncIterator[tuple[int, Any]][source]¶ Async version of enumerate function.
lang/api.python.langchain.com/en/latest/adapters/langchain.adapters.openai.aenumerate.html
27f3a88ba7f3-0
langchain.adapters.openai.convert_message_to_dict¶ langchain.adapters.openai.convert_message_to_dict(message: BaseMessage) → dict[source]¶ Convert a LangChain message to a dictionary. Parameters message – The LangChain message. Returns The dictionary. Examples using convert_message_to_dict¶ Twitter (via Apify)
lang/api.python.langchain.com/en/latest/adapters/langchain.adapters.openai.convert_message_to_dict.html
c499d7d7dfcd-0
langchain.adapters.openai.convert_dict_to_message¶ langchain.adapters.openai.convert_dict_to_message(_dict: Mapping[str, Any]) → BaseMessage[source]¶ Convert a dictionary to a LangChain message. Parameters _dict – The dictionary. Returns The LangChain message.
lang/api.python.langchain.com/en/latest/adapters/langchain.adapters.openai.convert_dict_to_message.html
4a6eb603e9e4-0
langchain.adapters.openai.convert_messages_for_finetuning¶ langchain.adapters.openai.convert_messages_for_finetuning(sessions: Iterable[ChatSession]) → List[List[dict]][source]¶ Convert messages to a list of lists of dictionaries for fine-tuning. Parameters sessions – The chat sessions. Returns The list of lists of dic...
lang/api.python.langchain.com/en/latest/adapters/langchain.adapters.openai.convert_messages_for_finetuning.html
f2b5823127a0-0
langchain.document_loaders.modern_treasury.ModernTreasuryLoader¶ class langchain.document_loaders.modern_treasury.ModernTreasuryLoader(resource: str, organization_id: Optional[str] = None, api_key: Optional[str] = None)[source]¶ Load from Modern Treasury. Parameters resource – The Modern Treasury resource to load. orga...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.modern_treasury.ModernTreasuryLoader.html
f2b5823127a0-1
Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using ModernTreasuryLoader¶ Modern Treasury
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.modern_treasury.ModernTreasuryLoader.html
f3932bf51abe-0
langchain.document_loaders.docusaurus.DocusaurusLoader¶ class langchain.document_loaders.docusaurus.DocusaurusLoader(url: str, custom_html_tags: Optional[List[str]] = None, **kwargs: Any)[source]¶ Loader that leverages the SitemapLoader to loop through the generated pages of a Docusaurus Documentation website and extra...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.docusaurus.DocusaurusLoader.html
f3932bf51abe-1
Parse sitemap xml and load into a list of dicts. scrape([parser]) Scrape data from webpage and return it in BeautifulSoup format. scrape_all(urls[, parser]) Fetch all urls, then return soups for all results. __init__(url: str, custom_html_tags: Optional[List[str]] = None, **kwargs: Any)[source]¶ Initialize DocusaurusLo...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.docusaurus.DocusaurusLoader.html
f3932bf51abe-2
Fetch all urls, then return soups for all results.
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.docusaurus.DocusaurusLoader.html
5bf5ec6ead27-0
langchain.document_loaders.twitter.TwitterTweetLoader¶ class langchain.document_loaders.twitter.TwitterTweetLoader(auth_handler: Union[OAuthHandler, OAuth2BearerHandler], twitter_users: Sequence[str], number_tweets: Optional[int] = 100)[source]¶ Load Twitter tweets. Read tweets of the user’s Twitter handle. First you n...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.twitter.TwitterTweetLoader.html
5bf5ec6ead27-1
load() → List[Document][source]¶ Load tweets. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.twitter.TwitterTweetLoader.html
8270e2e39ded-0
langchain.document_loaders.json_loader.JSONLoader¶ class langchain.document_loaders.json_loader.JSONLoader(file_path: Union[str, Path], jq_schema: str, content_key: Optional[str] = None, metadata_func: Optional[Callable[[Dict, Dict], Dict]] = None, text_content: bool = True, json_lines: bool = False)[source]¶ Load a JS...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.json_loader.JSONLoader.html
8270e2e39ded-1
load_and_split([text_splitter]) Load Documents and split into chunks. __init__(file_path: Union[str, Path], jq_schema: str, content_key: Optional[str] = None, metadata_func: Optional[Callable[[Dict, Dict], Dict]] = None, text_content: bool = True, json_lines: bool = False)[source]¶ Initialize the JSONLoader. Parameters...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.json_loader.JSONLoader.html
5a5379be6863-0
langchain.document_loaders.evernote.EverNoteLoader¶ class langchain.document_loaders.evernote.EverNoteLoader(file_path: str, load_single_document: bool = True)[source]¶ Load from EverNote. Loads an EverNote notebook export file e.g. my_notebook.enex into Documents. Instructions on producing this file can be found at ht...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.evernote.EverNoteLoader.html
5a5379be6863-1
Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using EverNoteLoader¶ EverNote
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.evernote.EverNoteLoader.html
4f3dd7d76eb9-0
langchain.document_loaders.word_document.UnstructuredWordDocumentLoader¶ class langchain.document_loaders.word_document.UnstructuredWordDocumentLoader(file_path: Union[str, List[str]], mode: str = 'single', **unstructured_kwargs: Any)[source]¶ Load Microsoft Word file using Unstructured. Works with both .docx and .doc ...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.word_document.UnstructuredWordDocumentLoader.html
4f3dd7d76eb9-1
Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using UnstructuredWordDocumentLoader¶ Microsoft Word
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.word_document.UnstructuredWordDocumentLoader.html
6a49c6183b64-0
langchain.document_loaders.chromium.AsyncChromiumLoader¶ class langchain.document_loaders.chromium.AsyncChromiumLoader(urls: List[str])[source]¶ Scrape HTML pages from URLs using a headless instance of the Chromium. Initialize the loader with a list of URL paths. Parameters urls (List[str]) – A list of URLs to scrape c...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.chromium.AsyncChromiumLoader.html
6a49c6183b64-1
Load and return all Documents from the provided URLs. Returns A list of Document objects containing the scraped content from each URL. Return type List[Document] load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Paramet...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.chromium.AsyncChromiumLoader.html
0de79a96b404-0
langchain.document_loaders.parsers.language.cobol.CobolSegmenter¶ class langchain.document_loaders.parsers.language.cobol.CobolSegmenter(code: str)[source]¶ Code segmenter for COBOL. Attributes DIVISION_PATTERN PARAGRAPH_PATTERN SECTION_PATTERN Methods __init__(code) extract_functions_classes() is_valid() simplify_code...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.parsers.language.cobol.CobolSegmenter.html
814670f5fe92-0
langchain.document_loaders.notion.NotionDirectoryLoader¶ class langchain.document_loaders.notion.NotionDirectoryLoader(path: str, *, encoding: str = 'utf-8')[source]¶ Load Notion directory dump. Initialize with a file path. Methods __init__(path, *[, encoding]) Initialize with a file path. lazy_load() A lazy loader for...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.notion.NotionDirectoryLoader.html
b29ef58c3d16-0
langchain.document_loaders.rocksetdb.RocksetLoader¶ class langchain.document_loaders.rocksetdb.RocksetLoader(client: ~typing.Any, query: ~typing.Any, content_keys: ~typing.List[str], metadata_keys: ~typing.Optional[~typing.List[str]] = None, content_columns_joiner: ~typing.Callable[[~typing.List[~typing.Tuple[str, ~typ...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.rocksetdb.RocksetLoader.html
b29ef58c3d16-1
line. This method is only relevant if there are multiple content_keys. Methods __init__(client, query, content_keys[, ...]) Initialize with Rockset client. lazy_load() A lazy loader for Documents. load() Load data into Document objects. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(clie...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.rocksetdb.RocksetLoader.html
b29ef58c3d16-2
Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using RocksetLoader¶ Rockset
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.rocksetdb.RocksetLoader.html
7a4cfe196fb7-0
langchain.document_loaders.acreom.AcreomLoader¶ class langchain.document_loaders.acreom.AcreomLoader(path: str, encoding: str = 'UTF-8', collect_metadata: bool = True)[source]¶ Load acreom vault from a directory. Initialize the loader. Attributes FRONT_MATTER_REGEX Regex to match front matter metadata in markdown files...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.acreom.AcreomLoader.html
897c712d1597-0
langchain.document_loaders.onedrive_file.OneDriveFileLoader¶ class langchain.document_loaders.onedrive_file.OneDriveFileLoader[source]¶ Bases: BaseLoader, BaseModel Load a file from Microsoft OneDrive. Create a new model by parsing and validating input data from keyword arguments. Raises ValidationError if the input da...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.onedrive_file.OneDriveFileLoader.html
897c712d1597-1
deep – set to True to make a deep copy of the model Returns new model instance dict(*, include: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, exclude: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, by_alias: bool = False, skip_defaults: Optional[bool] = None, exclude_unset: bool = False, ex...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.onedrive_file.OneDriveFileLoader.html
897c712d1597-2
Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. classmethod parse_file(path: Union[str, Path], *, content_type: unicode = None, encoding: unicode = 'utf8', proto: Protocol = None, allow_pickle: bool = False) → Model¶ classmethod parse_obj(obj: Any) → Model¶ classmethod parse_raw(b: Union[str, byt...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.onedrive_file.OneDriveFileLoader.html
f1ffabac1ec4-0
langchain.document_loaders.chatgpt.ChatGPTLoader¶ class langchain.document_loaders.chatgpt.ChatGPTLoader(log_file: str, num_logs: int = - 1)[source]¶ Load conversations from exported ChatGPT data. Initialize a class object. Parameters log_file – Path to the log file num_logs – Number of logs to load. If 0, load all log...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.chatgpt.ChatGPTLoader.html
23c4aeee6af5-0
langchain.document_loaders.hugging_face_dataset.HuggingFaceDatasetLoader¶ class langchain.document_loaders.hugging_face_dataset.HuggingFaceDatasetLoader(path: str, page_content_column: str = 'text', name: Optional[str] = None, data_dir: Optional[str] = None, data_files: Optional[Union[str, Sequence[str], Mapping[str, U...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.hugging_face_dataset.HuggingFaceDatasetLoader.html
23c4aeee6af5-1
load() Load documents. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(path: str, page_content_column: str = 'text', name: Optional[str] = None, data_dir: Optional[str] = None, data_files: Optional[Union[str, Sequence[str], Mapping[str, Union[str, Sequence[str]]]]] = None, cache_dir: Opti...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.hugging_face_dataset.HuggingFaceDatasetLoader.html
23c4aeee6af5-2
Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using HuggingFaceDatasetLoader¶ HuggingFace dataset
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.hugging_face_dataset.HuggingFaceDatasetLoader.html
6145247f66c1-0
langchain.document_loaders.embaas.BaseEmbaasLoader¶ class langchain.document_loaders.embaas.BaseEmbaasLoader[source]¶ Bases: BaseModel Base loader for Embaas document extraction API. Create a new model by parsing and validating input data from keyword arguments. Raises ValidationError if the input data cannot be parsed...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.embaas.BaseEmbaasLoader.html
6145247f66c1-1
the new model: you should trust this data deep – set to True to make a deep copy of the model Returns new model instance dict(*, include: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, exclude: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, by_alias: bool = False, skip_defaults: Optional[boo...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.embaas.BaseEmbaasLoader.html
6145247f66c1-2
classmethod schema(by_alias: bool = True, ref_template: unicode = '#/definitions/{model}') → DictStrAny¶ classmethod schema_json(*, by_alias: bool = True, ref_template: unicode = '#/definitions/{model}', **dumps_kwargs: Any) → unicode¶ classmethod update_forward_refs(**localns: Any) → None¶ Try to update ForwardRefs on...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.embaas.BaseEmbaasLoader.html
ceee09be4ddd-0
langchain.document_loaders.airbyte.AirbyteShopifyLoader¶ class langchain.document_loaders.airbyte.AirbyteShopifyLoader(config: Mapping[str, Any], stream_name: str, record_handler: Optional[Callable[[Any, Optional[str]], Document]] = None, state: Optional[Any] = None)[source]¶ Load from Shopify using an Airbyte source c...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.airbyte.AirbyteShopifyLoader.html
ceee09be4ddd-1
load() → List[Document]¶ Load data into Document objects. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacter...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.airbyte.AirbyteShopifyLoader.html
e07b962197e9-0
langchain.document_loaders.geodataframe.GeoDataFrameLoader¶ class langchain.document_loaders.geodataframe.GeoDataFrameLoader(data_frame: Any, page_content_column: str = 'geometry')[source]¶ Load geopandas Dataframe. Initialize with geopandas Dataframe. Parameters data_frame – geopandas DataFrame object. page_content_co...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.geodataframe.GeoDataFrameLoader.html
7dff003674bd-0
langchain.document_loaders.onedrive.OneDriveLoader¶ class langchain.document_loaders.onedrive.OneDriveLoader[source]¶ Bases: O365BaseLoader Load from Microsoft OneDrive. Create a new model by parsing and validating input data from keyword arguments. Raises ValidationError if the input data cannot be parsed to form a va...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.onedrive.OneDriveLoader.html
7dff003674bd-1
Parameters include – fields to include in new model exclude – fields to exclude from new model, as with values this takes precedence over include update – values to change/add in the new model. Note: the data is not validated before creating the new model: you should trust this data deep – set to True to make a deep co...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.onedrive.OneDriveLoader.html
7dff003674bd-2
load() → List[Document][source]¶ Load all documents. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextS...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.onedrive.OneDriveLoader.html
6848344c7eec-0
langchain.document_loaders.googledrive.GoogleDriveLoader¶ class langchain.document_loaders.googledrive.GoogleDriveLoader[source]¶ Bases: BaseLoader, BaseModel Load Google Docs from Google Drive. Create a new model by parsing and validating input data from keyword arguments. Raises ValidationError if the input data cann...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.googledrive.GoogleDriveLoader.html
6848344c7eec-1
Default values are respected, but no other validation is performed. Behaves as if Config.extra = ‘allow’ was set since it adds all passed values copy(*, include: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, exclude: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, update: Optional[DictStrAny...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.googledrive.GoogleDriveLoader.html
6848344c7eec-2
classmethod from_orm(obj: Any) → Model¶ json(*, include: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, exclude: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, by_alias: bool = False, skip_defaults: Optional[bool] = None, exclude_unset: bool = False, exclude_defaults: bool = False, exclude_n...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.googledrive.GoogleDriveLoader.html
6848344c7eec-3
classmethod schema_json(*, by_alias: bool = True, ref_template: unicode = '#/definitions/{model}', **dumps_kwargs: Any) → unicode¶ classmethod update_forward_refs(**localns: Any) → None¶ Try to update ForwardRefs on fields based on this Model, globalns and localns. classmethod validate(value: Any) → Model¶ Examples usi...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.googledrive.GoogleDriveLoader.html
7afa020b2c42-0
langchain.document_loaders.apify_dataset.ApifyDatasetLoader¶ class langchain.document_loaders.apify_dataset.ApifyDatasetLoader[source]¶ Bases: BaseLoader, BaseModel Load datasets from Apify web scraping, crawling, and data extraction platform. For details, see https://docs.apify.com/platform/integrations/langchain Exam...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.apify_dataset.ApifyDatasetLoader.html
7afa020b2c42-1
Behaves as if Config.extra = ‘allow’ was set since it adds all passed values copy(*, include: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, exclude: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, update: Optional[DictStrAny] = None, deep: bool = False) → Model¶ Duplicate a model, optionally...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.apify_dataset.ApifyDatasetLoader.html
7afa020b2c42-2
classmethod from_orm(obj: Any) → Model¶ json(*, include: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, exclude: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, by_alias: bool = False, skip_defaults: Optional[bool] = None, exclude_unset: bool = False, exclude_defaults: bool = False, exclude_n...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.apify_dataset.ApifyDatasetLoader.html
7afa020b2c42-3
classmethod schema_json(*, by_alias: bool = True, ref_template: unicode = '#/definitions/{model}', **dumps_kwargs: Any) → unicode¶ classmethod update_forward_refs(**localns: Any) → None¶ Try to update ForwardRefs on fields based on this Model, globalns and localns. classmethod validate(value: Any) → Model¶ Examples usi...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.apify_dataset.ApifyDatasetLoader.html
a42b73913211-0
langchain.document_loaders.github.GitHubIssuesLoader¶ class langchain.document_loaders.github.GitHubIssuesLoader[source]¶ Bases: BaseGitHubLoader Load issues of a GitHub repository. Create a new model by parsing and validating input data from keyword arguments. Raises ValidationError if the input data cannot be parsed ...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.github.GitHubIssuesLoader.html
a42b73913211-1
param since: Optional[str] = None¶ Only show notifications updated after the given time. This is a timestamp in ISO 8601 format: YYYY-MM-DDTHH:MM:SSZ. param sort: Optional[Literal['created', 'updated', 'comments']] = None¶ What to sort results by. Can be one of: ‘created’, ‘updated’, ‘comments’. Default is ‘created’. p...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.github.GitHubIssuesLoader.html
a42b73913211-2
deep – set to True to make a deep copy of the model Returns new model instance dict(*, include: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, exclude: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, by_alias: bool = False, skip_defaults: Optional[bool] = None, exclude_unset: bool = False, ex...
lang/api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.github.GitHubIssuesLoader.html