id
stringlengths
14
15
text
stringlengths
49
2.47k
source
stringlengths
61
166
d69fed5fe3a9-0
langchain.document_loaders.odt.UnstructuredODTLoader¶ class langchain.document_loaders.odt.UnstructuredODTLoader(file_path: str, mode: str = 'single', **unstructured_kwargs: Any)[source]¶ Loader that uses unstructured to load OpenOffice ODT files. You can run the loader in one of two modes: “single” and “elements”. If you use “single” mode, the document will be returned as a single langchain Document object. If you use “elements” mode, the unstructured library will split the document into elements such as Title and NarrativeText. You can pass in additional unstructured kwargs after mode to apply different unstructured settings. Examples from langchain.document_loaders import UnstructuredODTLoader loader = UnstructuredODTLoader(“example.odt”, mode=”elements”, strategy=”fast”, ) docs = loader.load() References https://unstructured-io.github.io/unstructured/bricks.html#partition-odt Parameters file_path – The path to the file to load. mode – The mode to use when loading the file. Can be one of “single”, “multi”, or “all”. Default is “single”. **unstructured_kwargs – Any kwargs to pass to the unstructured. Methods __init__(file_path[, mode]) param file_path The path to the file to load. lazy_load() A lazy loader for Documents. load() Load file. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(file_path: str, mode: str = 'single', **unstructured_kwargs: Any)[source]¶ Parameters file_path – The path to the file to load. mode – The mode to use when loading the file. Can be one of “single”,
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.odt.UnstructuredODTLoader.html
d69fed5fe3a9-1
mode – The mode to use when loading the file. Can be one of “single”, “multi”, or “all”. Default is “single”. **unstructured_kwargs – Any kwargs to pass to the unstructured. lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document]¶ Load file. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using UnstructuredODTLoader¶ Open Document Format (ODT)
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.odt.UnstructuredODTLoader.html
45b9b6547be4-0
langchain.document_loaders.rocksetdb.ColumnNotFoundError¶ class langchain.document_loaders.rocksetdb.ColumnNotFoundError(missing_key: str, query: str)[source]¶ Column not found error.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.rocksetdb.ColumnNotFoundError.html
a61cbf18b322-0
langchain.document_loaders.youtube.YoutubeLoader¶ class langchain.document_loaders.youtube.YoutubeLoader(video_id: str, add_video_info: bool = False, language: Union[str, Sequence[str]] = 'en', translation: str = 'en', continue_on_failure: bool = False)[source]¶ Loads Youtube transcripts. Initialize with YouTube video ID. Methods __init__(video_id[, add_video_info, ...]) Initialize with YouTube video ID. extract_video_id(youtube_url) Extract video id from common YT urls. from_youtube_url(youtube_url, **kwargs) Given youtube URL, load video. lazy_load() A lazy loader for Documents. load() Load documents. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(video_id: str, add_video_info: bool = False, language: Union[str, Sequence[str]] = 'en', translation: str = 'en', continue_on_failure: bool = False)[source]¶ Initialize with YouTube video ID. static extract_video_id(youtube_url: str) → str[source]¶ Extract video id from common YT urls. classmethod from_youtube_url(youtube_url: str, **kwargs: Any) → YoutubeLoader[source]¶ Given youtube URL, load video. lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document][source]¶ Load documents. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using YoutubeLoader¶ YouTube YouTube transcripts
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.youtube.YoutubeLoader.html
0025be90cba0-0
langchain.document_loaders.s3_directory.S3DirectoryLoader¶ class langchain.document_loaders.s3_directory.S3DirectoryLoader(bucket: str, prefix: str = '')[source]¶ Loading logic for loading documents from an AWS S3. Initialize with bucket and key name. Parameters bucket – The name of the S3 bucket. prefix – The prefix of the S3 key. Defaults to “”. Methods __init__(bucket[, prefix]) Initialize with bucket and key name. lazy_load() A lazy loader for Documents. load() Load documents. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(bucket: str, prefix: str = '')[source]¶ Initialize with bucket and key name. Parameters bucket – The name of the S3 bucket. prefix – The prefix of the S3 key. Defaults to “”. lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document][source]¶ Load documents. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using S3DirectoryLoader¶ AWS S3 Directory
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.s3_directory.S3DirectoryLoader.html
28010a374d55-0
langchain.document_loaders.twitter.TwitterTweetLoader¶ class langchain.document_loaders.twitter.TwitterTweetLoader(auth_handler: Union[OAuthHandler, OAuth2BearerHandler], twitter_users: Sequence[str], number_tweets: Optional[int] = 100)[source]¶ Twitter tweets loader. Read tweets of user twitter handle. First you need to go to https://developer.twitter.com/en/docs/twitter-api /getting-started/getting-access-to-the-twitter-api to get your token. And create a v2 version of the app. Methods __init__(auth_handler, twitter_users[, ...]) from_bearer_token(oauth2_bearer_token, ...) Create a TwitterTweetLoader from OAuth2 bearer token. from_secrets(access_token, ...[, number_tweets]) Create a TwitterTweetLoader from access tokens and secrets. lazy_load() A lazy loader for Documents. load() Load tweets. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(auth_handler: Union[OAuthHandler, OAuth2BearerHandler], twitter_users: Sequence[str], number_tweets: Optional[int] = 100)[source]¶ classmethod from_bearer_token(oauth2_bearer_token: str, twitter_users: Sequence[str], number_tweets: Optional[int] = 100) → TwitterTweetLoader[source]¶ Create a TwitterTweetLoader from OAuth2 bearer token. classmethod from_secrets(access_token: str, access_token_secret: str, consumer_key: str, consumer_secret: str, twitter_users: Sequence[str], number_tweets: Optional[int] = 100) → TwitterTweetLoader[source]¶ Create a TwitterTweetLoader from access tokens and secrets. lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document][source]¶ Load tweets.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.twitter.TwitterTweetLoader.html
28010a374d55-1
load() → List[Document][source]¶ Load tweets. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using TwitterTweetLoader¶ Twitter
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.twitter.TwitterTweetLoader.html
7fdd89a73dac-0
langchain.document_loaders.powerpoint.UnstructuredPowerPointLoader¶ class langchain.document_loaders.powerpoint.UnstructuredPowerPointLoader(file_path: Union[str, List[str]], mode: str = 'single', **unstructured_kwargs: Any)[source]¶ Loader that uses unstructured to load PowerPoint files. Works with both .ppt and .pptx files. You can run the loader in one of two modes: “single” and “elements”. If you use “single” mode, the document will be returned as a single langchain Document object. If you use “elements” mode, the unstructured library will split the document into elements such as Title and NarrativeText. You can pass in additional unstructured kwargs after mode to apply different unstructured settings. Examples from langchain.document_loaders import UnstructuredPowerPointLoader loader = UnstructuredPowerPointLoader(“example.pptx”, mode=”elements”, strategy=”fast”, ) docs = loader.load() References https://unstructured-io.github.io/unstructured/bricks.html#partition-pptx Initialize with file path. Methods __init__(file_path[, mode]) Initialize with file path. lazy_load() A lazy loader for Documents. load() Load file. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(file_path: Union[str, List[str]], mode: str = 'single', **unstructured_kwargs: Any)¶ Initialize with file path. lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document]¶ Load file. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.powerpoint.UnstructuredPowerPointLoader.html
7fdd89a73dac-1
Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using UnstructuredPowerPointLoader¶ Microsoft PowerPoint
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.powerpoint.UnstructuredPowerPointLoader.html
23f44904a546-0
langchain.document_loaders.iugu.IuguLoader¶ class langchain.document_loaders.iugu.IuguLoader(resource: str, api_token: Optional[str] = None)[source]¶ Loader that fetches data from IUGU. Initialize the IUGU resource. Parameters resource – The name of the resource to fetch. api_token – The IUGU API token to use. Methods __init__(resource[, api_token]) Initialize the IUGU resource. lazy_load() A lazy loader for Documents. load() Load data into Document objects. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(resource: str, api_token: Optional[str] = None) → None[source]¶ Initialize the IUGU resource. Parameters resource – The name of the resource to fetch. api_token – The IUGU API token to use. lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document][source]¶ Load data into Document objects. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using IuguLoader¶ Iugu
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.iugu.IuguLoader.html
0e59a30a7f0d-0
langchain.document_loaders.unstructured.get_elements_from_api¶ langchain.document_loaders.unstructured.get_elements_from_api(file_path: Optional[Union[str, List[str]]] = None, file: Optional[Union[IO, Sequence[IO]]] = None, api_url: str = 'https://api.unstructured.io/general/v0/general', api_key: str = '', **unstructured_kwargs: Any) → List[source]¶ Retrieves a list of elements from the Unstructured API.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.unstructured.get_elements_from_api.html
b0653925e790-0
langchain.document_loaders.pdf.PyPDFDirectoryLoader¶ class langchain.document_loaders.pdf.PyPDFDirectoryLoader(path: str, glob: str = '**/[!.]*.pdf', silent_errors: bool = False, load_hidden: bool = False, recursive: bool = False)[source]¶ Loads a directory with PDF files with pypdf and chunks at character level. Loader also stores page numbers in metadata. Methods __init__(path[, glob, silent_errors, ...]) lazy_load() A lazy loader for Documents. load() Load data into Document objects. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(path: str, glob: str = '**/[!.]*.pdf', silent_errors: bool = False, load_hidden: bool = False, recursive: bool = False)[source]¶ lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document][source]¶ Load data into Document objects. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.pdf.PyPDFDirectoryLoader.html
30e93d04c999-0
langchain.document_loaders.larksuite.LarkSuiteDocLoader¶ class langchain.document_loaders.larksuite.LarkSuiteDocLoader(domain: str, access_token: str, document_id: str)[source]¶ Loads LarkSuite (FeiShu) document. Initialize with domain, access_token (tenant / user), and document_id. Parameters domain – The domain to load the LarkSuite. access_token – The access_token to use. document_id – The document_id to load. Methods __init__(domain, access_token, document_id) Initialize with domain, access_token (tenant / user), and document_id. lazy_load() Lazy load LarkSuite (FeiShu) document. load() Load LarkSuite (FeiShu) document. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(domain: str, access_token: str, document_id: str)[source]¶ Initialize with domain, access_token (tenant / user), and document_id. Parameters domain – The domain to load the LarkSuite. access_token – The access_token to use. document_id – The document_id to load. lazy_load() → Iterator[Document][source]¶ Lazy load LarkSuite (FeiShu) document. load() → List[Document][source]¶ Load LarkSuite (FeiShu) document. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using LarkSuiteDocLoader¶
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.larksuite.LarkSuiteDocLoader.html
30e93d04c999-1
Returns List of Documents. Examples using LarkSuiteDocLoader¶ LarkSuite (FeiShu)
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.larksuite.LarkSuiteDocLoader.html
47291d148cd8-0
langchain.document_loaders.mediawikidump.MWDumpLoader¶ class langchain.document_loaders.mediawikidump.MWDumpLoader(file_path: Union[str, Path], encoding: Optional[str] = 'utf8', namespaces: Optional[Sequence[int]] = None, skip_redirects: Optional[bool] = False, stop_on_error: Optional[bool] = True)[source]¶ Load MediaWiki dump from XML file .. rubric:: Example from langchain.document_loaders import MWDumpLoader loader = MWDumpLoader( file_path="myWiki.xml", encoding="utf8" ) docs = loader.load() from langchain.text_splitter import RecursiveCharacterTextSplitter text_splitter = RecursiveCharacterTextSplitter( chunk_size=1000, chunk_overlap=0 ) texts = text_splitter.split_documents(docs) Parameters file_path (str) – XML local file path encoding (str, optional) – Charset encoding, defaults to “utf8” namespaces (List[int],optional) – The namespace of pages you want to parse. See https://www.mediawiki.org/wiki/Help:Namespaces#Localisation for a list of all common namespaces skip_redirects (bool, optional) – TR=rue to skip pages that redirect to other pages, False to keep them. False by default stop_on_error (bool, optional) – False to skip over pages that cause parsing errors, True to stop. True by default Methods __init__(file_path[, encoding, namespaces, ...]) lazy_load() A lazy loader for Documents. load() Load from a file path. load_and_split([text_splitter]) Load Documents and split into chunks.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.mediawikidump.MWDumpLoader.html
47291d148cd8-1
load_and_split([text_splitter]) Load Documents and split into chunks. __init__(file_path: Union[str, Path], encoding: Optional[str] = 'utf8', namespaces: Optional[Sequence[int]] = None, skip_redirects: Optional[bool] = False, stop_on_error: Optional[bool] = True)[source]¶ lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document][source]¶ Load from a file path. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using MWDumpLoader¶ MediaWikiDump
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.mediawikidump.MWDumpLoader.html
233c00f23a9a-0
langchain.document_loaders.airbyte.AirbyteShopifyLoader¶ class langchain.document_loaders.airbyte.AirbyteShopifyLoader(config: Mapping[str, Any], stream_name: str, record_handler: Optional[Callable[[Any, Optional[str]], Document]] = None, state: Optional[Any] = None)[source]¶ Methods __init__(config, stream_name[, ...]) lazy_load() A lazy loader for Documents. load() Load data into Document objects. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(config: Mapping[str, Any], stream_name: str, record_handler: Optional[Callable[[Any, Optional[str]], Document]] = None, state: Optional[Any] = None) → None[source]¶ lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document]¶ Load data into Document objects. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.airbyte.AirbyteShopifyLoader.html
06d48cd91af0-0
langchain.document_loaders.whatsapp_chat.concatenate_rows¶ langchain.document_loaders.whatsapp_chat.concatenate_rows(date: str, sender: str, text: str) → str[source]¶ Combine message information in a readable format ready to be used.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.whatsapp_chat.concatenate_rows.html
7ad33d557147-0
langchain.document_loaders.image.UnstructuredImageLoader¶ class langchain.document_loaders.image.UnstructuredImageLoader(file_path: Union[str, List[str]], mode: str = 'single', **unstructured_kwargs: Any)[source]¶ Loader that uses Unstructured to load PNG and JPG files. You can run the loader in one of two modes: “single” and “elements”. If you use “single” mode, the document will be returned as a single langchain Document object. If you use “elements” mode, the unstructured library will split the document into elements such as Title and NarrativeText. You can pass in additional unstructured kwargs after mode to apply different unstructured settings. Examples from langchain.document_loaders import UnstructuredImageLoader loader = UnstructuredImageLoader(“example.png”, mode=”elements”, strategy=”fast”, ) docs = loader.load() References https://unstructured-io.github.io/unstructured/bricks.html#partition-image Initialize with file path. Methods __init__(file_path[, mode]) Initialize with file path. lazy_load() A lazy loader for Documents. load() Load file. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(file_path: Union[str, List[str]], mode: str = 'single', **unstructured_kwargs: Any)¶ Initialize with file path. lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document]¶ Load file. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.image.UnstructuredImageLoader.html
7ad33d557147-1
Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using UnstructuredImageLoader¶ Images
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.image.UnstructuredImageLoader.html
f8315e0f5a0d-0
langchain.document_loaders.nuclia.NucliaLoader¶ class langchain.document_loaders.nuclia.NucliaLoader(path: str, nuclia_tool: NucliaUnderstandingAPI)[source]¶ Extract text from any file type. Methods __init__(path, nuclia_tool) lazy_load() A lazy loader for Documents. load() Load documents. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(path: str, nuclia_tool: NucliaUnderstandingAPI)[source]¶ lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document][source]¶ Load documents. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.nuclia.NucliaLoader.html
750c0971f93a-0
langchain.document_loaders.rocksetdb.default_joiner¶ langchain.document_loaders.rocksetdb.default_joiner(docs: List[Tuple[str, Any]]) → str[source]¶ Default joiner for content columns.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.rocksetdb.default_joiner.html
ffd3e18035e2-0
langchain.document_loaders.url_selenium.SeleniumURLLoader¶ class langchain.document_loaders.url_selenium.SeleniumURLLoader(urls: List[str], continue_on_failure: bool = True, browser: Literal['chrome', 'firefox'] = 'chrome', binary_location: Optional[str] = None, executable_path: Optional[str] = None, headless: bool = True, arguments: List[str] = [])[source]¶ Loader that uses Selenium and to load a page and unstructured to load the html. This is useful for loading pages that require javascript to render. urls¶ List of URLs to load. Type List[str] continue_on_failure¶ If True, continue loading other URLs on failure. Type bool browser¶ The browser to use, either ‘chrome’ or ‘firefox’. Type str binary_location¶ The location of the browser binary. Type Optional[str] executable_path¶ The path to the browser executable. Type Optional[str] headless¶ If True, the browser will run in headless mode. Type bool arguments [List[str]] List of arguments to pass to the browser. Load a list of URLs using Selenium and unstructured. Methods __init__(urls[, continue_on_failure, ...]) Load a list of URLs using Selenium and unstructured. lazy_load() A lazy loader for Documents. load() Load the specified URLs using Selenium and create Document instances. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(urls: List[str], continue_on_failure: bool = True, browser: Literal['chrome', 'firefox'] = 'chrome', binary_location: Optional[str] = None, executable_path: Optional[str] = None, headless: bool = True, arguments: List[str] = [])[source]¶
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.url_selenium.SeleniumURLLoader.html
ffd3e18035e2-1
Load a list of URLs using Selenium and unstructured. lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document][source]¶ Load the specified URLs using Selenium and create Document instances. Returns A list of Document instances with loaded content. Return type List[Document] load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using SeleniumURLLoader¶ URL
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.url_selenium.SeleniumURLLoader.html
b5ba825c8dc5-0
langchain.document_loaders.brave_search.BraveSearchLoader¶ class langchain.document_loaders.brave_search.BraveSearchLoader(query: str, api_key: str, search_kwargs: Optional[dict] = None)[source]¶ Loads a query result from Brave Search engine into a list of Documents. Initializes the BraveLoader. Parameters query – The query to search for. api_key – The API key to use. search_kwargs – The search kwargs to use. Methods __init__(query, api_key[, search_kwargs]) Initializes the BraveLoader. lazy_load() A lazy loader for Documents. load() Load data into Document objects. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(query: str, api_key: str, search_kwargs: Optional[dict] = None)[source]¶ Initializes the BraveLoader. Parameters query – The query to search for. api_key – The API key to use. search_kwargs – The search kwargs to use. lazy_load() → Iterator[Document][source]¶ A lazy loader for Documents. load() → List[Document][source]¶ Load data into Document objects. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using BraveSearchLoader¶ Brave Search
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.brave_search.BraveSearchLoader.html
132264150f7f-0
langchain.document_loaders.github.GitHubIssuesLoader¶ class langchain.document_loaders.github.GitHubIssuesLoader[source]¶ Bases: BaseGitHubLoader Load issues of a GitHub repository. Create a new model by parsing and validating input data from keyword arguments. Raises ValidationError if the input data cannot be parsed to form a valid model. param access_token: str [Required]¶ Personal access token - see https://github.com/settings/tokens?type=beta param assignee: Optional[str] = None¶ Filter on assigned user. Pass ‘none’ for no user and ‘*’ for any user. param creator: Optional[str] = None¶ Filter on the user that created the issue. param direction: Optional[Literal['asc', 'desc']] = None¶ The direction to sort the results by. Can be one of: ‘asc’, ‘desc’. param include_prs: bool = True¶ If True include Pull Requests in results, otherwise ignore them. param labels: Optional[List[str]] = None¶ Label names to filter one. Example: bug,ui,@high. param mentioned: Optional[str] = None¶ Filter on a user that’s mentioned in the issue. param milestone: Optional[Union[int, Literal['*', 'none']]] = None¶ If integer is passed, it should be a milestone’s number field. If the string ‘*’ is passed, issues with any milestone are accepted. If the string ‘none’ is passed, issues without milestones are returned. param repo: str [Required]¶ Name of repository param since: Optional[str] = None¶ Only show notifications updated after the given time. This is a timestamp in ISO 8601 format: YYYY-MM-DDTHH:MM:SSZ.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.github.GitHubIssuesLoader.html
132264150f7f-1
param sort: Optional[Literal['created', 'updated', 'comments']] = None¶ What to sort results by. Can be one of: ‘created’, ‘updated’, ‘comments’. Default is ‘created’. param state: Optional[Literal['open', 'closed', 'all']] = None¶ Filter on issue state. Can be one of: ‘open’, ‘closed’, ‘all’. classmethod construct(_fields_set: Optional[SetStr] = None, **values: Any) → Model¶ Creates a new model setting __dict__ and __fields_set__ from trusted or pre-validated data. Default values are respected, but no other validation is performed. Behaves as if Config.extra = ‘allow’ was set since it adds all passed values copy(*, include: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, exclude: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, update: Optional[DictStrAny] = None, deep: bool = False) → Model¶ Duplicate a model, optionally choose which fields to include, exclude and change. Parameters include – fields to include in new model exclude – fields to exclude from new model, as with values this takes precedence over include update – values to change/add in the new model. Note: the data is not validated before creating the new model: you should trust this data deep – set to True to make a deep copy of the model Returns new model instance
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.github.GitHubIssuesLoader.html
132264150f7f-2
deep – set to True to make a deep copy of the model Returns new model instance dict(*, include: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, exclude: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, by_alias: bool = False, skip_defaults: Optional[bool] = None, exclude_unset: bool = False, exclude_defaults: bool = False, exclude_none: bool = False) → DictStrAny¶ Generate a dictionary representation of the model, optionally specifying which fields to include or exclude. classmethod from_orm(obj: Any) → Model¶ json(*, include: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, exclude: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, by_alias: bool = False, skip_defaults: Optional[bool] = None, exclude_unset: bool = False, exclude_defaults: bool = False, exclude_none: bool = False, encoder: Optional[Callable[[Any], Any]] = None, models_as_dict: bool = True, **dumps_kwargs: Any) → unicode¶ Generate a JSON representation of the model, include and exclude arguments as per dict(). encoder is an optional function to supply as default to json.dumps(), other arguments as per json.dumps(). lazy_load() → Iterator[Document][source]¶ Get issues of a GitHub repository. Returns page_content metadata url title creator created_at last_update_time closed_time number of comments state labels assignee assignees milestone locked number is_pull_request Return type A list of Documents with attributes load() → List[Document][source]¶ Get issues of a GitHub repository. Returns page_content metadata url title creator created_at
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.github.GitHubIssuesLoader.html
132264150f7f-3
Returns page_content metadata url title creator created_at last_update_time closed_time number of comments state labels assignee assignees milestone locked number is_pull_request Return type A list of Documents with attributes load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. classmethod parse_file(path: Union[str, Path], *, content_type: unicode = None, encoding: unicode = 'utf8', proto: Protocol = None, allow_pickle: bool = False) → Model¶ parse_issue(issue: dict) → Document[source]¶ Create Document objects from a list of GitHub issues. classmethod parse_obj(obj: Any) → Model¶ classmethod parse_raw(b: Union[str, bytes], *, content_type: unicode = None, encoding: unicode = 'utf8', proto: Protocol = None, allow_pickle: bool = False) → Model¶ classmethod schema(by_alias: bool = True, ref_template: unicode = '#/definitions/{model}') → DictStrAny¶ classmethod schema_json(*, by_alias: bool = True, ref_template: unicode = '#/definitions/{model}', **dumps_kwargs: Any) → unicode¶ classmethod update_forward_refs(**localns: Any) → None¶ Try to update ForwardRefs on fields based on this Model, globalns and localns. classmethod validate(value: Any) → Model¶ property headers: Dict[str, str]¶ property query_params: str¶ Create query parameters for GitHub API. property url: str¶ Create URL for GitHub API. Examples using GitHubIssuesLoader¶ GitHub
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.github.GitHubIssuesLoader.html
d7b2dc79cc9c-0
langchain.document_loaders.parsers.txt.TextParser¶ class langchain.document_loaders.parsers.txt.TextParser[source]¶ Parser for text blobs. Methods __init__() lazy_parse(blob) Lazily parse the blob. parse(blob) Eagerly parse the blob into a document or documents. __init__()¶ lazy_parse(blob: Blob) → Iterator[Document][source]¶ Lazily parse the blob. parse(blob: Blob) → List[Document]¶ Eagerly parse the blob into a document or documents. This is a convenience method for interactive development environment. Production applications should favor the lazy_parse method instead. Subclasses should generally not over-ride this parse method. Parameters blob – Blob instance Returns List of documents
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.parsers.txt.TextParser.html
d6a16697f4f6-0
langchain.document_loaders.parsers.language.javascript.JavaScriptSegmenter¶ class langchain.document_loaders.parsers.language.javascript.JavaScriptSegmenter(code: str)[source]¶ The code segmenter for JavaScript. Methods __init__(code) extract_functions_classes() is_valid() simplify_code() __init__(code: str)[source]¶ extract_functions_classes() → List[str][source]¶ is_valid() → bool[source]¶ simplify_code() → str[source]¶
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.parsers.language.javascript.JavaScriptSegmenter.html
27bb085732c7-0
langchain.document_loaders.max_compute.MaxComputeLoader¶ class langchain.document_loaders.max_compute.MaxComputeLoader(query: str, api_wrapper: MaxComputeAPIWrapper, *, page_content_columns: Optional[Sequence[str]] = None, metadata_columns: Optional[Sequence[str]] = None)[source]¶ Loads a query result from Alibaba Cloud MaxCompute table into documents. Initialize Alibaba Cloud MaxCompute document loader. Parameters query – SQL query to execute. api_wrapper – MaxCompute API wrapper. page_content_columns – The columns to write into the page_content of the Document. If unspecified, all columns will be written to page_content. metadata_columns – The columns to write into the metadata of the Document. If unspecified, all columns not added to page_content will be written. Methods __init__(query, api_wrapper, *[, ...]) Initialize Alibaba Cloud MaxCompute document loader. from_params(query, endpoint, project, *[, ...]) Convenience constructor that builds the MaxCompute API wrapper from lazy_load() A lazy loader for Documents. load() Load data into Document objects. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(query: str, api_wrapper: MaxComputeAPIWrapper, *, page_content_columns: Optional[Sequence[str]] = None, metadata_columns: Optional[Sequence[str]] = None)[source]¶ Initialize Alibaba Cloud MaxCompute document loader. Parameters query – SQL query to execute. api_wrapper – MaxCompute API wrapper. page_content_columns – The columns to write into the page_content of the Document. If unspecified, all columns will be written to page_content. metadata_columns – The columns to write into the metadata of the Document. If unspecified, all columns not added to page_content will be written.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.max_compute.MaxComputeLoader.html
27bb085732c7-1
If unspecified, all columns not added to page_content will be written. classmethod from_params(query: str, endpoint: str, project: str, *, access_id: Optional[str] = None, secret_access_key: Optional[str] = None, **kwargs: Any) → MaxComputeLoader[source]¶ Convenience constructor that builds the MaxCompute API wrapper fromgiven parameters. Parameters query – SQL query to execute. endpoint – MaxCompute endpoint. project – A project is a basic organizational unit of MaxCompute, which is similar to a database. access_id – MaxCompute access ID. Should be passed in directly or set as the environment variable MAX_COMPUTE_ACCESS_ID. secret_access_key – MaxCompute secret access key. Should be passed in directly or set as the environment variable MAX_COMPUTE_SECRET_ACCESS_KEY. lazy_load() → Iterator[Document][source]¶ A lazy loader for Documents. load() → List[Document][source]¶ Load data into Document objects. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using MaxComputeLoader¶ Alibaba Cloud MaxCompute
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.max_compute.MaxComputeLoader.html
449b3a1ca562-0
langchain.document_loaders.pubmed.PubMedLoader¶ class langchain.document_loaders.pubmed.PubMedLoader(query: str, load_max_docs: Optional[int] = 3)[source]¶ Loads a query result from PubMed biomedical library into a list of Documents. query¶ The query to be passed to the PubMed API. load_max_docs¶ The maximum number of documents to load. Initialize the PubMedLoader. Parameters query – The query to be passed to the PubMed API. load_max_docs – The maximum number of documents to load. Defaults to 3. Methods __init__(query[, load_max_docs]) Initialize the PubMedLoader. lazy_load() A lazy loader for Documents. load() Load data into Document objects. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(query: str, load_max_docs: Optional[int] = 3)[source]¶ Initialize the PubMedLoader. Parameters query – The query to be passed to the PubMed API. load_max_docs – The maximum number of documents to load. Defaults to 3. lazy_load() → Iterator[Document][source]¶ A lazy loader for Documents. load() → List[Document][source]¶ Load data into Document objects. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.pubmed.PubMedLoader.html
e3a3c2be7661-0
langchain.document_loaders.csv_loader.CSVLoader¶ class langchain.document_loaders.csv_loader.CSVLoader(file_path: str, source_column: Optional[str] = None, csv_args: Optional[Dict] = None, encoding: Optional[str] = None)[source]¶ Loads a CSV file into a list of documents. Each document represents one row of the CSV file. Every row is converted into a key/value pair and outputted to a new line in the document’s page_content. The source for each document loaded from csv is set to the value of the file_path argument for all documents by default. You can override this by setting the source_column argument to the name of a column in the CSV file. The source of each document will then be set to the value of the column with the name specified in source_column. Output Example:column1: value1 column2: value2 column3: value3 Parameters file_path – The path to the CSV file. source_column – The name of the column in the CSV file to use as the source. Optional. Defaults to None. csv_args – A dictionary of arguments to pass to the csv.DictReader. Optional. Defaults to None. encoding – The encoding of the CSV file. Optional. Defaults to None. Methods __init__(file_path[, source_column, ...]) param file_path The path to the CSV file. lazy_load() A lazy loader for Documents. load() Load data into document objects. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(file_path: str, source_column: Optional[str] = None, csv_args: Optional[Dict] = None, encoding: Optional[str] = None)[source]¶ Parameters file_path – The path to the CSV file.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.csv_loader.CSVLoader.html
e3a3c2be7661-1
Parameters file_path – The path to the CSV file. source_column – The name of the column in the CSV file to use as the source. Optional. Defaults to None. csv_args – A dictionary of arguments to pass to the csv.DictReader. Optional. Defaults to None. encoding – The encoding of the CSV file. Optional. Defaults to None. lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document][source]¶ Load data into document objects. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using CSVLoader¶ ChatGPT Plugin CSV
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.csv_loader.CSVLoader.html
9c030e957492-0
langchain.document_loaders.airbyte.AirbyteZendeskSupportLoader¶ class langchain.document_loaders.airbyte.AirbyteZendeskSupportLoader(config: Mapping[str, Any], stream_name: str, record_handler: Optional[Callable[[Any, Optional[str]], Document]] = None, state: Optional[Any] = None)[source]¶ Methods __init__(config, stream_name[, ...]) lazy_load() A lazy loader for Documents. load() Load data into Document objects. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(config: Mapping[str, Any], stream_name: str, record_handler: Optional[Callable[[Any, Optional[str]], Document]] = None, state: Optional[Any] = None) → None[source]¶ lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document]¶ Load data into Document objects. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.airbyte.AirbyteZendeskSupportLoader.html
05ca5479de27-0
langchain.document_loaders.dropbox.DropboxLoader¶ class langchain.document_loaders.dropbox.DropboxLoader[source]¶ Bases: BaseLoader, BaseModel Loads files from Dropbox. In addition to common files such as text and PDF files, it also supports Dropbox Paper files. Create a new model by parsing and validating input data from keyword arguments. Raises ValidationError if the input data cannot be parsed to form a valid model. param dropbox_access_token: str [Required]¶ Dropbox access token. param dropbox_file_paths: Optional[List[str]] = None¶ The file paths to load from. param dropbox_folder_path: Optional[str] = None¶ The folder path to load from. param recursive: bool = False¶ Flag to indicate whether to load files recursively from subfolders. classmethod construct(_fields_set: Optional[SetStr] = None, **values: Any) → Model¶ Creates a new model setting __dict__ and __fields_set__ from trusted or pre-validated data. Default values are respected, but no other validation is performed. Behaves as if Config.extra = ‘allow’ was set since it adds all passed values copy(*, include: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, exclude: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, update: Optional[DictStrAny] = None, deep: bool = False) → Model¶ Duplicate a model, optionally choose which fields to include, exclude and change. Parameters include – fields to include in new model exclude – fields to exclude from new model, as with values this takes precedence over include update – values to change/add in the new model. Note: the data is not validated before creating the new model: you should trust this data
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.dropbox.DropboxLoader.html
05ca5479de27-1
the new model: you should trust this data deep – set to True to make a deep copy of the model Returns new model instance dict(*, include: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, exclude: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, by_alias: bool = False, skip_defaults: Optional[bool] = None, exclude_unset: bool = False, exclude_defaults: bool = False, exclude_none: bool = False) → DictStrAny¶ Generate a dictionary representation of the model, optionally specifying which fields to include or exclude. classmethod from_orm(obj: Any) → Model¶ json(*, include: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, exclude: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, by_alias: bool = False, skip_defaults: Optional[bool] = None, exclude_unset: bool = False, exclude_defaults: bool = False, exclude_none: bool = False, encoder: Optional[Callable[[Any], Any]] = None, models_as_dict: bool = True, **dumps_kwargs: Any) → unicode¶ Generate a JSON representation of the model, include and exclude arguments as per dict(). encoder is an optional function to supply as default to json.dumps(), other arguments as per json.dumps(). lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document][source]¶ Load documents. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.dropbox.DropboxLoader.html
05ca5479de27-2
Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. classmethod parse_file(path: Union[str, Path], *, content_type: unicode = None, encoding: unicode = 'utf8', proto: Protocol = None, allow_pickle: bool = False) → Model¶ classmethod parse_obj(obj: Any) → Model¶ classmethod parse_raw(b: Union[str, bytes], *, content_type: unicode = None, encoding: unicode = 'utf8', proto: Protocol = None, allow_pickle: bool = False) → Model¶ classmethod schema(by_alias: bool = True, ref_template: unicode = '#/definitions/{model}') → DictStrAny¶ classmethod schema_json(*, by_alias: bool = True, ref_template: unicode = '#/definitions/{model}', **dumps_kwargs: Any) → unicode¶ classmethod update_forward_refs(**localns: Any) → None¶ Try to update ForwardRefs on fields based on this Model, globalns and localns. classmethod validate(value: Any) → Model¶ Examples using DropboxLoader¶ Dropbox
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.dropbox.DropboxLoader.html
98c423e826c9-0
langchain.document_loaders.notebook.remove_newlines¶ langchain.document_loaders.notebook.remove_newlines(x: Any) → Any[source]¶ Recursively removes newlines, no matter the data structure they are stored in.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.notebook.remove_newlines.html
063917237600-0
langchain.document_loaders.parsers.pdf.PDFMinerParser¶ class langchain.document_loaders.parsers.pdf.PDFMinerParser[source]¶ Parse PDFs with PDFMiner. Methods __init__() lazy_parse(blob) Lazily parse the blob. parse(blob) Eagerly parse the blob into a document or documents. __init__()¶ lazy_parse(blob: Blob) → Iterator[Document][source]¶ Lazily parse the blob. parse(blob: Blob) → List[Document]¶ Eagerly parse the blob into a document or documents. This is a convenience method for interactive development environment. Production applications should favor the lazy_parse method instead. Subclasses should generally not over-ride this parse method. Parameters blob – Blob instance Returns List of documents
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.parsers.pdf.PDFMinerParser.html
f59e139a63a9-0
langchain.document_loaders.async_html.AsyncHtmlLoader¶ class langchain.document_loaders.async_html.AsyncHtmlLoader(web_path: Union[str, List[str]], header_template: Optional[dict] = None, verify_ssl: Optional[bool] = True, proxies: Optional[dict] = None, requests_per_second: int = 2, requests_kwargs: Dict[str, Any] = {}, raise_for_status: bool = False)[source]¶ Loads HTML asynchronously. Initialize with webpage path. Methods __init__(web_path[, header_template, ...]) Initialize with webpage path. fetch_all(urls) Fetch all urls concurrently with rate limiting. lazy_load() Lazy load text from the url(s) in web_path. load() Load text from the url(s) in web_path. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(web_path: Union[str, List[str]], header_template: Optional[dict] = None, verify_ssl: Optional[bool] = True, proxies: Optional[dict] = None, requests_per_second: int = 2, requests_kwargs: Dict[str, Any] = {}, raise_for_status: bool = False)[source]¶ Initialize with webpage path. async fetch_all(urls: List[str]) → Any[source]¶ Fetch all urls concurrently with rate limiting. lazy_load() → Iterator[Document][source]¶ Lazy load text from the url(s) in web_path. load() → List[Document][source]¶ Load text from the url(s) in web_path. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.async_html.AsyncHtmlLoader.html
f59e139a63a9-1
Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using AsyncHtmlLoader¶ html2text AsyncHtmlLoader
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.async_html.AsyncHtmlLoader.html
065e39812c4e-0
langchain.document_loaders.gitbook.GitbookLoader¶ class langchain.document_loaders.gitbook.GitbookLoader(web_page: str, load_all_paths: bool = False, base_url: Optional[str] = None, content_selector: str = 'main', continue_on_failure: Optional[bool] = False)[source]¶ Load GitBook data. load from either a single page, or load all (relative) paths in the navbar. Initialize with web page and whether to load all paths. Parameters web_page – The web page to load or the starting point from where relative paths are discovered. load_all_paths – If set to True, all relative paths in the navbar are loaded instead of only web_page. base_url – If load_all_paths is True, the relative paths are appended to this base url. Defaults to web_page. content_selector – The CSS selector for the content to load. Defaults to “main”. continue_on_failure – whether to continue loading the sitemap if an error occurs loading a url, emitting a warning instead of raising an exception. Setting this to True makes the loader more robust, but also may result in missing data. Default: False Attributes bs_get_text_kwargs kwargs for beatifulsoup4 get_text default_parser Default parser to use for BeautifulSoup. raise_for_status Raise an exception if http status code denotes an error. requests_kwargs kwargs for requests requests_per_second Max number of concurrent requests to make. web_path Methods __init__(web_page[, load_all_paths, ...]) Initialize with web page and whether to load all paths. aload() Load text from the urls in web_path async into Documents. fetch_all(urls) Fetch all urls concurrently with rate limiting. lazy_load() Lazy load text from the url(s) in web_path.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.gitbook.GitbookLoader.html
065e39812c4e-1
lazy_load() Lazy load text from the url(s) in web_path. load() Fetch text from one single GitBook page. load_and_split([text_splitter]) Load Documents and split into chunks. scrape([parser]) Scrape data from webpage and return it in BeautifulSoup format. scrape_all(urls[, parser]) Fetch all urls, then return soups for all results. __init__(web_page: str, load_all_paths: bool = False, base_url: Optional[str] = None, content_selector: str = 'main', continue_on_failure: Optional[bool] = False)[source]¶ Initialize with web page and whether to load all paths. Parameters web_page – The web page to load or the starting point from where relative paths are discovered. load_all_paths – If set to True, all relative paths in the navbar are loaded instead of only web_page. base_url – If load_all_paths is True, the relative paths are appended to this base url. Defaults to web_page. content_selector – The CSS selector for the content to load. Defaults to “main”. continue_on_failure – whether to continue loading the sitemap if an error occurs loading a url, emitting a warning instead of raising an exception. Setting this to True makes the loader more robust, but also may result in missing data. Default: False aload() → List[Document]¶ Load text from the urls in web_path async into Documents. async fetch_all(urls: List[str]) → Any¶ Fetch all urls concurrently with rate limiting. lazy_load() → Iterator[Document]¶ Lazy load text from the url(s) in web_path. load() → List[Document][source]¶ Fetch text from one single GitBook page.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.gitbook.GitbookLoader.html
065e39812c4e-2
Fetch text from one single GitBook page. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. scrape(parser: Optional[str] = None) → Any¶ Scrape data from webpage and return it in BeautifulSoup format. scrape_all(urls: List[str], parser: Optional[str] = None) → List[Any]¶ Fetch all urls, then return soups for all results. Examples using GitbookLoader¶ GitBook
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.gitbook.GitbookLoader.html
32c899e809f1-0
langchain.document_loaders.tencent_cos_directory.TencentCOSDirectoryLoader¶ class langchain.document_loaders.tencent_cos_directory.TencentCOSDirectoryLoader(conf: Any, bucket: str, prefix: str = '')[source]¶ Loader for Tencent Cloud COS directory. Initialize with COS config, bucket and prefix. :param conf(CosConfig): COS config. :param bucket(str): COS bucket. :param prefix(str): prefix. Methods __init__(conf, bucket[, prefix]) Initialize with COS config, bucket and prefix. lazy_load() Load documents. load() Load data into Document objects. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(conf: Any, bucket: str, prefix: str = '')[source]¶ Initialize with COS config, bucket and prefix. :param conf(CosConfig): COS config. :param bucket(str): COS bucket. :param prefix(str): prefix. lazy_load() → Iterator[Document][source]¶ Load documents. load() → List[Document][source]¶ Load data into Document objects. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using TencentCOSDirectoryLoader¶ Tencent COS Directory
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.tencent_cos_directory.TencentCOSDirectoryLoader.html
411b0a8eebd6-0
langchain.document_loaders.youtube.GoogleApiYoutubeLoader¶ class langchain.document_loaders.youtube.GoogleApiYoutubeLoader(google_api_client: GoogleApiClient, channel_name: Optional[str] = None, video_ids: Optional[List[str]] = None, add_video_info: bool = True, captions_language: str = 'en', continue_on_failure: bool = False)[source]¶ Loads all Videos from a Channel To use, you should have the googleapiclient,youtube_transcript_api python package installed. As the service needs a google_api_client, you first have to initialize the GoogleApiClient. Additionally you have to either provide a channel name or a list of videoids “https://developers.google.com/docs/api/quickstart/python” Example from langchain.document_loaders import GoogleApiClient from langchain.document_loaders import GoogleApiYoutubeLoader google_api_client = GoogleApiClient( service_account_path=Path("path_to_your_sec_file.json") ) loader = GoogleApiYoutubeLoader( google_api_client=google_api_client, channel_name = "CodeAesthetic" ) load.load() Attributes add_video_info captions_language channel_name continue_on_failure video_ids google_api_client Methods __init__(google_api_client[, channel_name, ...]) lazy_load() A lazy loader for Documents. load() Load documents. load_and_split([text_splitter]) Load Documents and split into chunks. validate_channel_or_videoIds_is_set(values) Validate that either folder_id or document_ids is set, but not both. __init__(google_api_client: GoogleApiClient, channel_name: Optional[str] = None, video_ids: Optional[List[str]] = None, add_video_info: bool = True, captions_language: str = 'en', continue_on_failure: bool = False) → None¶
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.youtube.GoogleApiYoutubeLoader.html
411b0a8eebd6-1
lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document][source]¶ Load documents. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. classmethod validate_channel_or_videoIds_is_set(values: Dict[str, Any]) → Dict[str, Any][source]¶ Validate that either folder_id or document_ids is set, but not both. Examples using GoogleApiYoutubeLoader¶ YouTube YouTube transcripts
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.youtube.GoogleApiYoutubeLoader.html
b6fdd3ffa7a1-0
langchain.document_loaders.parsers.generic.MimeTypeBasedParser¶ class langchain.document_loaders.parsers.generic.MimeTypeBasedParser(handlers: Mapping[str, BaseBlobParser], *, fallback_parser: Optional[BaseBlobParser] = None)[source]¶ A parser that uses mime-types to determine how to parse a blob. This parser is useful for simple pipelines where the mime-type is sufficient to determine how to parse a blob. To use, configure handlers based on mime-types and pass them to the initializer. Example from langchain.document_loaders.parsers.generic import MimeTypeBasedParser parser = MimeTypeBasedParser( handlers={“application/pdf”: …, }, fallback_parser=…, ) Define a parser that uses mime-types to determine how to parse a blob. Parameters handlers – A mapping from mime-types to functions that take a blob, parse it and return a document. fallback_parser – A fallback_parser parser to use if the mime-type is not found in the handlers. If provided, this parser will be used to parse blobs with all mime-types not found in the handlers. If not provided, a ValueError will be raised if the mime-type is not found in the handlers. Methods __init__(handlers, *[, fallback_parser]) Define a parser that uses mime-types to determine how to parse a blob. lazy_parse(blob) Load documents from a blob. parse(blob) Eagerly parse the blob into a document or documents. __init__(handlers: Mapping[str, BaseBlobParser], *, fallback_parser: Optional[BaseBlobParser] = None) → None[source]¶ Define a parser that uses mime-types to determine how to parse a blob. Parameters handlers – A mapping from mime-types to functions that take a blob, parse it and return a document.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.parsers.generic.MimeTypeBasedParser.html
b6fdd3ffa7a1-1
and return a document. fallback_parser – A fallback_parser parser to use if the mime-type is not found in the handlers. If provided, this parser will be used to parse blobs with all mime-types not found in the handlers. If not provided, a ValueError will be raised if the mime-type is not found in the handlers. lazy_parse(blob: Blob) → Iterator[Document][source]¶ Load documents from a blob. parse(blob: Blob) → List[Document]¶ Eagerly parse the blob into a document or documents. This is a convenience method for interactive development environment. Production applications should favor the lazy_parse method instead. Subclasses should generally not over-ride this parse method. Parameters blob – Blob instance Returns List of documents
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.parsers.generic.MimeTypeBasedParser.html
ef681deb9b47-0
langchain.document_loaders.evernote.EverNoteLoader¶ class langchain.document_loaders.evernote.EverNoteLoader(file_path: str, load_single_document: bool = True)[source]¶ EverNote Loader. Loads an EverNote notebook export file e.g. my_notebook.enex into Documents. Instructions on producing this file can be found at https://help.evernote.com/hc/en-us/articles/209005557-Export-notes-and-notebooks-as-ENEX-or-HTML Currently only the plain text in the note is extracted and stored as the contents of the Document, any non content metadata (e.g. ‘author’, ‘created’, ‘updated’ etc. but not ‘content-raw’ or ‘resource’) tags on the note will be extracted and stored as metadata on the Document. Parameters file_path (str) – The path to the notebook export with a .enex extension load_single_document (bool) – Whether or not to concatenate the content of all notes into a single long Document. True (If this is set to) – the ‘source’ which contains the file name of the export. Initialize with file path. Methods __init__(file_path[, load_single_document]) Initialize with file path. lazy_load() A lazy loader for Documents. load() Load documents from EverNote export file. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(file_path: str, load_single_document: bool = True)[source]¶ Initialize with file path. lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document][source]¶ Load documents from EverNote export file. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.evernote.EverNoteLoader.html
ef681deb9b47-1
Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using EverNoteLoader¶ EverNote
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.evernote.EverNoteLoader.html
b43e66b706e4-0
langchain.document_loaders.unstructured.UnstructuredAPIFileLoader¶ class langchain.document_loaders.unstructured.UnstructuredAPIFileLoader(file_path: Union[str, List[str]] = '', mode: str = 'single', url: str = 'https://api.unstructured.io/general/v0/general', api_key: str = '', **unstructured_kwargs: Any)[source]¶ Loader that uses the Unstructured API to load files. By default, the loader makes a call to the hosted Unstructured API. If you are running the unstructured API locally, you can change the API rule by passing in the url parameter when you initialize the loader. The hosted Unstructured API requires an API key. See https://www.unstructured.io/api-key/ if you need to generate a key. You can run the loader in one of two modes: “single” and “elements”. If you use “single” mode, the document will be returned as a single langchain Document object. If you use “elements” mode, the unstructured library will split the document into elements such as Title and NarrativeText. You can pass in additional unstructured kwargs after mode to apply different unstructured settings. Examples ```python from langchain.document_loaders import UnstructuredAPIFileLoader loader = UnstructuredFileAPILoader(“example.pdf”, mode=”elements”, strategy=”fast”, api_key=”MY_API_KEY”, ) docs = loader.load() References https://unstructured-io.github.io/unstructured/bricks.html#partition https://www.unstructured.io/api-key/ https://github.com/Unstructured-IO/unstructured-api Initialize with file path. Methods __init__([file_path, mode, url, api_key]) Initialize with file path. lazy_load() A lazy loader for Documents. load() Load file.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.unstructured.UnstructuredAPIFileLoader.html
b43e66b706e4-1
lazy_load() A lazy loader for Documents. load() Load file. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(file_path: Union[str, List[str]] = '', mode: str = 'single', url: str = 'https://api.unstructured.io/general/v0/general', api_key: str = '', **unstructured_kwargs: Any)[source]¶ Initialize with file path. lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document]¶ Load file. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using UnstructuredAPIFileLoader¶ Unstructured File
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.unstructured.UnstructuredAPIFileLoader.html
15a2dc1d830a-0
langchain.document_loaders.html_bs.BSHTMLLoader¶ class langchain.document_loaders.html_bs.BSHTMLLoader(file_path: str, open_encoding: Optional[str] = None, bs_kwargs: Optional[dict] = None, get_text_separator: str = '')[source]¶ Loader that uses beautiful soup to parse HTML files. Initialise with path, and optionally, file encoding to use, and any kwargs to pass to the BeautifulSoup object. Parameters file_path – The path to the file to load. open_encoding – The encoding to use when opening the file. bs_kwargs – Any kwargs to pass to the BeautifulSoup object. get_text_separator – The separator to use when calling get_text on the soup. Methods __init__(file_path[, open_encoding, ...]) Initialise with path, and optionally, file encoding to use, and any kwargs to pass to the BeautifulSoup object. lazy_load() A lazy loader for Documents. load() Load HTML document into document objects. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(file_path: str, open_encoding: Optional[str] = None, bs_kwargs: Optional[dict] = None, get_text_separator: str = '') → None[source]¶ Initialise with path, and optionally, file encoding to use, and any kwargs to pass to the BeautifulSoup object. Parameters file_path – The path to the file to load. open_encoding – The encoding to use when opening the file. bs_kwargs – Any kwargs to pass to the BeautifulSoup object. get_text_separator – The separator to use when calling get_text on the soup. lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document][source]¶ Load HTML document into document objects.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.html_bs.BSHTMLLoader.html
15a2dc1d830a-1
load() → List[Document][source]¶ Load HTML document into document objects. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.html_bs.BSHTMLLoader.html
d832507761b6-0
langchain.document_loaders.parsers.pdf.PyPDFium2Parser¶ class langchain.document_loaders.parsers.pdf.PyPDFium2Parser[source]¶ Parse PDFs with PyPDFium2. Initialize the parser. Methods __init__() Initialize the parser. lazy_parse(blob) Lazily parse the blob. parse(blob) Eagerly parse the blob into a document or documents. __init__() → None[source]¶ Initialize the parser. lazy_parse(blob: Blob) → Iterator[Document][source]¶ Lazily parse the blob. parse(blob: Blob) → List[Document]¶ Eagerly parse the blob into a document or documents. This is a convenience method for interactive development environment. Production applications should favor the lazy_parse method instead. Subclasses should generally not over-ride this parse method. Parameters blob – Blob instance Returns List of documents
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.parsers.pdf.PyPDFium2Parser.html
a8e2aea47f21-0
langchain.document_loaders.web_base.WebBaseLoader¶ class langchain.document_loaders.web_base.WebBaseLoader(web_path: Union[str, List[str]], header_template: Optional[dict] = None, verify_ssl: Optional[bool] = True, proxies: Optional[dict] = None, continue_on_failure: Optional[bool] = False)[source]¶ Loader that uses urllib and beautiful soup to load webpages. Initialize with webpage path. Attributes bs_get_text_kwargs kwargs for beatifulsoup4 get_text default_parser Default parser to use for BeautifulSoup. raise_for_status Raise an exception if http status code denotes an error. requests_kwargs kwargs for requests requests_per_second Max number of concurrent requests to make. web_path web_paths Methods __init__(web_path[, header_template, ...]) Initialize with webpage path. aload() Load text from the urls in web_path async into Documents. fetch_all(urls) Fetch all urls concurrently with rate limiting. lazy_load() Lazy load text from the url(s) in web_path. load() Load text from the url(s) in web_path. load_and_split([text_splitter]) Load Documents and split into chunks. scrape([parser]) Scrape data from webpage and return it in BeautifulSoup format. scrape_all(urls[, parser]) Fetch all urls, then return soups for all results. __init__(web_path: Union[str, List[str]], header_template: Optional[dict] = None, verify_ssl: Optional[bool] = True, proxies: Optional[dict] = None, continue_on_failure: Optional[bool] = False)[source]¶ Initialize with webpage path. aload() → List[Document][source]¶ Load text from the urls in web_path async into Documents.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.web_base.WebBaseLoader.html
a8e2aea47f21-1
Load text from the urls in web_path async into Documents. async fetch_all(urls: List[str]) → Any[source]¶ Fetch all urls concurrently with rate limiting. lazy_load() → Iterator[Document][source]¶ Lazy load text from the url(s) in web_path. load() → List[Document][source]¶ Load text from the url(s) in web_path. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. scrape(parser: Optional[str] = None) → Any[source]¶ Scrape data from webpage and return it in BeautifulSoup format. scrape_all(urls: List[str], parser: Optional[str] = None) → List[Any][source]¶ Fetch all urls, then return soups for all results. Examples using WebBaseLoader¶ Vectorstore Agent WebBaseLoader MergeDocLoader QA over Documents Running LLMs locally Use local LLMs MultiQueryRetriever Combine agents and vector stores
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.web_base.WebBaseLoader.html
2bedab5ab11a-0
langchain.document_loaders.dataframe.DataFrameLoader¶ class langchain.document_loaders.dataframe.DataFrameLoader(data_frame: Any, page_content_column: str = 'text')[source]¶ Load Pandas DataFrame. Initialize with dataframe object. Parameters data_frame – Pandas DataFrame object. page_content_column – Name of the column containing the page content. Defaults to “text”. Methods __init__(data_frame[, page_content_column]) Initialize with dataframe object. lazy_load() Lazy load records from dataframe. load() Load full dataframe. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(data_frame: Any, page_content_column: str = 'text')[source]¶ Initialize with dataframe object. Parameters data_frame – Pandas DataFrame object. page_content_column – Name of the column containing the page content. Defaults to “text”. lazy_load() → Iterator[Document][source]¶ Lazy load records from dataframe. load() → List[Document][source]¶ Load full dataframe. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using DataFrameLoader¶ Pandas DataFrame
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.dataframe.DataFrameLoader.html
19e4c9212a24-0
langchain.document_loaders.base.BaseLoader¶ class langchain.document_loaders.base.BaseLoader[source]¶ Interface for loading Documents. Implementations should implement the lazy-loading method using generators to avoid loading all Documents into memory at once. The load method will remain as is for backwards compatibility, but its implementation should be just list(self.lazy_load()). Methods __init__() lazy_load() A lazy loader for Documents. load() Load data into Document objects. load_and_split([text_splitter]) Load Documents and split into chunks. __init__()¶ lazy_load() → Iterator[Document][source]¶ A lazy loader for Documents. abstract load() → List[Document][source]¶ Load data into Document objects. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document][source]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.base.BaseLoader.html
40b14b4bcd3a-0
langchain.document_loaders.chatgpt.ChatGPTLoader¶ class langchain.document_loaders.chatgpt.ChatGPTLoader(log_file: str, num_logs: int = - 1)[source]¶ Load conversations from exported ChatGPT data. Initialize a class object. Parameters log_file – Path to the log file num_logs – Number of logs to load. If 0, load all logs. Methods __init__(log_file[, num_logs]) Initialize a class object. lazy_load() A lazy loader for Documents. load() Load data into Document objects. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(log_file: str, num_logs: int = - 1)[source]¶ Initialize a class object. Parameters log_file – Path to the log file num_logs – Number of logs to load. If 0, load all logs. lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document][source]¶ Load data into Document objects. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using ChatGPTLoader¶ OpenAI ChatGPT Data
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.chatgpt.ChatGPTLoader.html
c4b06677db99-0
langchain.document_loaders.bibtex.BibtexLoader¶ class langchain.document_loaders.bibtex.BibtexLoader(file_path: str, *, parser: Optional[BibtexparserWrapper] = None, max_docs: Optional[int] = None, max_content_chars: Optional[int] = 4000, load_extra_metadata: bool = False, file_pattern: str = '[^:]+\\.pdf')[source]¶ Loads a bibtex file into a list of Documents. Each document represents one entry from the bibtex file. If a PDF file is present in the file bibtex field, the original PDF is loaded into the document text. If no such file entry is present, the abstract field is used instead. Initialize the BibtexLoader. Parameters file_path – Path to the bibtex file. parser – The parser to use. If None, a default parser is used. max_docs – Max number of associated documents to load. Use -1 means no limit. max_content_chars – Maximum number of characters to load from the PDF. load_extra_metadata – Whether to load extra metadata from the PDF. file_pattern – Regex pattern to match the file name in the bibtex. Methods __init__(file_path, *[, parser, max_docs, ...]) Initialize the BibtexLoader. lazy_load() Load bibtex file using bibtexparser and get the article texts plus the article metadata. load() Load bibtex file documents from the given bibtex file path. load_and_split([text_splitter]) Load Documents and split into chunks.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.bibtex.BibtexLoader.html
c4b06677db99-1
load_and_split([text_splitter]) Load Documents and split into chunks. __init__(file_path: str, *, parser: Optional[BibtexparserWrapper] = None, max_docs: Optional[int] = None, max_content_chars: Optional[int] = 4000, load_extra_metadata: bool = False, file_pattern: str = '[^:]+\\.pdf')[source]¶ Initialize the BibtexLoader. Parameters file_path – Path to the bibtex file. parser – The parser to use. If None, a default parser is used. max_docs – Max number of associated documents to load. Use -1 means no limit. max_content_chars – Maximum number of characters to load from the PDF. load_extra_metadata – Whether to load extra metadata from the PDF. file_pattern – Regex pattern to match the file name in the bibtex. lazy_load() → Iterator[Document][source]¶ Load bibtex file using bibtexparser and get the article texts plus the article metadata. See https://bibtexparser.readthedocs.io/en/master/ Returns a list of documents with the document.page_content in text format load() → List[Document][source]¶ Load bibtex file documents from the given bibtex file path. See https://bibtexparser.readthedocs.io/en/master/ Parameters file_path – the path to the bibtex file Returns a list of documents with the document.page_content in text format load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.bibtex.BibtexLoader.html
c4b06677db99-2
Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using BibtexLoader¶ BibTeX
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.bibtex.BibtexLoader.html
c74f16fda89e-0
langchain.document_loaders.news.NewsURLLoader¶ class langchain.document_loaders.news.NewsURLLoader(urls: List[str], text_mode: bool = True, nlp: bool = False, continue_on_failure: bool = True, show_progress_bar: bool = False, **newspaper_kwargs: Any)[source]¶ Loader that uses newspaper to load news articles from URLs. Parameters urls – URLs to load. Each is loaded into its own document. text_mode – If True, extract text from URL and use that for page content. Otherwise, extract raw HTML. nlp – If True, perform NLP on the extracted contents, like providing a summary and extracting keywords. continue_on_failure – If True, continue loading documents even if loading fails for a particular URL. show_progress_bar – If True, use tqdm to show a loading progress bar. Requires tqdm to be installed, pip install tqdm. **newspaper_kwargs – Any additional named arguments to pass to newspaper.Article(). Example from langchain.document_loaders import NewsURLLoader loader = NewsURLLoader( urls=["<url-1>", "<url-2>"], ) docs = loader.load() Newspaper reference:https://newspaper.readthedocs.io/en/latest/ Initialize with file path. Methods __init__(urls[, text_mode, nlp, ...]) Initialize with file path. lazy_load() A lazy loader for Documents. load() Load data into Document objects. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(urls: List[str], text_mode: bool = True, nlp: bool = False, continue_on_failure: bool = True, show_progress_bar: bool = False, **newspaper_kwargs: Any) → None[source]¶
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.news.NewsURLLoader.html
c74f16fda89e-1
Initialize with file path. lazy_load() → Iterator[Document][source]¶ A lazy loader for Documents. load() → List[Document][source]¶ Load data into Document objects. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.news.NewsURLLoader.html
d3a194889543-0
langchain.document_loaders.telegram.TelegramChatApiLoader¶ class langchain.document_loaders.telegram.TelegramChatApiLoader(chat_entity: Optional[EntityLike] = None, api_id: Optional[int] = None, api_hash: Optional[str] = None, username: Optional[str] = None, file_path: str = 'telegram_data.json')[source]¶ Loads Telegram chat json directory dump. Initialize with API parameters. Parameters chat_entity – The chat entity to fetch data from. api_id – The API ID. api_hash – The API hash. username – The username. file_path – The file path to save the data to. Defaults to “telegram_data.json”. Methods __init__([chat_entity, api_id, api_hash, ...]) Initialize with API parameters. fetch_data_from_telegram() Fetch data from Telegram API and save it as a JSON file. lazy_load() A lazy loader for Documents. load() Load documents. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(chat_entity: Optional[EntityLike] = None, api_id: Optional[int] = None, api_hash: Optional[str] = None, username: Optional[str] = None, file_path: str = 'telegram_data.json')[source]¶ Initialize with API parameters. Parameters chat_entity – The chat entity to fetch data from. api_id – The API ID. api_hash – The API hash. username – The username. file_path – The file path to save the data to. Defaults to “telegram_data.json”. async fetch_data_from_telegram() → None[source]¶ Fetch data from Telegram API and save it as a JSON file. lazy_load() → Iterator[Document]¶ A lazy loader for Documents.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.telegram.TelegramChatApiLoader.html
d3a194889543-1
lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document][source]¶ Load documents. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using TelegramChatApiLoader¶ Telegram
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.telegram.TelegramChatApiLoader.html
d3d579ff40df-0
langchain.document_loaders.college_confidential.CollegeConfidentialLoader¶ class langchain.document_loaders.college_confidential.CollegeConfidentialLoader(web_path: Union[str, List[str]], header_template: Optional[dict] = None, verify_ssl: Optional[bool] = True, proxies: Optional[dict] = None, continue_on_failure: Optional[bool] = False)[source]¶ Loads College Confidential webpages. Initialize with webpage path. Attributes bs_get_text_kwargs kwargs for beatifulsoup4 get_text default_parser Default parser to use for BeautifulSoup. raise_for_status Raise an exception if http status code denotes an error. requests_kwargs kwargs for requests requests_per_second Max number of concurrent requests to make. web_path Methods __init__(web_path[, header_template, ...]) Initialize with webpage path. aload() Load text from the urls in web_path async into Documents. fetch_all(urls) Fetch all urls concurrently with rate limiting. lazy_load() Lazy load text from the url(s) in web_path. load() Load webpages as Documents. load_and_split([text_splitter]) Load Documents and split into chunks. scrape([parser]) Scrape data from webpage and return it in BeautifulSoup format. scrape_all(urls[, parser]) Fetch all urls, then return soups for all results. __init__(web_path: Union[str, List[str]], header_template: Optional[dict] = None, verify_ssl: Optional[bool] = True, proxies: Optional[dict] = None, continue_on_failure: Optional[bool] = False)¶ Initialize with webpage path. aload() → List[Document]¶ Load text from the urls in web_path async into Documents. async fetch_all(urls: List[str]) → Any¶
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.college_confidential.CollegeConfidentialLoader.html
d3d579ff40df-1
async fetch_all(urls: List[str]) → Any¶ Fetch all urls concurrently with rate limiting. lazy_load() → Iterator[Document]¶ Lazy load text from the url(s) in web_path. load() → List[Document][source]¶ Load webpages as Documents. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. scrape(parser: Optional[str] = None) → Any¶ Scrape data from webpage and return it in BeautifulSoup format. scrape_all(urls: List[str], parser: Optional[str] = None) → List[Any]¶ Fetch all urls, then return soups for all results. Examples using CollegeConfidentialLoader¶ College Confidential
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.college_confidential.CollegeConfidentialLoader.html
e63381c4dd8d-0
langchain.document_loaders.obsidian.ObsidianLoader¶ class langchain.document_loaders.obsidian.ObsidianLoader(path: str, encoding: str = 'UTF-8', collect_metadata: bool = True)[source]¶ Loads Obsidian files from disk. Initialize with a path. Parameters path – Path to the directory containing the Obsidian files. encoding – Charset encoding, defaults to “UTF-8” collect_metadata – Whether to collect metadata from the front matter. Defaults to True. Attributes FRONT_MATTER_REGEX Methods __init__(path[, encoding, collect_metadata]) Initialize with a path. lazy_load() A lazy loader for Documents. load() Load documents. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(path: str, encoding: str = 'UTF-8', collect_metadata: bool = True)[source]¶ Initialize with a path. Parameters path – Path to the directory containing the Obsidian files. encoding – Charset encoding, defaults to “UTF-8” collect_metadata – Whether to collect metadata from the front matter. Defaults to True. lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document][source]¶ Load documents. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using ObsidianLoader¶ Obsidian
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.obsidian.ObsidianLoader.html
d47ff00c8e63-0
langchain.document_loaders.html.UnstructuredHTMLLoader¶ class langchain.document_loaders.html.UnstructuredHTMLLoader(file_path: Union[str, List[str]], mode: str = 'single', **unstructured_kwargs: Any)[source]¶ Loader that uses Unstructured to load HTML files. You can run the loader in one of two modes: “single” and “elements”. If you use “single” mode, the document will be returned as a single langchain Document object. If you use “elements” mode, the unstructured library will split the document into elements such as Title and NarrativeText. You can pass in additional unstructured kwargs after mode to apply different unstructured settings. Examples from langchain.document_loaders import UnstructuredHTMLLoader loader = UnstructuredHTMLLoader(“example.html”, mode=”elements”, strategy=”fast”, ) docs = loader.load() References https://unstructured-io.github.io/unstructured/bricks.html#partition-html Initialize with file path. Methods __init__(file_path[, mode]) Initialize with file path. lazy_load() A lazy loader for Documents. load() Load file. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(file_path: Union[str, List[str]], mode: str = 'single', **unstructured_kwargs: Any)¶ Initialize with file path. lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document]¶ Load file. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.html.UnstructuredHTMLLoader.html
d47ff00c8e63-1
Defaults to RecursiveCharacterTextSplitter. Returns List of Documents.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.html.UnstructuredHTMLLoader.html
8464546ee988-0
langchain.document_loaders.obs_directory.OBSDirectoryLoader¶ class langchain.document_loaders.obs_directory.OBSDirectoryLoader(bucket: str, endpoint: str, config: Optional[dict] = None, prefix: str = '')[source]¶ Loading logic for loading documents from Huawei OBS. Initialize the OBSDirectoryLoader with the specified settings. Parameters bucket (str) – The name of the OBS bucket to be used. endpoint (str) – The endpoint URL of your OBS bucket. config (dict) – The parameters for connecting to OBS, provided as a dictionary. The dictionary could have the following keys: - “ak” (str, optional): Your OBS access key (required if get_token_from_ecs is False and bucket policy is not public read). - “sk” (str, optional): Your OBS secret key (required if get_token_from_ecs is False and bucket policy is not public read). - “token” (str, optional): Your security token (required if using temporary credentials). - “get_token_from_ecs” (bool, optional): Whether to retrieve the security token from ECS. Defaults to False if not provided. If set to True, ak, sk, and token will be ignored. prefix (str, optional) – The prefix to be added to the OBS key. Defaults to “”. Note Before using this class, make sure you have registered with OBS and have the necessary credentials. The ak, sk, and endpoint values are mandatory unless get_token_from_ecs is True or the bucket policy is public read. token is required when using temporary credentials. Example To create a new OBSDirectoryLoader: ``` config = { “ak”: “your-access-key”, “sk”: “your-secret-key” directory_loader = OBSDirectoryLoader(“your-bucket-name”, “your-end-endpoint”, config, “your-prefix”)
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.obs_directory.OBSDirectoryLoader.html
8464546ee988-1
Methods __init__(bucket, endpoint[, config, prefix]) Initialize the OBSDirectoryLoader with the specified settings. lazy_load() A lazy loader for Documents. load() Load documents. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(bucket: str, endpoint: str, config: Optional[dict] = None, prefix: str = '')[source]¶ Initialize the OBSDirectoryLoader with the specified settings. Parameters bucket (str) – The name of the OBS bucket to be used. endpoint (str) – The endpoint URL of your OBS bucket. config (dict) – The parameters for connecting to OBS, provided as a dictionary. The dictionary could have the following keys: - “ak” (str, optional): Your OBS access key (required if get_token_from_ecs is False and bucket policy is not public read). - “sk” (str, optional): Your OBS secret key (required if get_token_from_ecs is False and bucket policy is not public read). - “token” (str, optional): Your security token (required if using temporary credentials). - “get_token_from_ecs” (bool, optional): Whether to retrieve the security token from ECS. Defaults to False if not provided. If set to True, ak, sk, and token will be ignored. prefix (str, optional) – The prefix to be added to the OBS key. Defaults to “”. Note Before using this class, make sure you have registered with OBS and have the necessary credentials. The ak, sk, and endpoint values are mandatory unless get_token_from_ecs is True or the bucket policy is public read. token is required when using temporary credentials. Example To create a new OBSDirectoryLoader: ``` config = { “ak”: “your-access-key”,
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.obs_directory.OBSDirectoryLoader.html
8464546ee988-2
``` config = { “ak”: “your-access-key”, “sk”: “your-secret-key” directory_loader = OBSDirectoryLoader(“your-bucket-name”, “your-end-endpoint”, config, “your-prefix”) lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document][source]¶ Load documents. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.obs_directory.OBSDirectoryLoader.html
40ef9c463798-0
langchain.document_loaders.figma.FigmaFileLoader¶ class langchain.document_loaders.figma.FigmaFileLoader(access_token: str, ids: str, key: str)[source]¶ Loads Figma file json. Initialize with access token, ids, and key. Parameters access_token – The access token for the Figma REST API. ids – The ids of the Figma file. key – The key for the Figma file Methods __init__(access_token, ids, key) Initialize with access token, ids, and key. lazy_load() A lazy loader for Documents. load() Load file load_and_split([text_splitter]) Load Documents and split into chunks. __init__(access_token: str, ids: str, key: str)[source]¶ Initialize with access token, ids, and key. Parameters access_token – The access token for the Figma REST API. ids – The ids of the Figma file. key – The key for the Figma file lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document][source]¶ Load file load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using FigmaFileLoader¶ Figma
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.figma.FigmaFileLoader.html
408338ee7ccf-0
langchain.document_loaders.pdf.PDFPlumberLoader¶ class langchain.document_loaders.pdf.PDFPlumberLoader(file_path: str, text_kwargs: Optional[Mapping[str, Any]] = None)[source]¶ Loader that uses pdfplumber to load PDF files. Initialize with a file path. Attributes source Methods __init__(file_path[, text_kwargs]) Initialize with a file path. lazy_load() A lazy loader for Documents. load() Load file. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(file_path: str, text_kwargs: Optional[Mapping[str, Any]] = None) → None[source]¶ Initialize with a file path. lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document][source]¶ Load file. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.pdf.PDFPlumberLoader.html
942cc9f62749-0
langchain.document_loaders.pdf.UnstructuredPDFLoader¶ class langchain.document_loaders.pdf.UnstructuredPDFLoader(file_path: Union[str, List[str]], mode: str = 'single', **unstructured_kwargs: Any)[source]¶ Loader that uses unstructured to load PDF files. You can run the loader in one of two modes: “single” and “elements”. If you use “single” mode, the document will be returned as a single langchain Document object. If you use “elements” mode, the unstructured library will split the document into elements such as Title and NarrativeText. You can pass in additional unstructured kwargs after mode to apply different unstructured settings. Examples from langchain.document_loaders import UnstructuredPDFLoader loader = UnstructuredPDFLoader(“example.pdf”, mode=”elements”, strategy=”fast”, ) docs = loader.load() References https://unstructured-io.github.io/unstructured/bricks.html#partition-pdf Initialize with file path. Methods __init__(file_path[, mode]) Initialize with file path. lazy_load() A lazy loader for Documents. load() Load file. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(file_path: Union[str, List[str]], mode: str = 'single', **unstructured_kwargs: Any)¶ Initialize with file path. lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document]¶ Load file. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.pdf.UnstructuredPDFLoader.html
942cc9f62749-1
Defaults to RecursiveCharacterTextSplitter. Returns List of Documents.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.pdf.UnstructuredPDFLoader.html
9e0bcad37d3f-0
langchain.document_loaders.fauna.FaunaLoader¶ class langchain.document_loaders.fauna.FaunaLoader(query: str, page_content_field: str, secret: str, metadata_fields: Optional[Sequence[str]] = None)[source]¶ FaunaDB Loader. query¶ The FQL query string to execute. Type str page_content_field¶ The field that contains the content of each page. Type str secret¶ The secret key for authenticating to FaunaDB. Type str metadata_fields¶ Optional list of field names to include in metadata. Type Optional[Sequence[str]] Methods __init__(query, page_content_field, secret) lazy_load() A lazy loader for Documents. load() Load data into Document objects. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(query: str, page_content_field: str, secret: str, metadata_fields: Optional[Sequence[str]] = None)[source]¶ lazy_load() → Iterator[Document][source]¶ A lazy loader for Documents. load() → List[Document][source]¶ Load data into Document objects. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using FaunaLoader¶ Fauna
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.fauna.FaunaLoader.html
e2cc60795dc4-0
langchain.document_loaders.telegram.TelegramChatFileLoader¶ class langchain.document_loaders.telegram.TelegramChatFileLoader(path: str)[source]¶ Loads Telegram chat json directory dump. Initialize with a path. Methods __init__(path) Initialize with a path. lazy_load() A lazy loader for Documents. load() Load documents. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(path: str)[source]¶ Initialize with a path. lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document][source]¶ Load documents. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using TelegramChatFileLoader¶ Telegram
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.telegram.TelegramChatFileLoader.html
929e05472513-0
langchain.document_loaders.git.GitLoader¶ class langchain.document_loaders.git.GitLoader(repo_path: str, clone_url: Optional[str] = None, branch: Optional[str] = 'main', file_filter: Optional[Callable[[str], bool]] = None)[source]¶ Loads files from a Git repository into a list of documents. The Repository can be local on disk available at repo_path, or remote at clone_url that will be cloned to repo_path. Currently, supports only text files. Each document represents one file in the repository. The path points to the local Git repository, and the branch specifies the branch to load files from. By default, it loads from the main branch. Parameters repo_path – The path to the Git repository. clone_url – Optional. The URL to clone the repository from. branch – Optional. The branch to load files from. Defaults to main. file_filter – Optional. A function that takes a file path and returns a boolean indicating whether to load the file. Defaults to None. Methods __init__(repo_path[, clone_url, branch, ...]) param repo_path The path to the Git repository. lazy_load() A lazy loader for Documents. load() Load data into Document objects. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(repo_path: str, clone_url: Optional[str] = None, branch: Optional[str] = 'main', file_filter: Optional[Callable[[str], bool]] = None)[source]¶ Parameters repo_path – The path to the Git repository. clone_url – Optional. The URL to clone the repository from. branch – Optional. The branch to load files from. Defaults to main. file_filter – Optional. A function that takes a file path and returns
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.git.GitLoader.html
929e05472513-1
file_filter – Optional. A function that takes a file path and returns a boolean indicating whether to load the file. Defaults to None. lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document][source]¶ Load data into Document objects. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using GitLoader¶ Git
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.git.GitLoader.html
a900c0808446-0
langchain.document_loaders.url.UnstructuredURLLoader¶ class langchain.document_loaders.url.UnstructuredURLLoader(urls: List[str], continue_on_failure: bool = True, mode: str = 'single', show_progress_bar: bool = False, **unstructured_kwargs: Any)[source]¶ Loader that use Unstructured to load files from remote URLs. Use the unstructured partition function to detect the MIME type and route the file to the appropriate partitioner. You can run the loader in one of two modes: “single” and “elements”. If you use “single” mode, the document will be returned as a single langchain Document object. If you use “elements” mode, the unstructured library will split the document into elements such as Title and NarrativeText. You can pass in additional unstructured kwargs after mode to apply different unstructured settings. Examples from langchain.document_loaders import UnstructuredURLLoader loader = UnstructuredURLLoader(ursl=[“<url-1>”, “<url-2>”], mode=”elements”, strategy=”fast”, ) docs = loader.load() References https://unstructured-io.github.io/unstructured/bricks.html#partition Initialize with file path. Methods __init__(urls[, continue_on_failure, mode, ...]) Initialize with file path. lazy_load() A lazy loader for Documents. load() Load file. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(urls: List[str], continue_on_failure: bool = True, mode: str = 'single', show_progress_bar: bool = False, **unstructured_kwargs: Any)[source]¶ Initialize with file path. lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document][source]¶
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.url.UnstructuredURLLoader.html
a900c0808446-1
A lazy loader for Documents. load() → List[Document][source]¶ Load file. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using UnstructuredURLLoader¶ URL
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.url.UnstructuredURLLoader.html
4a30e3012878-0
langchain.document_loaders.azlyrics.AZLyricsLoader¶ class langchain.document_loaders.azlyrics.AZLyricsLoader(web_path: Union[str, List[str]], header_template: Optional[dict] = None, verify_ssl: Optional[bool] = True, proxies: Optional[dict] = None, continue_on_failure: Optional[bool] = False)[source]¶ Loads AZLyrics webpages. Initialize with webpage path. Attributes bs_get_text_kwargs kwargs for beatifulsoup4 get_text default_parser Default parser to use for BeautifulSoup. raise_for_status Raise an exception if http status code denotes an error. requests_kwargs kwargs for requests requests_per_second Max number of concurrent requests to make. web_path Methods __init__(web_path[, header_template, ...]) Initialize with webpage path. aload() Load text from the urls in web_path async into Documents. fetch_all(urls) Fetch all urls concurrently with rate limiting. lazy_load() Lazy load text from the url(s) in web_path. load() Load webpages into Documents. load_and_split([text_splitter]) Load Documents and split into chunks. scrape([parser]) Scrape data from webpage and return it in BeautifulSoup format. scrape_all(urls[, parser]) Fetch all urls, then return soups for all results. __init__(web_path: Union[str, List[str]], header_template: Optional[dict] = None, verify_ssl: Optional[bool] = True, proxies: Optional[dict] = None, continue_on_failure: Optional[bool] = False)¶ Initialize with webpage path. aload() → List[Document]¶ Load text from the urls in web_path async into Documents. async fetch_all(urls: List[str]) → Any¶
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.azlyrics.AZLyricsLoader.html
4a30e3012878-1
async fetch_all(urls: List[str]) → Any¶ Fetch all urls concurrently with rate limiting. lazy_load() → Iterator[Document]¶ Lazy load text from the url(s) in web_path. load() → List[Document][source]¶ Load webpages into Documents. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. scrape(parser: Optional[str] = None) → Any¶ Scrape data from webpage and return it in BeautifulSoup format. scrape_all(urls: List[str], parser: Optional[str] = None) → List[Any]¶ Fetch all urls, then return soups for all results. Examples using AZLyricsLoader¶ AZLyrics
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.azlyrics.AZLyricsLoader.html
e86cfe15631d-0
langchain.document_loaders.sitemap.SitemapLoader¶ class langchain.document_loaders.sitemap.SitemapLoader(web_path: str, filter_urls: Optional[List[str]] = None, parsing_function: Optional[Callable] = None, blocksize: Optional[int] = None, blocknum: int = 0, meta_function: Optional[Callable] = None, is_local: bool = False, continue_on_failure: bool = False)[source]¶ Loader that fetches a sitemap and loads those URLs. Initialize with webpage path and optional filter URLs. Parameters web_path – url of the sitemap. can also be a local path filter_urls – list of strings or regexes that will be applied to filter the urls that are parsed and loaded parsing_function – Function to parse bs4.Soup output blocksize – number of sitemap locations per block blocknum – the number of the block that should be loaded - zero indexed. Default: 0 meta_function – Function to parse bs4.Soup output for metadata remember when setting this method to also copy metadata[“loc”] to metadata[“source”] if you are using this field is_local – whether the sitemap is a local file. Default: False continue_on_failure – whether to continue loading the sitemap if an error occurs loading a url, emitting a warning instead of raising an exception. Setting this to True makes the loader more robust, but also may result in missing data. Default: False Attributes bs_get_text_kwargs kwargs for beatifulsoup4 get_text default_parser Default parser to use for BeautifulSoup. raise_for_status Raise an exception if http status code denotes an error. requests_kwargs kwargs for requests requests_per_second Max number of concurrent requests to make. web_path Methods __init__(web_path[, filter_urls, ...])
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.sitemap.SitemapLoader.html
e86cfe15631d-1
web_path Methods __init__(web_path[, filter_urls, ...]) Initialize with webpage path and optional filter URLs. aload() Load text from the urls in web_path async into Documents. fetch_all(urls) Fetch all urls concurrently with rate limiting. lazy_load() Lazy load text from the url(s) in web_path. load() Load sitemap. load_and_split([text_splitter]) Load Documents and split into chunks. parse_sitemap(soup) Parse sitemap xml and load into a list of dicts. scrape([parser]) Scrape data from webpage and return it in BeautifulSoup format. scrape_all(urls[, parser]) Fetch all urls, then return soups for all results. __init__(web_path: str, filter_urls: Optional[List[str]] = None, parsing_function: Optional[Callable] = None, blocksize: Optional[int] = None, blocknum: int = 0, meta_function: Optional[Callable] = None, is_local: bool = False, continue_on_failure: bool = False)[source]¶ Initialize with webpage path and optional filter URLs. Parameters web_path – url of the sitemap. can also be a local path filter_urls – list of strings or regexes that will be applied to filter the urls that are parsed and loaded parsing_function – Function to parse bs4.Soup output blocksize – number of sitemap locations per block blocknum – the number of the block that should be loaded - zero indexed. Default: 0 meta_function – Function to parse bs4.Soup output for metadata remember when setting this method to also copy metadata[“loc”] to metadata[“source”] if you are using this field is_local – whether the sitemap is a local file. Default: False
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.sitemap.SitemapLoader.html
e86cfe15631d-2
is_local – whether the sitemap is a local file. Default: False continue_on_failure – whether to continue loading the sitemap if an error occurs loading a url, emitting a warning instead of raising an exception. Setting this to True makes the loader more robust, but also may result in missing data. Default: False aload() → List[Document]¶ Load text from the urls in web_path async into Documents. async fetch_all(urls: List[str]) → Any¶ Fetch all urls concurrently with rate limiting. lazy_load() → Iterator[Document]¶ Lazy load text from the url(s) in web_path. load() → List[Document][source]¶ Load sitemap. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. parse_sitemap(soup: Any) → List[dict][source]¶ Parse sitemap xml and load into a list of dicts. Parameters soup – BeautifulSoup object. Returns List of dicts. scrape(parser: Optional[str] = None) → Any¶ Scrape data from webpage and return it in BeautifulSoup format. scrape_all(urls: List[str], parser: Optional[str] = None) → List[Any]¶ Fetch all urls, then return soups for all results. Examples using SitemapLoader¶ Sitemap
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.sitemap.SitemapLoader.html
b9fc2766020a-0
langchain.document_loaders.org_mode.UnstructuredOrgModeLoader¶ class langchain.document_loaders.org_mode.UnstructuredOrgModeLoader(file_path: str, mode: str = 'single', **unstructured_kwargs: Any)[source]¶ Loader that uses unstructured to load Org-Mode files. You can run the loader in one of two modes: “single” and “elements”. If you use “single” mode, the document will be returned as a single langchain Document object. If you use “elements” mode, the unstructured library will split the document into elements such as Title and NarrativeText. You can pass in additional unstructured kwargs after mode to apply different unstructured settings. Examples from langchain.document_loaders import UnstructuredOrgModeLoader loader = UnstructuredOrgModeLoader(“example.org”, mode=”elements”, strategy=”fast”, ) docs = loader.load() References https://unstructured-io.github.io/unstructured/bricks.html#partition-org Parameters file_path – The path to the file to load. mode – The mode to load the file from. Default is “single”. **unstructured_kwargs – Any additional keyword arguments to pass to the unstructured. Methods __init__(file_path[, mode]) param file_path The path to the file to load. lazy_load() A lazy loader for Documents. load() Load file. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(file_path: str, mode: str = 'single', **unstructured_kwargs: Any)[source]¶ Parameters file_path – The path to the file to load. mode – The mode to load the file from. Default is “single”. **unstructured_kwargs – Any additional keyword arguments to pass to the unstructured.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.org_mode.UnstructuredOrgModeLoader.html
b9fc2766020a-1
**unstructured_kwargs – Any additional keyword arguments to pass to the unstructured. lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document]¶ Load file. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using UnstructuredOrgModeLoader¶ Org-mode
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.org_mode.UnstructuredOrgModeLoader.html
86a2b2a3a666-0
langchain.document_loaders.email.UnstructuredEmailLoader¶ class langchain.document_loaders.email.UnstructuredEmailLoader(file_path: str, mode: str = 'single', **unstructured_kwargs: Any)[source]¶ Loader that uses unstructured to load email files. Works with both .eml and .msg files. You can process attachments in addition to the e-mail message itself by passing process_attachments=True into the constructor for the loader. By default, attachments will be processed with the unstructured partition function. If you already know the document types of the attachments, you can specify another partitioning function with the attachment partitioner kwarg. Example from langchain.document_loaders import UnstructuredEmailLoader loader = UnstructuredEmailLoader(“example_data/fake-email.eml”, mode=”elements”) loader.load() Example from langchain.document_loaders import UnstructuredEmailLoader loader = UnstructuredEmailLoader(“example_data/fake-email-attachment.eml”, mode=”elements”, process_attachments=True, ) loader.load() Initialize with file path. Methods __init__(file_path[, mode]) Initialize with file path. lazy_load() A lazy loader for Documents. load() Load file. load_and_split([text_splitter]) Load Documents and split into chunks. __init__(file_path: str, mode: str = 'single', **unstructured_kwargs: Any)[source]¶ Initialize with file path. lazy_load() → Iterator[Document]¶ A lazy loader for Documents. load() → List[Document]¶ Load file. load_and_split(text_splitter: Optional[TextSplitter] = None) → List[Document]¶ Load Documents and split into chunks. Chunks are returned as Documents. Parameters text_splitter – TextSplitter instance to use for splitting documents.
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.email.UnstructuredEmailLoader.html
86a2b2a3a666-1
Parameters text_splitter – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter. Returns List of Documents. Examples using UnstructuredEmailLoader¶ Email
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.email.UnstructuredEmailLoader.html
e30990cd7c9a-0
langchain.document_loaders.parsers.pdf.AmazonTextractPDFParser¶ class langchain.document_loaders.parsers.pdf.AmazonTextractPDFParser(textract_features: Optional[Sequence[int]] = None, client: Optional[Any] = None)[source]¶ Sends PDF files to Amazon Textract and parses them to generate Documents. For parsing multi-page PDFs, they have to reside on S3. Initializes the parser. Parameters textract_features – Features to be used for extraction, each feature should be passed as an int that conforms to the enum Textract_Features, see amazon-textract-caller pkg client – boto3 textract client Methods __init__([textract_features, client]) Initializes the parser. lazy_parse(blob) Iterates over the Blob pages and returns an Iterator with a Document for each page, like the other parsers If multi-page document, blob.path has to be set to the S3 URI and for single page docs the blob.data is taken parse(blob) Eagerly parse the blob into a document or documents. __init__(textract_features: Optional[Sequence[int]] = None, client: Optional[Any] = None) → None[source]¶ Initializes the parser. Parameters textract_features – Features to be used for extraction, each feature should be passed as an int that conforms to the enum Textract_Features, see amazon-textract-caller pkg client – boto3 textract client lazy_parse(blob: Blob) → Iterator[Document][source]¶ Iterates over the Blob pages and returns an Iterator with a Document for each page, like the other parsers If multi-page document, blob.path has to be set to the S3 URI and for single page docs the blob.data is taken parse(blob: Blob) → List[Document]¶
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.parsers.pdf.AmazonTextractPDFParser.html
e30990cd7c9a-1
parse(blob: Blob) → List[Document]¶ Eagerly parse the blob into a document or documents. This is a convenience method for interactive development environment. Production applications should favor the lazy_parse method instead. Subclasses should generally not over-ride this parse method. Parameters blob – Blob instance Returns List of documents
https://api.python.langchain.com/en/latest/document_loaders/langchain.document_loaders.parsers.pdf.AmazonTextractPDFParser.html