id
stringlengths
14
15
text
stringlengths
35
2.51k
source
stringlengths
61
154
00b928ef36c7-0
Source code for langchain.output_parsers.rail_parser from __future__ import annotations from typing import Any, Callable, Dict, Optional from langchain.schema import BaseOutputParser [docs]class GuardrailsOutputParser(BaseOutputParser): guard: Any api: Optional[Callable] args: Any kwargs: Any @prope...
https://api.python.langchain.com/en/latest/_modules/langchain/output_parsers/rail_parser.html
00b928ef36c7-1
) return cls( guard=Guard.from_rail_string(rail_str, num_reasks=num_reasks), api=api, args=args, kwargs=kwargs, ) [docs] @classmethod def from_pydantic( cls, output_class: Any, num_reasks: int = 1, api: Optional[Calla...
https://api.python.langchain.com/en/latest/_modules/langchain/output_parsers/rail_parser.html
1c7528234fbc-0
Source code for langchain.document_loaders.blockchain import os import re import time from enum import Enum from typing import List, Optional import requests from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader [docs]class BlockchainType(Enum): """Enumerator of the...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/blockchain.html
1c7528234fbc-1
""" def __init__( self, contract_address: str, blockchainType: BlockchainType = BlockchainType.ETH_MAINNET, api_key: str = "docs-demo", startToken: str = "", get_all_tokens: bool = False, max_execution_time: Optional[int] = None, ): self.contract_a...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/blockchain.html
1c7528234fbc-2
tokenId = item["id"]["tokenId"] metadata = { "source": self.contract_address, "blockchain": self.blockchainType, "tokenId": tokenId, } result.append(Document(page_content=content, metadata=metadata)) ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/blockchain.html
1c7528234fbc-3
elif value_type == "hex_0xbf": return "0xbf" + format(result, "0" + str(len(tokenId) - 4) + "x") else: return str(result) # A smart contract can use different formats for the tokenId @staticmethod def _detect_value_type(tokenId: str) -> str: if isinstance(tokenId, int...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/blockchain.html
27ada31af51a-0
Source code for langchain.document_loaders.airtable from typing import Iterator, List from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader [docs]class AirtableLoader(BaseLoader): """Loader for Airtable tables.""" def __init__(self, api_token: str, table_id: str...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/airtable.html
7b23cafd55cc-0
Source code for langchain.document_loaders.sitemap """Loader that fetches a sitemap and loads those URLs.""" import itertools import re from typing import Any, Callable, Generator, Iterable, List, Optional from langchain.document_loaders.web_base import WebBaseLoader from langchain.schema import Document def _default_p...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/sitemap.html
7b23cafd55cc-1
meta_function: Function to parse bs4.Soup output for metadata remember when setting this method to also copy metadata["loc"] to metadata["source"] if you are using this field is_local: whether the sitemap is a local file """ if blocksize is not None and blocks...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/sitemap.html
7b23cafd55cc-2
if (prop := url.find(tag)) } ) for sitemap in soup.find_all("sitemap"): loc = sitemap.find("loc") if not loc: continue soup_child = self.scrape_all([loc.text], "xml")[0] els.extend(self.parse_sitemap(soup_child)) ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/sitemap.html
bb6b466b38c2-0
Source code for langchain.document_loaders.notiondb """Notion DB loader for langchain""" from typing import Any, Dict, List, Optional import requests from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader NOTION_BASE_URL = "https://api.notion.com/v1" DATABASE_URL = NOTIO...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/notiondb.html
bb6b466b38c2-1
Returns: List[Document]: List of documents. """ page_summaries = self._retrieve_page_summaries() return list(self.load_page(page_summary) for page_summary in page_summaries) def _retrieve_page_summaries( self, query_dict: Dict[str, Any] = {"page_size": 100} ) -> List[...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/notiondb.html
bb6b466b38c2-2
if prop_data["multi_select"] else [] ) elif prop_type == "url": value = prop_data["url"] else: value = None metadata[prop_name.lower()] = value metadata["id"] = page_id return Document(page_content=se...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/notiondb.html
bb6b466b38c2-3
method, url, headers=self.headers, json=query_dict, timeout=self.request_timeout_sec, ) res.raise_for_status() return res.json()
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/notiondb.html
e7da210922b2-0
Source code for langchain.document_loaders.larksuite """Loader that loads LarkSuite (FeiShu) document json dump.""" import json import urllib.request from typing import Any, Iterator, List from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader [docs]class LarkSuiteDocLoa...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/larksuite.html
e7da210922b2-1
metadata = { "document_id": self.document_id, "revision_id": metadata_json["data"]["document"]["revision_id"], "title": metadata_json["data"]["document"]["title"], } yield Document(page_content=text, metadata=metadata) [docs] def load(self) -> List[Document]: ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/larksuite.html
5f2eb59bc49c-0
Source code for langchain.document_loaders.trello """Loader that loads cards from Trello""" from __future__ import annotations from typing import TYPE_CHECKING, Any, List, Literal, Optional, Tuple from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader from langchain.util...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/trello.html
5f2eb59bc49c-1
self.board_name = board_name self.include_card_name = include_card_name self.include_comments = include_comments self.include_checklist = include_checklist self.extra_metadata = extra_metadata self.card_filter = card_filter [docs] @classmethod def from_credentials( ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/trello.html
5f2eb59bc49c-2
token = token or get_from_env("token", "TRELLO_TOKEN") client = TrelloClient(api_key=api_key, token=token) return cls(client, board_name, **kwargs) [docs] def load(self) -> List[Document]: """Loads all cards from the specified Trello board. You can filter the cards, metadata and text ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/trello.html
5f2eb59bc49c-3
if self.include_card_name: text_content = card.name + "\n" if card.description.strip(): text_content += BeautifulSoup(card.description, "lxml").get_text() if self.include_checklist: # Get all the checklist items on the card for checklist in card.checklists...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/trello.html
aba0121c9013-0
Source code for langchain.document_loaders.gcs_directory """Loading logic for loading documents from an GCS directory.""" from typing import List from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader from langchain.document_loaders.gcs_file import GCSFileLoader [docs]cl...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/gcs_directory.html
36e7751bc8e6-0
Source code for langchain.document_loaders.acreom """Loader that loads acreom vault from a directory.""" import re from pathlib import Path from typing import Iterator, List from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader [docs]class AcreomLoader(BaseLoader): ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/acreom.html
36e7751bc8e6-1
# do not contribute to the context of current document content = re.sub("\s*-\s\[\s\]\s.*|\s*\[\s\]\s.*", "", content) # rm tasks content = re.sub("#", "", content) # rm hashtags content = re.sub("\[\[.*?\]\]", "", content) # rm doclinks return content [docs] def lazy_load(self) ->...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/acreom.html
f0d5cd018459-0
Source code for langchain.document_loaders.embaas import base64 import warnings from typing import Any, Dict, Iterator, List, Optional import requests from pydantic import BaseModel, root_validator, validator from typing_extensions import NotRequired, TypedDict from langchain.docstore.document import Document from lang...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/embaas.html
f0d5cd018459-1
"""The instruction to pass to the Embaas document extraction API.""" [docs]class EmbaasDocumentExtractionPayload(EmbaasDocumentExtractionParameters): """Payload for the Embaas document extraction API.""" bytes: str """The base64 encoded bytes of the document to extract text from.""" [docs]class BaseEmbaasLo...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/embaas.html
f0d5cd018459-2
documents = loader.parse(blob=blob) # Custom api parameters (create embeddings automatically) from langchain.document_loaders.embaas import EmbaasBlobLoader loader = EmbaasBlobLoader( params={ "should_embed": True, "model": "e5-...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/embaas.html
f0d5cd018459-3
# type: ignore **self.params, ) if blob.mimetype is not None and payload.get("mime_type", None) is None: payload["mime_type"] = blob.mimetype return payload def _handle_request( self, payload: EmbaasDocumentExtractionPayload ) -> List[Document]: ""...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/embaas.html
f0d5cd018459-4
To use, you should have the environment variable ``EMBAAS_API_KEY`` set with your API key, or pass it as a named parameter to the constructor. Example: .. code-block:: python # Default parsing from langchain.document_loaders.embaas import EmbaasLoader loader = Emb...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/embaas.html
f0d5cd018459-5
assert self.blob_loader is not None # Should never be None, but mypy doesn't know that. yield from self.blob_loader.lazy_parse(blob=blob) [docs] def load(self) -> List[Document]: return list(self.lazy_load()) [docs] def load_and_split( self, text_splitter: Optional[TextSplitter] = ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/embaas.html
e5429edf82f2-0
Source code for langchain.document_loaders.college_confidential """Loader that loads College Confidential.""" from typing import List from langchain.docstore.document import Document from langchain.document_loaders.web_base import WebBaseLoader [docs]class CollegeConfidentialLoader(WebBaseLoader): """Loader that lo...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/college_confidential.html
05ec8c196b0c-0
Source code for langchain.document_loaders.googledrive """Loader that loads data from Google Drive.""" # Prerequisites: # 1. Create a Google Cloud project # 2. Enable the Google Drive API: # https://console.cloud.google.com/flows/enableapi?apiid=drive.googleapis.com # 3. Authorize credentials for desktop app: # htt...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/googledrive.html
05ec8c196b0c-1
# results in pydantic validation errors file_loader_cls: Any = None file_loader_kwargs: Dict["str", Any] = {} [docs] @root_validator def validate_inputs(cls, values: Dict[str, Any]) -> Dict[str, Any]: """Validate that either folder_id or document_ids is set, but not both.""" if values.get...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/googledrive.html
05ec8c196b0c-2
if file_type not in allowed_types: raise ValueError( f"Given file type {file_type} is not supported. " f"Supported values are: {short_names}; and " f"their full-form names: {full_names}" ) # repla...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/googledrive.html
05ec8c196b0c-3
) if self.token_path.exists(): creds = Credentials.from_authorized_user_file(str(self.token_path), SCOPES) if not creds or not creds.valid: if creds and creds.expired and creds.refresh_token: creds.refresh(Request()) elif "GOOGLE_APPLICATION_CREDENTIAL...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/googledrive.html
05ec8c196b0c-4
metadata = { "source": ( f"https://docs.google.com/spreadsheets/d/{id}/" f"edit?gid={sheet['properties']['sheetId']}" ), "title": f"{spreadsheet['properties']['title']} - {sheet_name}", "row":...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/googledrive.html
05ec8c196b0c-5
text = fh.getvalue().decode("utf-8") metadata = { "source": f"https://docs.google.com/document/d/{id}/edit", "title": f"{file.get('name')}", } return Document(page_content=text, metadata=metadata) def _load_documents_from_folder( self, folder_id: str, *, file_...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/googledrive.html
05ec8c196b0c-6
else: pass return returns def _fetch_files_recursive( self, service: Any, folder_id: str ) -> List[Dict[str, Union[str, List[str]]]]: """Fetch all files and subfolders recursively.""" results = ( service.files() .list( q=f"'...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/googledrive.html
05ec8c196b0c-7
file = service.files().get(fileId=id, supportsAllDrives=True).execute() request = service.files().get_media(fileId=id) fh = BytesIO() downloader = MediaIoBaseDownload(fh, request) done = False while done is False: status, done = downloader.next_chunk() if self...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/googledrive.html
05ec8c196b0c-8
) elif self.document_ids: return self._load_documents_from_ids() else: return self._load_file_from_ids()
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/googledrive.html
568944512e6f-0
Source code for langchain.document_loaders.stripe """Loader that fetches data from Stripe""" import json import urllib.request from typing import List, Optional from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader from langchain.utils import get_from_env, stringify_dic...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/stripe.html
568944512e6f-1
if endpoint is None: return [] return self._make_request(endpoint) [docs] def load(self) -> List[Document]: return self._get_resource()
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/stripe.html
9ca4d2d127b8-0
Source code for langchain.document_loaders.fauna from typing import Iterator, List, Optional, Sequence from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader [docs]class FaunaLoader(BaseLoader): """FaunaDB Loader. Attributes: query (str): The FQL query st...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/fauna.html
9ca4d2d127b8-1
document_dict = dict(result.items()) page_content = "" for key, value in document_dict.items(): if key == self.page_content_field: page_content = value document: Document = Document( page_content=page_content...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/fauna.html
a4d0ecf1935d-0
Source code for langchain.document_loaders.duckdb_loader from typing import Dict, List, Optional, cast from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader [docs]class DuckDBLoader(BaseLoader): """Loads a query result from DuckDB into a list of documents. Each ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/duckdb_loader.html
a4d0ecf1935d-1
results = query_result.fetchall() description = cast(list, query_result.description) field_names = [c[0] for c in description] if self.page_content_columns is None: page_content_columns = field_names else: page_content_columns = self.page_c...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/duckdb_loader.html
60aaa137ac4b-0
Source code for langchain.document_loaders.gutenberg """Loader that loads .txt web files.""" from typing import List from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader [docs]class GutenbergLoader(BaseLoader): """Loader that uses urllib to load .txt web files.""" ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/gutenberg.html
1227e3d833ee-0
Source code for langchain.document_loaders.youtube """Loader that loads YouTube transcript.""" from __future__ import annotations import logging from pathlib import Path from typing import Any, Dict, List, Optional, Sequence, Union from urllib.parse import parse_qs, urlparse from pydantic import root_validator from pyd...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/youtube.html
1227e3d833ee-1
"""Validate that either folder_id or document_ids is set, but not both.""" if not values.get("credentials_path") and not values.get( "service_account_path" ): raise ValueError("Must specify either channel_name or video_ids") return values def _load_credentials(self) -...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/youtube.html
1227e3d833ee-2
token.write(creds.to_json()) return creds ALLOWED_SCHEMAS = {"http", "https"} ALLOWED_NETLOCK = { "youtu.be", "m.youtube.com", "youtube.com", "www.youtube.com", "www.youtube-nocookie.com", "vid.plus", } def _parse_video_id(url: str) -> Optional[str]: """Parse a youtube url and return...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/youtube.html
1227e3d833ee-3
self.add_video_info = add_video_info self.language = language if isinstance(language, str): self.language = [language] else: self.language = language self.translation = translation self.continue_on_failure = continue_on_failure [docs] @staticmethod ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/youtube.html
1227e3d833ee-4
except TranscriptsDisabled: return [] try: transcript = transcript_list.find_transcript(self.language) except NoTranscriptFound: en_transcript = transcript_list.find_transcript(["en"]) transcript = en_transcript.translate(self.translation) transcri...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/youtube.html
1227e3d833ee-5
To use, you should have the ``googleapiclient,youtube_transcript_api`` python package installed. As the service needs a google_api_client, you first have to initialize the GoogleApiClient. Additionally you have to either provide a channel name or a list of videoids "https://developers.google.com/doc...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/youtube.html
1227e3d833ee-6
"to use the Google Drive loader" ) return build("youtube", "v3", credentials=creds) [docs] @root_validator def validate_channel_or_videoIds_is_set( cls, values: Dict[str, Any] ) -> Dict[str, Any]: """Validate that either folder_id or document_ids is set, but not both.""" ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/youtube.html
1227e3d833ee-7
request = self.youtube_client.search().list( part="id", q=channel_name, type="channel", maxResults=1, # we only need one result since channel names are unique ) response = request.execute() channel_id = response["items"][0]["id"]["channelId"] ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/youtube.html
1227e3d833ee-8
metadata=meta_data, ) ) except (TranscriptsDisabled, NoTranscriptFound) as e: if self.continue_on_failure: logger.error( "Error fetching transscript " + f" {ite...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/youtube.html
98c132bf4325-0
Source code for langchain.document_loaders.modern_treasury """Loader that fetches data from Modern Treasury""" import json import urllib.request from base64 import b64encode from typing import List, Optional from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader from lan...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/modern_treasury.html
98c132bf4325-1
def __init__( self, resource: str, organization_id: Optional[str] = None, api_key: Optional[str] = None, ) -> None: self.resource = resource organization_id = organization_id or get_from_env( "organization_id", "MODERN_TREASURY_ORGANIZATION_ID" ) ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/modern_treasury.html
2385dbb1d587-0
Source code for langchain.document_loaders.twitter """Twitter document loader.""" from __future__ import annotations from typing import TYPE_CHECKING, Any, Dict, Iterable, List, Optional, Sequence, Union from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader if TYPE_CHEC...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/twitter.html
2385dbb1d587-1
user = api.get_user(screen_name=username) docs = self._format_tweets(tweets, user) results.extend(docs) return results def _format_tweets( self, tweets: List[Dict[str, Any]], user_info: dict ) -> Iterable[Document]: """Format tweets into a string.""" for t...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/twitter.html
2385dbb1d587-2
access_token=access_token, access_token_secret=access_token_secret, consumer_key=consumer_key, consumer_secret=consumer_secret, ) return cls( auth_handler=auth, twitter_users=twitter_users, number_tweets=number_tweets, )
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/twitter.html
047f0c43440f-0
Source code for langchain.document_loaders.toml import json from pathlib import Path from typing import Iterator, List, Union from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader [docs]class TomlLoader(BaseLoader): """ A TOML document loader that inherits from ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/toml.html
ca00ee739fa6-0
Source code for langchain.document_loaders.tencent_cos_directory """Loading logic for loading documents from Tencent Cloud COS directory.""" from typing import Any, Iterator, List from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader from langchain.document_loaders.tenc...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/tencent_cos_directory.html
ca00ee739fa6-1
for content in contents: if content["Key"].endswith("/"): continue loader = TencentCOSFileLoader(self.conf, self.bucket, content["Key"]) yield loader.load()[0]
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/tencent_cos_directory.html
2d76eb4740b4-0
Source code for langchain.document_loaders.pdf """Loader that loads PDF files.""" import json import logging import os import tempfile import time from abc import ABC from io import StringIO from pathlib import Path from typing import Any, Iterator, List, Mapping, Optional, Union from urllib.parse import urlparse impor...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/pdf.html
2d76eb4740b4-1
if not os.path.isfile(self.file_path) and self._is_valid_url(self.file_path): r = requests.get(self.file_path) if r.status_code != 200: raise ValueError( "Check the url of your file; returned status code %s" % r.status_code ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/pdf.html
2d76eb4740b4-2
""" def __init__( self, file_path: str, password: Optional[Union[str, bytes]] = None ) -> None: """Initialize with file path.""" try: import pypdf # noqa:F401 except ImportError: raise ImportError( "pypdf package not found, please install ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/pdf.html
2d76eb4740b4-3
Loader also stores page numbers in metadatas. """ def __init__( self, path: str, glob: str = "**/[!.]*.pdf", silent_errors: bool = False, load_hidden: bool = False, recursive: bool = False, ): self.path = path self.glob = glob self.load...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/pdf.html
2d76eb4740b4-4
raise ImportError( "`pdfminer` package not found, please install it with " "`pip install pdfminer.six`" ) super().__init__(file_path) self.parser = PDFMinerParser() [docs] def load(self) -> List[Document]: """Eagerly load the content.""" ret...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/pdf.html
2d76eb4740b4-5
) metadata = {"source": self.file_path} return [Document(page_content=output_string.getvalue(), metadata=metadata)] [docs]class PyMuPDFLoader(BasePDFLoader): """Loader that uses PyMuPDF to load PDF files.""" def __init__(self, file_path: str) -> None: """Initialize with file path.""" ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/pdf.html
2d76eb4740b4-6
kwargs, "mathpix_api_id", "MATHPIX_API_ID" ) self.processed_file_format = processed_file_format self.max_wait_time_seconds = max_wait_time_seconds self.should_clean_pdf = should_clean_pdf @property def headers(self) -> dict: return {"app_id": self.mathpix_api_id, "app_key...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/pdf.html
2d76eb4740b4-7
raise ValueError("Unable to retrieve PDF from Mathpix") else: print(f"Status: {status}, waiting for processing to complete") time.sleep(5) raise TimeoutError [docs] def get_processed_pdf(self, pdf_id: str) -> str: self.wait_for_processing(pdf_id) ur...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/pdf.html
2d76eb4740b4-8
) -> None: """Initialize with file path.""" try: import pdfplumber # noqa:F401 except ImportError: raise ImportError( "pdfplumber package not found, please install it with " "`pip install pdfplumber`" ) super().__init__...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/pdf.html
b94cbae8a157-0
Source code for langchain.document_loaders.markdown """Loader that loads Markdown files.""" from typing import List from langchain.document_loaders.unstructured import UnstructuredFileLoader [docs]class UnstructuredMarkdownLoader(UnstructuredFileLoader): """Loader that uses unstructured to load markdown files.""" ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/markdown.html
bbd45aaca7a3-0
Source code for langchain.document_loaders.dataframe """Load from Dataframe object""" from typing import Any, Iterator, List from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader [docs]class DataFrameLoader(BaseLoader): """Load Pandas DataFrames.""" def __init__...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/dataframe.html
5a6bb948480b-0
Source code for langchain.document_loaders.generic from __future__ import annotations from pathlib import Path from typing import Iterator, List, Literal, Optional, Sequence, Union from langchain.document_loaders.base import BaseBlobParser, BaseLoader from langchain.document_loaders.blob_loaders import BlobLoader, File...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/generic.html
5a6bb948480b-1
from langchain.document_loaders.parsers.pdf import PyPDFParser # Recursively load all text files in a directory. loader = GenericLoader.from_filesystem( "/path/to/dir", glob="**/*.pdf", parser=PyPDFParser() ) """ def __init__( ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/generic.html
5a6bb948480b-2
*, glob: str = "**/[!.]*", suffixes: Optional[Sequence[str]] = None, show_progress: bool = False, parser: Union[DEFAULT, BaseBlobParser] = "default", ) -> GenericLoader: """Create a generic document loader using a filesystem blob loader. Args: path: The pa...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/generic.html
7655bc962825-0
Source code for langchain.document_loaders.bigquery from __future__ import annotations from typing import TYPE_CHECKING, List, Optional from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader if TYPE_CHECKING: from google.auth.credentials import Credentials [docs]clas...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/bigquery.html
7655bc962825-1
self.project = project self.page_content_columns = page_content_columns self.metadata_columns = metadata_columns self.credentials = credentials [docs] def load(self) -> List[Document]: try: from google.cloud import bigquery except ImportError as ex: rai...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/bigquery.html
8964f0de1c8f-0
Source code for langchain.document_loaders.mastodon """Mastodon document loader.""" from __future__ import annotations import os from typing import TYPE_CHECKING, Any, Dict, Iterable, List, Optional, Sequence from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader if TYPE...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/mastodon.html
8964f0de1c8f-1
access_token = access_token or os.environ.get("MASTODON_ACCESS_TOKEN") self.api = mastodon.Mastodon( access_token=access_token, api_base_url=api_base_url ) self.mastodon_accounts = mastodon_accounts self.number_toots = number_toots self.exclude_replies = exclude_repli...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/mastodon.html
53880898c7c5-0
Source code for langchain.document_loaders.iugu """Loader that fetches data from IUGU""" import json import urllib.request from typing import List, Optional from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader from langchain.utils import get_from_env, stringify_dict IU...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/iugu.html
53880898c7c5-1
[docs] def load(self) -> List[Document]: return self._get_resource()
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/iugu.html
c003b5158f19-0
Source code for langchain.document_loaders.snowflake_loader from __future__ import annotations from typing import Any, Dict, Iterator, List, Optional, Tuple from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader [docs]class SnowflakeLoader(BaseLoader): """Loads a que...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/snowflake_loader.html
c003b5158f19-1
self.password = password self.account = account self.warehouse = warehouse self.role = role self.database = database self.schema = schema self.parameters = parameters self.page_content_columns = ( page_content_columns if page_content_columns is not Non...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/snowflake_loader.html
c003b5158f19-2
) -> Tuple[List[str], List[str]]: page_content_columns = ( self.page_content_columns if self.page_content_columns else [] ) metadata_columns = self.metadata_columns if self.metadata_columns else [] if page_content_columns is None and query_result: page_content_col...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/snowflake_loader.html
8a7fd2ccf1f9-0
Source code for langchain.document_loaders.s3_directory """Loading logic for loading documents from an s3 directory.""" from typing import List from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader from langchain.document_loaders.s3_file import S3FileLoader [docs]class ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/s3_directory.html
7e338c5724ab-0
Source code for langchain.document_loaders.unstructured """Loader that uses unstructured to load files.""" import collections from abc import ABC, abstractmethod from typing import IO, Any, Dict, List, Sequence, Union from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoade...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/unstructured.html
7e338c5724ab-1
import unstructured # noqa:F401 except ImportError: raise ValueError( "unstructured package not found, please install it with " "`pip install unstructured`" ) _valid_modes = {"single", "elements", "paged"} if mode not in _valid_modes: ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/unstructured.html
7e338c5724ab-2
for idx, element in enumerate(elements): metadata = self._get_metadata() if hasattr(element, "metadata"): metadata.update(element.metadata.to_dict()) page_number = metadata.get("page_number", 1) # Check if this page_number already exist...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/unstructured.html
7e338c5724ab-3
def _get_elements(self) -> List: from unstructured.partition.auto import partition return partition(filename=self.file_path, **self.unstructured_kwargs) def _get_metadata(self) -> dict: return {"source": self.file_path} [docs]def get_elements_from_api( file_path: Union[str, List[str], No...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/unstructured.html
7e338c5724ab-4
url: str = "https://api.unstructured.io/general/v0/general", api_key: str = "", **unstructured_kwargs: Any, ): """Initialize with file path.""" if isinstance(file_path, str): validate_unstructured_version(min_unstructured_version="0.6.2") else: validat...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/unstructured.html
7e338c5724ab-5
def __init__( self, file: Union[IO, Sequence[IO]], mode: str = "single", url: str = "https://api.unstructured.io/general/v0/general", api_key: str = "", **unstructured_kwargs: Any, ): """Initialize with file path.""" if isinstance(file, collections.abc...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/unstructured.html
c4e771a67bb1-0
Source code for langchain.document_loaders.merge from typing import Iterator, List from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader [docs]class MergedDataLoader(BaseLoader): """Merge documents from a list of loaders""" def __init__(self, loaders: List): ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/merge.html
f1f78b2b45e8-0
Source code for langchain.document_loaders.apify_dataset """Logic for loading documents from Apify datasets.""" from typing import Any, Callable, Dict, List from pydantic import BaseModel, root_validator from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader [docs]class ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/apify_dataset.html
f1f78b2b45e8-1
) return values [docs] def load(self) -> List[Document]: """Load documents.""" dataset_items = ( self.apify_client.dataset(self.dataset_id).list_items(clean=True).items ) return list(map(self.dataset_mapping_function, dataset_items))
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/apify_dataset.html
40fb9615847c-0
Source code for langchain.document_loaders.hugging_face_dataset """Loader that loads HuggingFace datasets.""" from typing import Iterator, List, Mapping, Optional, Sequence, Union from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader [docs]class HuggingFaceDatasetLoader...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/hugging_face_dataset.html
40fb9615847c-1
self.page_content_column = page_content_column self.name = name self.data_dir = data_dir self.data_files = data_files self.cache_dir = cache_dir self.keep_in_memory = keep_in_memory self.save_infos = save_infos self.use_auth_token = use_auth_token self.num...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/hugging_face_dataset.html
90a1e6df9f1e-0
Source code for langchain.document_loaders.epub """Loader that loads EPub files.""" from typing import List from langchain.document_loaders.unstructured import ( UnstructuredFileLoader, satisfies_min_unstructured_version, ) [docs]class UnstructuredEPubLoader(UnstructuredFileLoader): """Loader that uses unst...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/epub.html
e408e4b72331-0
Source code for langchain.document_loaders.discord """Load from Discord chat dump""" from __future__ import annotations from typing import TYPE_CHECKING, List from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader if TYPE_CHECKING: import pandas as pd [docs]class Dis...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/discord.html