id stringlengths 14 15 | text stringlengths 35 2.51k | source stringlengths 61 154 |
|---|---|---|
00b928ef36c7-0 | Source code for langchain.output_parsers.rail_parser
from __future__ import annotations
from typing import Any, Callable, Dict, Optional
from langchain.schema import BaseOutputParser
[docs]class GuardrailsOutputParser(BaseOutputParser):
guard: Any
api: Optional[Callable]
args: Any
kwargs: Any
@prope... | https://api.python.langchain.com/en/latest/_modules/langchain/output_parsers/rail_parser.html |
00b928ef36c7-1 | )
return cls(
guard=Guard.from_rail_string(rail_str, num_reasks=num_reasks),
api=api,
args=args,
kwargs=kwargs,
)
[docs] @classmethod
def from_pydantic(
cls,
output_class: Any,
num_reasks: int = 1,
api: Optional[Calla... | https://api.python.langchain.com/en/latest/_modules/langchain/output_parsers/rail_parser.html |
1c7528234fbc-0 | Source code for langchain.document_loaders.blockchain
import os
import re
import time
from enum import Enum
from typing import List, Optional
import requests
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
[docs]class BlockchainType(Enum):
"""Enumerator of the... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/blockchain.html |
1c7528234fbc-1 | """
def __init__(
self,
contract_address: str,
blockchainType: BlockchainType = BlockchainType.ETH_MAINNET,
api_key: str = "docs-demo",
startToken: str = "",
get_all_tokens: bool = False,
max_execution_time: Optional[int] = None,
):
self.contract_a... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/blockchain.html |
1c7528234fbc-2 | tokenId = item["id"]["tokenId"]
metadata = {
"source": self.contract_address,
"blockchain": self.blockchainType,
"tokenId": tokenId,
}
result.append(Document(page_content=content, metadata=metadata))
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/blockchain.html |
1c7528234fbc-3 | elif value_type == "hex_0xbf":
return "0xbf" + format(result, "0" + str(len(tokenId) - 4) + "x")
else:
return str(result)
# A smart contract can use different formats for the tokenId
@staticmethod
def _detect_value_type(tokenId: str) -> str:
if isinstance(tokenId, int... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/blockchain.html |
27ada31af51a-0 | Source code for langchain.document_loaders.airtable
from typing import Iterator, List
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
[docs]class AirtableLoader(BaseLoader):
"""Loader for Airtable tables."""
def __init__(self, api_token: str, table_id: str... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/airtable.html |
7b23cafd55cc-0 | Source code for langchain.document_loaders.sitemap
"""Loader that fetches a sitemap and loads those URLs."""
import itertools
import re
from typing import Any, Callable, Generator, Iterable, List, Optional
from langchain.document_loaders.web_base import WebBaseLoader
from langchain.schema import Document
def _default_p... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/sitemap.html |
7b23cafd55cc-1 | meta_function: Function to parse bs4.Soup output for metadata
remember when setting this method to also copy metadata["loc"]
to metadata["source"] if you are using this field
is_local: whether the sitemap is a local file
"""
if blocksize is not None and blocks... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/sitemap.html |
7b23cafd55cc-2 | if (prop := url.find(tag))
}
)
for sitemap in soup.find_all("sitemap"):
loc = sitemap.find("loc")
if not loc:
continue
soup_child = self.scrape_all([loc.text], "xml")[0]
els.extend(self.parse_sitemap(soup_child))
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/sitemap.html |
bb6b466b38c2-0 | Source code for langchain.document_loaders.notiondb
"""Notion DB loader for langchain"""
from typing import Any, Dict, List, Optional
import requests
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
NOTION_BASE_URL = "https://api.notion.com/v1"
DATABASE_URL = NOTIO... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/notiondb.html |
bb6b466b38c2-1 | Returns:
List[Document]: List of documents.
"""
page_summaries = self._retrieve_page_summaries()
return list(self.load_page(page_summary) for page_summary in page_summaries)
def _retrieve_page_summaries(
self, query_dict: Dict[str, Any] = {"page_size": 100}
) -> List[... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/notiondb.html |
bb6b466b38c2-2 | if prop_data["multi_select"]
else []
)
elif prop_type == "url":
value = prop_data["url"]
else:
value = None
metadata[prop_name.lower()] = value
metadata["id"] = page_id
return Document(page_content=se... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/notiondb.html |
bb6b466b38c2-3 | method,
url,
headers=self.headers,
json=query_dict,
timeout=self.request_timeout_sec,
)
res.raise_for_status()
return res.json() | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/notiondb.html |
e7da210922b2-0 | Source code for langchain.document_loaders.larksuite
"""Loader that loads LarkSuite (FeiShu) document json dump."""
import json
import urllib.request
from typing import Any, Iterator, List
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
[docs]class LarkSuiteDocLoa... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/larksuite.html |
e7da210922b2-1 | metadata = {
"document_id": self.document_id,
"revision_id": metadata_json["data"]["document"]["revision_id"],
"title": metadata_json["data"]["document"]["title"],
}
yield Document(page_content=text, metadata=metadata)
[docs] def load(self) -> List[Document]:
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/larksuite.html |
5f2eb59bc49c-0 | Source code for langchain.document_loaders.trello
"""Loader that loads cards from Trello"""
from __future__ import annotations
from typing import TYPE_CHECKING, Any, List, Literal, Optional, Tuple
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
from langchain.util... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/trello.html |
5f2eb59bc49c-1 | self.board_name = board_name
self.include_card_name = include_card_name
self.include_comments = include_comments
self.include_checklist = include_checklist
self.extra_metadata = extra_metadata
self.card_filter = card_filter
[docs] @classmethod
def from_credentials(
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/trello.html |
5f2eb59bc49c-2 | token = token or get_from_env("token", "TRELLO_TOKEN")
client = TrelloClient(api_key=api_key, token=token)
return cls(client, board_name, **kwargs)
[docs] def load(self) -> List[Document]:
"""Loads all cards from the specified Trello board.
You can filter the cards, metadata and text ... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/trello.html |
5f2eb59bc49c-3 | if self.include_card_name:
text_content = card.name + "\n"
if card.description.strip():
text_content += BeautifulSoup(card.description, "lxml").get_text()
if self.include_checklist:
# Get all the checklist items on the card
for checklist in card.checklists... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/trello.html |
aba0121c9013-0 | Source code for langchain.document_loaders.gcs_directory
"""Loading logic for loading documents from an GCS directory."""
from typing import List
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
from langchain.document_loaders.gcs_file import GCSFileLoader
[docs]cl... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/gcs_directory.html |
36e7751bc8e6-0 | Source code for langchain.document_loaders.acreom
"""Loader that loads acreom vault from a directory."""
import re
from pathlib import Path
from typing import Iterator, List
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
[docs]class AcreomLoader(BaseLoader):
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/acreom.html |
36e7751bc8e6-1 | # do not contribute to the context of current document
content = re.sub("\s*-\s\[\s\]\s.*|\s*\[\s\]\s.*", "", content) # rm tasks
content = re.sub("#", "", content) # rm hashtags
content = re.sub("\[\[.*?\]\]", "", content) # rm doclinks
return content
[docs] def lazy_load(self) ->... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/acreom.html |
f0d5cd018459-0 | Source code for langchain.document_loaders.embaas
import base64
import warnings
from typing import Any, Dict, Iterator, List, Optional
import requests
from pydantic import BaseModel, root_validator, validator
from typing_extensions import NotRequired, TypedDict
from langchain.docstore.document import Document
from lang... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/embaas.html |
f0d5cd018459-1 | """The instruction to pass to the Embaas document extraction API."""
[docs]class EmbaasDocumentExtractionPayload(EmbaasDocumentExtractionParameters):
"""Payload for the Embaas document extraction API."""
bytes: str
"""The base64 encoded bytes of the document to extract text from."""
[docs]class BaseEmbaasLo... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/embaas.html |
f0d5cd018459-2 | documents = loader.parse(blob=blob)
# Custom api parameters (create embeddings automatically)
from langchain.document_loaders.embaas import EmbaasBlobLoader
loader = EmbaasBlobLoader(
params={
"should_embed": True,
"model": "e5-... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/embaas.html |
f0d5cd018459-3 | # type: ignore
**self.params,
)
if blob.mimetype is not None and payload.get("mime_type", None) is None:
payload["mime_type"] = blob.mimetype
return payload
def _handle_request(
self, payload: EmbaasDocumentExtractionPayload
) -> List[Document]:
""... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/embaas.html |
f0d5cd018459-4 | To use, you should have the
environment variable ``EMBAAS_API_KEY`` set with your API key, or pass
it as a named parameter to the constructor.
Example:
.. code-block:: python
# Default parsing
from langchain.document_loaders.embaas import EmbaasLoader
loader = Emb... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/embaas.html |
f0d5cd018459-5 | assert self.blob_loader is not None
# Should never be None, but mypy doesn't know that.
yield from self.blob_loader.lazy_parse(blob=blob)
[docs] def load(self) -> List[Document]:
return list(self.lazy_load())
[docs] def load_and_split(
self, text_splitter: Optional[TextSplitter] = ... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/embaas.html |
e5429edf82f2-0 | Source code for langchain.document_loaders.college_confidential
"""Loader that loads College Confidential."""
from typing import List
from langchain.docstore.document import Document
from langchain.document_loaders.web_base import WebBaseLoader
[docs]class CollegeConfidentialLoader(WebBaseLoader):
"""Loader that lo... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/college_confidential.html |
05ec8c196b0c-0 | Source code for langchain.document_loaders.googledrive
"""Loader that loads data from Google Drive."""
# Prerequisites:
# 1. Create a Google Cloud project
# 2. Enable the Google Drive API:
# https://console.cloud.google.com/flows/enableapi?apiid=drive.googleapis.com
# 3. Authorize credentials for desktop app:
# htt... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/googledrive.html |
05ec8c196b0c-1 | # results in pydantic validation errors
file_loader_cls: Any = None
file_loader_kwargs: Dict["str", Any] = {}
[docs] @root_validator
def validate_inputs(cls, values: Dict[str, Any]) -> Dict[str, Any]:
"""Validate that either folder_id or document_ids is set, but not both."""
if values.get... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/googledrive.html |
05ec8c196b0c-2 | if file_type not in allowed_types:
raise ValueError(
f"Given file type {file_type} is not supported. "
f"Supported values are: {short_names}; and "
f"their full-form names: {full_names}"
)
# repla... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/googledrive.html |
05ec8c196b0c-3 | )
if self.token_path.exists():
creds = Credentials.from_authorized_user_file(str(self.token_path), SCOPES)
if not creds or not creds.valid:
if creds and creds.expired and creds.refresh_token:
creds.refresh(Request())
elif "GOOGLE_APPLICATION_CREDENTIAL... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/googledrive.html |
05ec8c196b0c-4 | metadata = {
"source": (
f"https://docs.google.com/spreadsheets/d/{id}/"
f"edit?gid={sheet['properties']['sheetId']}"
),
"title": f"{spreadsheet['properties']['title']} - {sheet_name}",
"row":... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/googledrive.html |
05ec8c196b0c-5 | text = fh.getvalue().decode("utf-8")
metadata = {
"source": f"https://docs.google.com/document/d/{id}/edit",
"title": f"{file.get('name')}",
}
return Document(page_content=text, metadata=metadata)
def _load_documents_from_folder(
self, folder_id: str, *, file_... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/googledrive.html |
05ec8c196b0c-6 | else:
pass
return returns
def _fetch_files_recursive(
self, service: Any, folder_id: str
) -> List[Dict[str, Union[str, List[str]]]]:
"""Fetch all files and subfolders recursively."""
results = (
service.files()
.list(
q=f"'... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/googledrive.html |
05ec8c196b0c-7 | file = service.files().get(fileId=id, supportsAllDrives=True).execute()
request = service.files().get_media(fileId=id)
fh = BytesIO()
downloader = MediaIoBaseDownload(fh, request)
done = False
while done is False:
status, done = downloader.next_chunk()
if self... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/googledrive.html |
05ec8c196b0c-8 | )
elif self.document_ids:
return self._load_documents_from_ids()
else:
return self._load_file_from_ids() | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/googledrive.html |
568944512e6f-0 | Source code for langchain.document_loaders.stripe
"""Loader that fetches data from Stripe"""
import json
import urllib.request
from typing import List, Optional
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
from langchain.utils import get_from_env, stringify_dic... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/stripe.html |
568944512e6f-1 | if endpoint is None:
return []
return self._make_request(endpoint)
[docs] def load(self) -> List[Document]:
return self._get_resource() | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/stripe.html |
9ca4d2d127b8-0 | Source code for langchain.document_loaders.fauna
from typing import Iterator, List, Optional, Sequence
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
[docs]class FaunaLoader(BaseLoader):
"""FaunaDB Loader.
Attributes:
query (str): The FQL query st... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/fauna.html |
9ca4d2d127b8-1 | document_dict = dict(result.items())
page_content = ""
for key, value in document_dict.items():
if key == self.page_content_field:
page_content = value
document: Document = Document(
page_content=page_content... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/fauna.html |
a4d0ecf1935d-0 | Source code for langchain.document_loaders.duckdb_loader
from typing import Dict, List, Optional, cast
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
[docs]class DuckDBLoader(BaseLoader):
"""Loads a query result from DuckDB into a list of documents.
Each ... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/duckdb_loader.html |
a4d0ecf1935d-1 | results = query_result.fetchall()
description = cast(list, query_result.description)
field_names = [c[0] for c in description]
if self.page_content_columns is None:
page_content_columns = field_names
else:
page_content_columns = self.page_c... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/duckdb_loader.html |
60aaa137ac4b-0 | Source code for langchain.document_loaders.gutenberg
"""Loader that loads .txt web files."""
from typing import List
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
[docs]class GutenbergLoader(BaseLoader):
"""Loader that uses urllib to load .txt web files."""
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/gutenberg.html |
1227e3d833ee-0 | Source code for langchain.document_loaders.youtube
"""Loader that loads YouTube transcript."""
from __future__ import annotations
import logging
from pathlib import Path
from typing import Any, Dict, List, Optional, Sequence, Union
from urllib.parse import parse_qs, urlparse
from pydantic import root_validator
from pyd... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/youtube.html |
1227e3d833ee-1 | """Validate that either folder_id or document_ids is set, but not both."""
if not values.get("credentials_path") and not values.get(
"service_account_path"
):
raise ValueError("Must specify either channel_name or video_ids")
return values
def _load_credentials(self) -... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/youtube.html |
1227e3d833ee-2 | token.write(creds.to_json())
return creds
ALLOWED_SCHEMAS = {"http", "https"}
ALLOWED_NETLOCK = {
"youtu.be",
"m.youtube.com",
"youtube.com",
"www.youtube.com",
"www.youtube-nocookie.com",
"vid.plus",
}
def _parse_video_id(url: str) -> Optional[str]:
"""Parse a youtube url and return... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/youtube.html |
1227e3d833ee-3 | self.add_video_info = add_video_info
self.language = language
if isinstance(language, str):
self.language = [language]
else:
self.language = language
self.translation = translation
self.continue_on_failure = continue_on_failure
[docs] @staticmethod
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/youtube.html |
1227e3d833ee-4 | except TranscriptsDisabled:
return []
try:
transcript = transcript_list.find_transcript(self.language)
except NoTranscriptFound:
en_transcript = transcript_list.find_transcript(["en"])
transcript = en_transcript.translate(self.translation)
transcri... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/youtube.html |
1227e3d833ee-5 | To use, you should have the ``googleapiclient,youtube_transcript_api``
python package installed.
As the service needs a google_api_client, you first have to initialize
the GoogleApiClient.
Additionally you have to either provide a channel name or a list of videoids
"https://developers.google.com/doc... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/youtube.html |
1227e3d833ee-6 | "to use the Google Drive loader"
)
return build("youtube", "v3", credentials=creds)
[docs] @root_validator
def validate_channel_or_videoIds_is_set(
cls, values: Dict[str, Any]
) -> Dict[str, Any]:
"""Validate that either folder_id or document_ids is set, but not both."""
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/youtube.html |
1227e3d833ee-7 | request = self.youtube_client.search().list(
part="id",
q=channel_name,
type="channel",
maxResults=1, # we only need one result since channel names are unique
)
response = request.execute()
channel_id = response["items"][0]["id"]["channelId"]
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/youtube.html |
1227e3d833ee-8 | metadata=meta_data,
)
)
except (TranscriptsDisabled, NoTranscriptFound) as e:
if self.continue_on_failure:
logger.error(
"Error fetching transscript "
+ f" {ite... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/youtube.html |
98c132bf4325-0 | Source code for langchain.document_loaders.modern_treasury
"""Loader that fetches data from Modern Treasury"""
import json
import urllib.request
from base64 import b64encode
from typing import List, Optional
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
from lan... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/modern_treasury.html |
98c132bf4325-1 | def __init__(
self,
resource: str,
organization_id: Optional[str] = None,
api_key: Optional[str] = None,
) -> None:
self.resource = resource
organization_id = organization_id or get_from_env(
"organization_id", "MODERN_TREASURY_ORGANIZATION_ID"
)
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/modern_treasury.html |
2385dbb1d587-0 | Source code for langchain.document_loaders.twitter
"""Twitter document loader."""
from __future__ import annotations
from typing import TYPE_CHECKING, Any, Dict, Iterable, List, Optional, Sequence, Union
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
if TYPE_CHEC... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/twitter.html |
2385dbb1d587-1 | user = api.get_user(screen_name=username)
docs = self._format_tweets(tweets, user)
results.extend(docs)
return results
def _format_tweets(
self, tweets: List[Dict[str, Any]], user_info: dict
) -> Iterable[Document]:
"""Format tweets into a string."""
for t... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/twitter.html |
2385dbb1d587-2 | access_token=access_token,
access_token_secret=access_token_secret,
consumer_key=consumer_key,
consumer_secret=consumer_secret,
)
return cls(
auth_handler=auth,
twitter_users=twitter_users,
number_tweets=number_tweets,
) | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/twitter.html |
047f0c43440f-0 | Source code for langchain.document_loaders.toml
import json
from pathlib import Path
from typing import Iterator, List, Union
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
[docs]class TomlLoader(BaseLoader):
"""
A TOML document loader that inherits from ... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/toml.html |
ca00ee739fa6-0 | Source code for langchain.document_loaders.tencent_cos_directory
"""Loading logic for loading documents from Tencent Cloud COS directory."""
from typing import Any, Iterator, List
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
from langchain.document_loaders.tenc... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/tencent_cos_directory.html |
ca00ee739fa6-1 | for content in contents:
if content["Key"].endswith("/"):
continue
loader = TencentCOSFileLoader(self.conf, self.bucket, content["Key"])
yield loader.load()[0] | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/tencent_cos_directory.html |
2d76eb4740b4-0 | Source code for langchain.document_loaders.pdf
"""Loader that loads PDF files."""
import json
import logging
import os
import tempfile
import time
from abc import ABC
from io import StringIO
from pathlib import Path
from typing import Any, Iterator, List, Mapping, Optional, Union
from urllib.parse import urlparse
impor... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/pdf.html |
2d76eb4740b4-1 | if not os.path.isfile(self.file_path) and self._is_valid_url(self.file_path):
r = requests.get(self.file_path)
if r.status_code != 200:
raise ValueError(
"Check the url of your file; returned status code %s"
% r.status_code
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/pdf.html |
2d76eb4740b4-2 | """
def __init__(
self, file_path: str, password: Optional[Union[str, bytes]] = None
) -> None:
"""Initialize with file path."""
try:
import pypdf # noqa:F401
except ImportError:
raise ImportError(
"pypdf package not found, please install ... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/pdf.html |
2d76eb4740b4-3 | Loader also stores page numbers in metadatas.
"""
def __init__(
self,
path: str,
glob: str = "**/[!.]*.pdf",
silent_errors: bool = False,
load_hidden: bool = False,
recursive: bool = False,
):
self.path = path
self.glob = glob
self.load... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/pdf.html |
2d76eb4740b4-4 | raise ImportError(
"`pdfminer` package not found, please install it with "
"`pip install pdfminer.six`"
)
super().__init__(file_path)
self.parser = PDFMinerParser()
[docs] def load(self) -> List[Document]:
"""Eagerly load the content."""
ret... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/pdf.html |
2d76eb4740b4-5 | )
metadata = {"source": self.file_path}
return [Document(page_content=output_string.getvalue(), metadata=metadata)]
[docs]class PyMuPDFLoader(BasePDFLoader):
"""Loader that uses PyMuPDF to load PDF files."""
def __init__(self, file_path: str) -> None:
"""Initialize with file path."""
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/pdf.html |
2d76eb4740b4-6 | kwargs, "mathpix_api_id", "MATHPIX_API_ID"
)
self.processed_file_format = processed_file_format
self.max_wait_time_seconds = max_wait_time_seconds
self.should_clean_pdf = should_clean_pdf
@property
def headers(self) -> dict:
return {"app_id": self.mathpix_api_id, "app_key... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/pdf.html |
2d76eb4740b4-7 | raise ValueError("Unable to retrieve PDF from Mathpix")
else:
print(f"Status: {status}, waiting for processing to complete")
time.sleep(5)
raise TimeoutError
[docs] def get_processed_pdf(self, pdf_id: str) -> str:
self.wait_for_processing(pdf_id)
ur... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/pdf.html |
2d76eb4740b4-8 | ) -> None:
"""Initialize with file path."""
try:
import pdfplumber # noqa:F401
except ImportError:
raise ImportError(
"pdfplumber package not found, please install it with "
"`pip install pdfplumber`"
)
super().__init__... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/pdf.html |
b94cbae8a157-0 | Source code for langchain.document_loaders.markdown
"""Loader that loads Markdown files."""
from typing import List
from langchain.document_loaders.unstructured import UnstructuredFileLoader
[docs]class UnstructuredMarkdownLoader(UnstructuredFileLoader):
"""Loader that uses unstructured to load markdown files."""
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/markdown.html |
bbd45aaca7a3-0 | Source code for langchain.document_loaders.dataframe
"""Load from Dataframe object"""
from typing import Any, Iterator, List
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
[docs]class DataFrameLoader(BaseLoader):
"""Load Pandas DataFrames."""
def __init__... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/dataframe.html |
5a6bb948480b-0 | Source code for langchain.document_loaders.generic
from __future__ import annotations
from pathlib import Path
from typing import Iterator, List, Literal, Optional, Sequence, Union
from langchain.document_loaders.base import BaseBlobParser, BaseLoader
from langchain.document_loaders.blob_loaders import BlobLoader, File... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/generic.html |
5a6bb948480b-1 | from langchain.document_loaders.parsers.pdf import PyPDFParser
# Recursively load all text files in a directory.
loader = GenericLoader.from_filesystem(
"/path/to/dir",
glob="**/*.pdf",
parser=PyPDFParser()
)
"""
def __init__(
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/generic.html |
5a6bb948480b-2 | *,
glob: str = "**/[!.]*",
suffixes: Optional[Sequence[str]] = None,
show_progress: bool = False,
parser: Union[DEFAULT, BaseBlobParser] = "default",
) -> GenericLoader:
"""Create a generic document loader using a filesystem blob loader.
Args:
path: The pa... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/generic.html |
7655bc962825-0 | Source code for langchain.document_loaders.bigquery
from __future__ import annotations
from typing import TYPE_CHECKING, List, Optional
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
if TYPE_CHECKING:
from google.auth.credentials import Credentials
[docs]clas... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/bigquery.html |
7655bc962825-1 | self.project = project
self.page_content_columns = page_content_columns
self.metadata_columns = metadata_columns
self.credentials = credentials
[docs] def load(self) -> List[Document]:
try:
from google.cloud import bigquery
except ImportError as ex:
rai... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/bigquery.html |
8964f0de1c8f-0 | Source code for langchain.document_loaders.mastodon
"""Mastodon document loader."""
from __future__ import annotations
import os
from typing import TYPE_CHECKING, Any, Dict, Iterable, List, Optional, Sequence
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
if TYPE... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/mastodon.html |
8964f0de1c8f-1 | access_token = access_token or os.environ.get("MASTODON_ACCESS_TOKEN")
self.api = mastodon.Mastodon(
access_token=access_token, api_base_url=api_base_url
)
self.mastodon_accounts = mastodon_accounts
self.number_toots = number_toots
self.exclude_replies = exclude_repli... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/mastodon.html |
53880898c7c5-0 | Source code for langchain.document_loaders.iugu
"""Loader that fetches data from IUGU"""
import json
import urllib.request
from typing import List, Optional
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
from langchain.utils import get_from_env, stringify_dict
IU... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/iugu.html |
53880898c7c5-1 | [docs] def load(self) -> List[Document]:
return self._get_resource() | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/iugu.html |
c003b5158f19-0 | Source code for langchain.document_loaders.snowflake_loader
from __future__ import annotations
from typing import Any, Dict, Iterator, List, Optional, Tuple
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
[docs]class SnowflakeLoader(BaseLoader):
"""Loads a que... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/snowflake_loader.html |
c003b5158f19-1 | self.password = password
self.account = account
self.warehouse = warehouse
self.role = role
self.database = database
self.schema = schema
self.parameters = parameters
self.page_content_columns = (
page_content_columns if page_content_columns is not Non... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/snowflake_loader.html |
c003b5158f19-2 | ) -> Tuple[List[str], List[str]]:
page_content_columns = (
self.page_content_columns if self.page_content_columns else []
)
metadata_columns = self.metadata_columns if self.metadata_columns else []
if page_content_columns is None and query_result:
page_content_col... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/snowflake_loader.html |
8a7fd2ccf1f9-0 | Source code for langchain.document_loaders.s3_directory
"""Loading logic for loading documents from an s3 directory."""
from typing import List
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
from langchain.document_loaders.s3_file import S3FileLoader
[docs]class ... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/s3_directory.html |
7e338c5724ab-0 | Source code for langchain.document_loaders.unstructured
"""Loader that uses unstructured to load files."""
import collections
from abc import ABC, abstractmethod
from typing import IO, Any, Dict, List, Sequence, Union
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoade... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/unstructured.html |
7e338c5724ab-1 | import unstructured # noqa:F401
except ImportError:
raise ValueError(
"unstructured package not found, please install it with "
"`pip install unstructured`"
)
_valid_modes = {"single", "elements", "paged"}
if mode not in _valid_modes:
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/unstructured.html |
7e338c5724ab-2 | for idx, element in enumerate(elements):
metadata = self._get_metadata()
if hasattr(element, "metadata"):
metadata.update(element.metadata.to_dict())
page_number = metadata.get("page_number", 1)
# Check if this page_number already exist... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/unstructured.html |
7e338c5724ab-3 | def _get_elements(self) -> List:
from unstructured.partition.auto import partition
return partition(filename=self.file_path, **self.unstructured_kwargs)
def _get_metadata(self) -> dict:
return {"source": self.file_path}
[docs]def get_elements_from_api(
file_path: Union[str, List[str], No... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/unstructured.html |
7e338c5724ab-4 | url: str = "https://api.unstructured.io/general/v0/general",
api_key: str = "",
**unstructured_kwargs: Any,
):
"""Initialize with file path."""
if isinstance(file_path, str):
validate_unstructured_version(min_unstructured_version="0.6.2")
else:
validat... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/unstructured.html |
7e338c5724ab-5 | def __init__(
self,
file: Union[IO, Sequence[IO]],
mode: str = "single",
url: str = "https://api.unstructured.io/general/v0/general",
api_key: str = "",
**unstructured_kwargs: Any,
):
"""Initialize with file path."""
if isinstance(file, collections.abc... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/unstructured.html |
c4e771a67bb1-0 | Source code for langchain.document_loaders.merge
from typing import Iterator, List
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
[docs]class MergedDataLoader(BaseLoader):
"""Merge documents from a list of loaders"""
def __init__(self, loaders: List):
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/merge.html |
f1f78b2b45e8-0 | Source code for langchain.document_loaders.apify_dataset
"""Logic for loading documents from Apify datasets."""
from typing import Any, Callable, Dict, List
from pydantic import BaseModel, root_validator
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
[docs]class ... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/apify_dataset.html |
f1f78b2b45e8-1 | )
return values
[docs] def load(self) -> List[Document]:
"""Load documents."""
dataset_items = (
self.apify_client.dataset(self.dataset_id).list_items(clean=True).items
)
return list(map(self.dataset_mapping_function, dataset_items)) | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/apify_dataset.html |
40fb9615847c-0 | Source code for langchain.document_loaders.hugging_face_dataset
"""Loader that loads HuggingFace datasets."""
from typing import Iterator, List, Mapping, Optional, Sequence, Union
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
[docs]class HuggingFaceDatasetLoader... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/hugging_face_dataset.html |
40fb9615847c-1 | self.page_content_column = page_content_column
self.name = name
self.data_dir = data_dir
self.data_files = data_files
self.cache_dir = cache_dir
self.keep_in_memory = keep_in_memory
self.save_infos = save_infos
self.use_auth_token = use_auth_token
self.num... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/hugging_face_dataset.html |
90a1e6df9f1e-0 | Source code for langchain.document_loaders.epub
"""Loader that loads EPub files."""
from typing import List
from langchain.document_loaders.unstructured import (
UnstructuredFileLoader,
satisfies_min_unstructured_version,
)
[docs]class UnstructuredEPubLoader(UnstructuredFileLoader):
"""Loader that uses unst... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/epub.html |
e408e4b72331-0 | Source code for langchain.document_loaders.discord
"""Load from Discord chat dump"""
from __future__ import annotations
from typing import TYPE_CHECKING, List
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
if TYPE_CHECKING:
import pandas as pd
[docs]class Dis... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/discord.html |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.