id stringlengths 14 15 | text stringlengths 35 2.51k | source stringlengths 61 154 |
|---|---|---|
159ac61bed59-3 | data = res.json()
text = "\n".join(
[
data[key]
for key in ["title", "description", "contents_raw"]
if key in data
]
).strip()
metadata = {"source": self.web_path, "title": data["title"]}
documents.append(Document(pa... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/ifixit.html |
159ac61bed59-4 | else:
for part in data["parts"]:
doc_parts.append("\n - " + part["text"])
for row in data["steps"]:
doc_parts.append(
"\n\n## "
+ (
row["title"]
if row["title"] != ""
else "Step {}... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/ifixit.html |
c31fe87c2f6b-0 | Source code for langchain.document_loaders.url
"""Loader that uses unstructured to load HTML files."""
import logging
from typing import Any, List
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
logger = logging.getLogger(__name__)
[docs]class UnstructuredURLLoade... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/url.html |
c31fe87c2f6b-1 | self.unstructured_kwargs = unstructured_kwargs
self.show_progress_bar = show_progress_bar
def _validate_mode(self, mode: str) -> None:
_valid_modes = {"single", "elements"}
if mode not in _valid_modes:
raise ValueError(
f"Got {mode} for `mode`, but should be one o... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/url.html |
c31fe87c2f6b-2 | "Please install with 'pip install tqdm' or set "
"show_progress_bar=False."
) from e
urls = tqdm(self.urls)
else:
urls = self.urls
for url in urls:
try:
if self.__is_non_html_available():
if self.... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/url.html |
e1b0f462dec1-0 | Source code for langchain.document_loaders.docugami
"""Loader that loads processed documents from Docugami."""
import io
import logging
import os
import re
from pathlib import Path
from typing import Any, Dict, List, Mapping, Optional, Sequence, Union
import requests
from pydantic import BaseModel, root_validator
from ... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/docugami.html |
e1b0f462dec1-1 | if values.get("file_paths") and values.get("docset_id"):
raise ValueError("Cannot specify both file_paths and remote API docset_id")
if not values.get("file_paths") and not values.get("docset_id"):
raise ValueError("Must specify either file_paths or remote API docset_id")
if valu... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/docugami.html |
e1b0f462dec1-2 | ancestor_chain = chunk.xpath("ancestor-or-self::*")
return "/" + "/".join(_xpath_qname_for_chunk(x) for x in ancestor_chain)
def _structure_value(node: Any) -> str:
"""Get the structure value for a node."""
structure = (
"table"
if node.tag == ... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/docugami.html |
e1b0f462dec1-3 | """Create a Document from a node and text."""
metadata = {
XPATH_KEY: _xpath_for_chunk(node),
DOCUMENT_ID_KEY: document["id"],
DOCUMENT_NAME_KEY: document["name"],
STRUCTURE_KEY: node.attrib.get("structure", ""),
TAG_KEY: re.sub... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/docugami.html |
e1b0f462dec1-4 | while url:
response = requests.get(
url,
headers={"Authorization": f"Bearer {self.access_token}"},
)
if response.ok:
data = response.json()
all_documents.extend(data["documents"])
url = data.get("next", N... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/docugami.html |
e1b0f462dec1-5 | data={},
)
if response.ok:
data = response.json()
all_artifacts.extend(data["artifacts"])
url = data.get("next", None)
else:
raise Exception(
f"Failed to download {url} (status: {response.status_code}... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/docugami.html |
e1b0f462dec1-6 | per_file_metadata[doc_id] = metadata
else:
raise Exception(
f"Failed to download {artifact_url}/content "
+ "(status: {response.status_code})"
)
return per_file_metadata
def _load_chunks_for_document(
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/docugami.html |
e1b0f462dec1-7 | for project in _project_details:
metadata = self._metadata_for_project(project)
combined_project_metadata.update(metadata)
for doc in _document_details:
doc_metadata = combined_project_metadata.get(doc["id"])
chunks += self._load_chunks... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/docugami.html |
d902a1612c4a-0 | Source code for langchain.document_loaders.gitbook
"""Loader that loads GitBook."""
from typing import Any, List, Optional
from urllib.parse import urljoin, urlparse
from langchain.docstore.document import Document
from langchain.document_loaders.web_base import WebBaseLoader
[docs]class GitbookLoader(WebBaseLoader):
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/gitbook.html |
d902a1612c4a-1 | [docs] def load(self) -> List[Document]:
"""Fetch text from one single GitBook page."""
if self.load_all_paths:
soup_info = self.scrape()
relative_paths = self._get_paths(soup_info)
documents = []
for path in relative_paths:
url = urljoi... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/gitbook.html |
f923c56c4337-0 | Source code for langchain.document_loaders.psychic
"""Loader that loads documents from Psychic.dev."""
from typing import List, Optional
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
[docs]class PsychicLoader(BaseLoader):
"""Loader that loads documents from ... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/psychic.html |
a422b6e7d949-0 | Source code for langchain.document_loaders.web_base
"""Web base loader class."""
import asyncio
import logging
import warnings
from typing import Any, Dict, Iterator, List, Optional, Union
import aiohttp
import requests
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoa... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/web_base.html |
a422b6e7d949-1 | requests_kwargs: Dict[str, Any] = {}
"""kwargs for requests"""
raise_for_status: bool = False
"""Raise an exception if http status code denotes an error."""
bs_get_text_kwargs: Dict[str, Any] = {}
"""kwargs for beatifulsoup4 get_text"""
def __init__(
self,
web_path: Union[str, Li... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/web_base.html |
a422b6e7d949-2 | if proxies:
self.session.proxies.update(proxies)
@property
def web_path(self) -> str:
if len(self.web_paths) > 1:
raise ValueError("Multiple webpaths found.")
return self.web_paths[0]
async def _fetch(
self, url: str, retries: int = 3, cooldown: int = 2, backo... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/web_base.html |
a422b6e7d949-3 | tasks = []
for url in urls:
task = asyncio.ensure_future(self._fetch_with_rate_limit(url, semaphore))
tasks.append(task)
try:
from tqdm.asyncio import tqdm_asyncio
return await tqdm_asyncio.gather(
*tasks, desc="Fetching pages", ascii=True,... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/web_base.html |
a422b6e7d949-4 | if parser is None:
if url.endswith(".xml"):
parser = "xml"
else:
parser = self.default_parser
self._check_parser(parser)
html_doc = self.session.get(url, verify=self.verify, **self.requests_kwargs)
if self.raise_for_status:
html... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/web_base.html |
aab3917220d1-0 | Source code for langchain.document_loaders.conllu
"""Load CoNLL-U files."""
import csv
from typing import List
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
[docs]class CoNLLULoader(BaseLoader):
"""Load CoNLL-U files."""
def __init__(self, file_path: str... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/conllu.html |
d58ad04281eb-0 | Source code for langchain.document_loaders.slack_directory
"""Loader for documents from a Slack export."""
import json
import zipfile
from pathlib import Path
from typing import Dict, List, Optional
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
[docs]class Slack... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/slack_directory.html |
d58ad04281eb-1 | if not channel_name:
continue
if channel_path.endswith(".json"):
messages = self._read_json(zip_file, channel_path)
for message in messages:
document = self._convert_message_to_document(
messa... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/slack_directory.html |
d58ad04281eb-2 | "timestamp": timestamp,
"user": user,
}
def _get_message_source(self, channel_name: str, user: str, timestamp: str) -> str:
"""
Get the message source as a string.
Args:
channel_name (str): The name of the channel the message belongs to.
user (str)... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/slack_directory.html |
23c0c681609a-0 | Source code for langchain.document_loaders.base
"""Abstract interface for document loader implementations."""
from abc import ABC, abstractmethod
from typing import Iterator, List, Optional
from langchain.document_loaders.blob_loaders import Blob
from langchain.schema import Document
from langchain.text_splitter import... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/base.html |
23c0c681609a-1 | f"{self.__class__.__name__} does not implement lazy_load()"
)
[docs]class BaseBlobParser(ABC):
"""Abstract interface for blob parsers.
A blob parser is provides a way to parse raw data stored in a blob into one
or more documents.
The parser can be composed with blob loaders, making it easy to re... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/base.html |
c49f1d4ca379-0 | Source code for langchain.document_loaders.whatsapp_chat
import re
from pathlib import Path
from typing import List
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
[docs]def concatenate_rows(date: str, sender: str, text: str) -> str:
"""Combine message informa... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/whatsapp_chat.html |
c49f1d4ca379-1 | )
if result:
date, sender, text = result.groups()
if text not in ignore_lines:
text_content += concatenate_rows(date, sender, text)
metadata = {"source": str(p)}
return [Document(page_content=text_content, metadata=metadata)] | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/whatsapp_chat.html |
9438a334eea5-0 | Source code for langchain.document_loaders.html_bs
"""Loader that uses bs4 to load HTML files, enriching metadata with page title."""
import logging
from typing import Dict, List, Union
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
logger = logging.getLogger(__n... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/html_bs.html |
9438a334eea5-1 | title = ""
metadata: Dict[str, Union[str, None]] = {
"source": self.file_path,
"title": title,
}
return [Document(page_content=text, metadata=metadata)] | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/html_bs.html |
159fdfcd17f9-0 | Source code for langchain.document_loaders.word_document
"""Loader that loads word documents."""
import os
import tempfile
from abc import ABC
from typing import List
from urllib.parse import urlparse
import requests
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/word_document.html |
159fdfcd17f9-1 | if hasattr(self, "temp_file"):
self.temp_file.close()
[docs] def load(self) -> List[Document]:
"""Load given path as single page."""
import docx2txt
return [
Document(
page_content=docx2txt.process(self.file_path),
metadata={"source": se... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/word_document.html |
159fdfcd17f9-2 | f"You are on unstructured version {__unstructured_version__}. "
"Partitioning .doc files is only supported in unstructured>=0.4.11. "
"Please upgrade the unstructured package and try again."
)
if is_doc:
from unstructured.partition.doc import partition_doc... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/word_document.html |
6a26d14f29d6-0 | Source code for langchain.document_loaders.recursive_url_loader
from typing import Iterator, List, Optional, Set
from urllib.parse import urlparse
import requests
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
[docs]class RecursiveUrlLoader(BaseLoader):
"""Lo... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/recursive_url_loader.html |
6a26d14f29d6-1 | ):
return visited
# Get all links that are relative to the root of the website
response = requests.get(url)
soup = BeautifulSoup(response.text, "html.parser")
all_links = [link.get("href") for link in soup.find_all("a")]
# Extract only the links that are children of t... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/recursive_url_loader.html |
580d9a90f133-0 | Source code for langchain.document_loaders.onedrive_file
from __future__ import annotations
import tempfile
from typing import TYPE_CHECKING, List
from pydantic import BaseModel, Field
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
from langchain.document_loaders... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/onedrive_file.html |
c74b082b9d77-0 | Source code for langchain.document_loaders.mediawikidump
"""Load Data from a MediaWiki dump xml."""
from typing import List, Optional
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
[docs]class MWDumpLoader(BaseLoader):
"""
Load MediaWiki dump from XML fil... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/mediawikidump.html |
c74b082b9d77-1 | )
metadata = {"source": page.title}
docs.append(Document(page_content=text, metadata=metadata))
return docs | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/mediawikidump.html |
d1546fa0098e-0 | Source code for langchain.document_loaders.helpers
"""Document loader helpers."""
import concurrent.futures
from typing import List, NamedTuple, Optional, cast
[docs]class FileEncoding(NamedTuple):
encoding: Optional[str]
confidence: float
language: Optional[str]
[docs]def detect_file_encodings(file_path: s... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/helpers.html |
c147353bc066-0 | Source code for langchain.document_loaders.bilibili
import json
import re
import warnings
from typing import List, Tuple
import requests
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
[docs]class BiliBiliLoader(BaseLoader):
"""Loader that loads bilibili trans... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/bilibili.html |
c147353bc066-1 | video_info = sync(v.get_info())
video_info.update({"url": url})
# Get subtitle url
subtitle = video_info.pop("subtitle")
sub_list = subtitle["list"]
if sub_list:
sub_url = sub_list[0]["subtitle_url"]
result = requests.get(sub_url)
raw_sub_title... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/bilibili.html |
70c2fe86e04e-0 | Source code for langchain.document_loaders.xml
"""Loader that loads Microsoft Excel files."""
from typing import Any, List
from langchain.document_loaders.unstructured import (
UnstructuredFileLoader,
validate_unstructured_version,
)
[docs]class UnstructuredXMLLoader(UnstructuredFileLoader):
"""Loader that ... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/xml.html |
7e4dd43c4336-0 | Source code for langchain.document_loaders.powerpoint
"""Loader that loads powerpoint files."""
import os
from typing import List
from langchain.document_loaders.unstructured import UnstructuredFileLoader
[docs]class UnstructuredPowerPointLoader(UnstructuredFileLoader):
"""Loader that uses unstructured to load powe... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/powerpoint.html |
567e70744c87-0 | Source code for langchain.document_loaders.chatgpt
"""Load conversations from ChatGPT data export"""
import datetime
import json
from typing import List
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
[docs]def concatenate_rows(message: dict, title: str) -> str:
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/chatgpt.html |
567e70744c87-1 | if not (
idx == 0
and messages[key]["message"]["author"]["role"] == "system"
)
]
)
metadata = {"source": str(self.log_file)}
documents.append(Document(page_content=text, metadata=metadata))
re... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/chatgpt.html |
3458a7a9d094-0 | Source code for langchain.document_loaders.odt
"""Loader that loads Open Office ODT files."""
from typing import Any, List
from langchain.document_loaders.unstructured import (
UnstructuredFileLoader,
validate_unstructured_version,
)
[docs]class UnstructuredODTLoader(UnstructuredFileLoader):
"""Loader that ... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/odt.html |
9bc93c2fa5f7-0 | Source code for langchain.document_loaders.python
import tokenize
from langchain.document_loaders.text import TextLoader
[docs]class PythonLoader(TextLoader):
"""
Load Python files, respecting any non-default encoding if specified.
"""
def __init__(self, file_path: str):
with open(file_path, "rb... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/python.html |
6acf9601d050-0 | Source code for langchain.document_loaders.telegram
"""Loader that loads Telegram chat json dump."""
from __future__ import annotations
import asyncio
import json
from pathlib import Path
from typing import TYPE_CHECKING, Dict, List, Optional, Union
from langchain.docstore.document import Document
from langchain.docume... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/telegram.html |
6acf9601d050-1 | if isinstance(text, str):
# Take a single string as one page
text = [text]
page_docs = [Document(page_content=page) for page in text]
# Add page numbers as metadata
for i, doc in enumerate(page_docs):
doc.metadata["page"] = i + 1
# Split pages into chunks
doc_chunks = []
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/telegram.html |
6acf9601d050-2 | [docs] async def fetch_data_from_telegram(self) -> None:
"""Fetch data from Telegram API and save it as a JSON file."""
from telethon.sync import TelegramClient
data = []
async with TelegramClient(self.username, self.api_id, self.api_hash) as client:
async for message in c... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/telegram.html |
6acf9601d050-3 | Args:
parent_id (int): The parent message ID.
reply_data (pd.DataFrame): A DataFrame containing reply messages.
Returns:
list: A list of message IDs that are replies to the parent message ID.
"""
# Find direct replies to the parent mess... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/telegram.html |
6acf9601d050-4 | message_threads (dict): A dictionary where the key is the parent message \
ID and the value is a list of message IDs in ascending order.
data (pd.DataFrame): A DataFrame containing the conversation data:
- message.sender_id
- text
- date
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/telegram.html |
6acf9601d050-5 | please install with `pip install pandas`
"""
)
normalized_messages = pd.json_normalize(d)
df = pd.DataFrame(normalized_messages)
message_threads = self._get_message_threads(df)
combined_texts = self._combine_message_texts(message_threads, df)
return te... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/telegram.html |
53e27423a128-0 | Source code for langchain.document_loaders.url_selenium
"""Loader that uses Selenium to load a page, then uses unstructured to load the html.
"""
import logging
from typing import TYPE_CHECKING, List, Literal, Optional, Union
if TYPE_CHECKING:
from selenium.webdriver import Chrome, Firefox
from langchain.docstore.d... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/url_selenium.html |
53e27423a128-1 | raise ImportError(
"selenium package not found, please install it with "
"`pip install selenium`"
)
try:
import unstructured # noqa:F401
except ImportError:
raise ImportError(
"unstructured package not found, please ins... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/url_selenium.html |
53e27423a128-2 | for arg in self.arguments:
firefox_options.add_argument(arg)
if self.headless:
firefox_options.add_argument("--headless")
if self.binary_location is not None:
firefox_options.binary_location = self.binary_location
if self.executable_pat... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/url_selenium.html |
297e03b416f4-0 | Source code for langchain.document_loaders.s3_file
"""Loading logic for loading documents from an s3 file."""
import os
import tempfile
from typing import List
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
from langchain.document_loaders.unstructured import Unst... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/s3_file.html |
c9fbfedaeec2-0 | Source code for langchain.document_loaders.gcs_file
"""Loading logic for loading documents from a GCS file."""
import os
import tempfile
from typing import List
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
from langchain.document_loaders.unstructured import Uns... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/gcs_file.html |
177b7fc24be5-0 | Source code for langchain.document_loaders.tomarkdown
"""Loader that loads HTML to markdown using 2markdown."""
from __future__ import annotations
from typing import Iterator, List
import requests
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
[docs]class ToMarkd... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/tomarkdown.html |
2f16b59bc02c-0 | Source code for langchain.document_loaders.pyspark_dataframe
"""Load from a Spark Dataframe object"""
import itertools
import logging
import sys
from typing import TYPE_CHECKING, Any, Iterator, List, Optional, Tuple
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/pyspark_dataframe.html |
2f16b59bc02c-1 | """Gets the amount of "feasible" rows for the DataFrame"""
try:
import psutil
except ImportError as e:
raise ImportError(
"psutil not installed. Please install it with `pip install psutil`."
) from e
row = self.df.limit(1).collect()[0]
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/pyspark_dataframe.html |
8c428a873dcd-0 | Source code for langchain.document_loaders.evernote
"""Load documents from Evernote.
https://gist.github.com/foxmask/7b29c43a161e001ff04afdb2f181e31c
"""
import hashlib
import logging
from base64 import b64decode
from time import strptime
from typing import Any, Dict, Iterator, List, Optional
from langchain.docstore.do... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/evernote.html |
8c428a873dcd-1 | self.file_path = file_path
self.load_single_document = load_single_document
[docs] def load(self) -> List[Document]:
"""Load documents from EverNote export file."""
documents = [
Document(
page_content=note["content"],
metadata={
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/evernote.html |
8c428a873dcd-2 | rsc_dict["hash"] = hashlib.md5(rsc_dict[elem.tag]).hexdigest()
else:
rsc_dict[elem.tag] = elem.text
return rsc_dict
@staticmethod
def _parse_note(note: List, prefix: Optional[str] = None) -> dict:
note_dict: Dict[str, Any] = {}
resources = []
def add_p... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/evernote.html |
8c428a873dcd-3 | # Without huge_tree set to True, parser may complain about huge text node
# Try to recover, because there may be " ", which will cause
# "XMLSyntaxError: Entity 'nbsp' not defined"
try:
from lxml import etree
except ImportError as e:
logging.error(
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/evernote.html |
1fe2193066c3-0 | Source code for langchain.document_loaders.tencent_cos_file
"""Loading logic for loading documents from Tencent Cloud COS file."""
import os
import tempfile
from typing import Any, Iterator, List
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
from langchain.docum... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/tencent_cos_file.html |
1fe2193066c3-1 | )
loader = UnstructuredFileLoader(file_path)
# UnstructuredFileLoader not implement lazy_load yet
return iter(loader.load()) | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/tencent_cos_file.html |
3bbcc0a92cff-0 | Source code for langchain.document_loaders.parsers.grobid
from typing import Dict, Iterator, List, Union
import requests
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseBlobParser
from langchain.document_loaders.blob_loaders import Blob
[docs]class ServerUnavailableExcep... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/grobid.html |
3bbcc0a92cff-1 | for i, paragraph in enumerate(section.find_all("p")):
chunk_bboxes = []
paragraph_text = []
for i, sentence in enumerate(paragraph.find_all("s")):
paragraph_text.append(sentence.text)
sbboxes = []
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/grobid.html |
3bbcc0a92cff-2 | Document(
page_content=chunk["text"],
metadata=dict(
{
"text": str(chunk["text"]),
"para": str(chunk["para"]),
"bboxes": str(chunk["bboxes"]),
"pages": str(chunk["pages"]),... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/grobid.html |
71f60526a274-0 | Source code for langchain.document_loaders.parsers.audio
from typing import Iterator
from langchain.document_loaders.base import BaseBlobParser
from langchain.document_loaders.blob_loaders import Blob
from langchain.schema import Document
[docs]class OpenAIWhisperParser(BaseBlobParser):
"""Transcribe and parse audi... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/audio.html |
71f60526a274-1 | # Transcribe
print(f"Transcribing part {split_number+1}!")
transcript = openai.Audio.transcribe("whisper-1", file_obj)
yield Document(
page_content=transcript.text,
metadata={"source": blob.source, "chunk": split_number},
) | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/audio.html |
b155063f625d-0 | Source code for langchain.document_loaders.parsers.pdf
"""Module contains common parsers for PDFs."""
from typing import Any, Iterator, Mapping, Optional, Union
from langchain.document_loaders.base import BaseBlobParser
from langchain.document_loaders.blob_loaders import Blob
from langchain.schema import Document
[docs... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/pdf.html |
b155063f625d-1 | """Initialize the parser.
Args:
text_kwargs: Keyword arguments to pass to ``fitz.Page.get_text()``.
"""
self.text_kwargs = text_kwargs or {}
[docs] def lazy_parse(self, blob: Blob) -> Iterator[Document]:
"""Lazily parse the blob."""
import fitz
with blob.as... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/pdf.html |
b155063f625d-2 | # if done incorrectly creates seg faults.
with blob.as_bytes_io() as file_path:
pdf_reader = pypdfium2.PdfDocument(file_path, autoclose=True)
try:
for page_number, page in enumerate(pdf_reader):
text_page = page.get_textpage()
conte... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/pdf.html |
b155063f625d-3 | },
),
)
for page in doc.pages
] | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/pdf.html |
4cd92ac7dba1-0 | Source code for langchain.document_loaders.parsers.generic
"""Code for generic / auxiliary parsers.
This module contains some logic to help assemble more sophisticated parsers.
"""
from typing import Iterator, Mapping, Optional
from langchain.document_loaders.base import BaseBlobParser
from langchain.document_loaders.b... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/generic.html |
4cd92ac7dba1-1 | """
self.handlers = handlers
self.fallback_parser = fallback_parser
[docs] def lazy_parse(self, blob: Blob) -> Iterator[Document]:
"""Load documents from a blob."""
mimetype = blob.mimetype
if mimetype is None:
raise ValueError(f"{blob} does not have a mimetype.")
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/generic.html |
7e7265154bc9-0 | Source code for langchain.document_loaders.parsers.txt
"""Module for parsing text files.."""
from typing import Iterator
from langchain.document_loaders.base import BaseBlobParser
from langchain.document_loaders.blob_loaders import Blob
from langchain.schema import Document
[docs]class TextParser(BaseBlobParser):
"... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/txt.html |
a5d0a177b784-0 | Source code for langchain.document_loaders.parsers.registry
"""Module includes a registry of default parser configurations."""
from langchain.document_loaders.base import BaseBlobParser
from langchain.document_loaders.parsers.generic import MimeTypeBasedParser
from langchain.document_loaders.parsers.pdf import PyMuPDFP... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/registry.html |
c6126d0c8db0-0 | Source code for langchain.document_loaders.parsers.html.bs4
"""Loader that uses bs4 to load HTML files, enriching metadata with page title."""
import logging
from typing import Any, Dict, Iterator, Union
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseBlobParser
from lan... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/html/bs4.html |
62ed8298bced-0 | Source code for langchain.document_loaders.parsers.language.code_segmenter
from abc import ABC, abstractmethod
from typing import List
[docs]class CodeSegmenter(ABC):
def __init__(self, code: str):
self.code = code
[docs] def is_valid(self) -> bool:
return True
[docs] @abstractmethod
def s... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/language/code_segmenter.html |
bec8ced10a2f-0 | Source code for langchain.document_loaders.parsers.language.javascript
from typing import Any, List
from langchain.document_loaders.parsers.language.code_segmenter import CodeSegmenter
[docs]class JavaScriptSegmenter(CodeSegmenter):
def __init__(self, code: str):
super().__init__(code)
self.source_l... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/language/javascript.html |
bec8ced10a2f-1 | for node in tree.body:
if isinstance(
node,
(esprima.nodes.FunctionDeclaration, esprima.nodes.ClassDeclaration),
):
start = node.loc.start.line - 1
simplified_lines[start] = f"// Code for: {simplified_lines[start]}"
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/language/javascript.html |
1a219eef5d7d-0 | Source code for langchain.document_loaders.parsers.language.language_parser
from typing import Any, Dict, Iterator, Optional
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseBlobParser
from langchain.document_loaders.blob_loaders import Blob
from langchain.document_loader... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/language/language_parser.html |
1a219eef5d7d-1 | docs = loader.load()
Example instantiations to manually select the language:
... code-block:: python
from langchain.text_splitter import Language
loader = GenericLoader.from_filesystem(
"./code",
glob="**/*",
suffixes=[".py"],
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/language/language_parser.html |
1a219eef5d7d-2 | "language": language,
},
)
return
self.Segmenter = LANGUAGE_SEGMENTERS[language]
segmenter = self.Segmenter(blob.as_string())
if not segmenter.is_valid():
yield Document(
page_content=code,
metadata={
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/language/language_parser.html |
fc6956191354-0 | Source code for langchain.document_loaders.parsers.language.python
import ast
from typing import Any, List
from langchain.document_loaders.parsers.language.code_segmenter import CodeSegmenter
[docs]class PythonSegmenter(CodeSegmenter):
def __init__(self, code: str):
super().__init__(code)
self.sourc... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/language/python.html |
d02f9d1cad25-0 | Source code for langchain.document_loaders.blob_loaders.file_system
"""Use to load blobs from the local file system."""
from pathlib import Path
from typing import Callable, Iterable, Iterator, Optional, Sequence, TypeVar, Union
from langchain.document_loaders.blob_loaders.schema import Blob, BlobLoader
T = TypeVar("T"... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/blob_loaders/file_system.html |
d02f9d1cad25-1 | *,
glob: str = "**/[!.]*",
suffixes: Optional[Sequence[str]] = None,
show_progress: bool = False,
) -> None:
"""Initialize with path to directory and how to glob over it.
Args:
path: Path to directory to load from
glob: Glob pattern relative to the spe... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/blob_loaders/file_system.html |
d02f9d1cad25-2 | self,
) -> Iterable[Blob]:
"""Yield blobs that match the requested pattern."""
iterator = _make_iterator(
length_func=self.count_matching_files, show_progress=self.show_progress
)
for path in iterator(self._yield_paths()):
yield Blob.from_path(path)
def _y... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/blob_loaders/file_system.html |
d14be4bb58da-0 | Source code for langchain.document_loaders.blob_loaders.youtube_audio
from typing import Iterable, List
from langchain.document_loaders.blob_loaders import FileSystemBlobLoader
from langchain.document_loaders.blob_loaders.schema import Blob, BlobLoader
[docs]class YoutubeAudioLoader(BlobLoader):
"""Load YouTube url... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/blob_loaders/youtube_audio.html |
52bea49e6b46-0 | Source code for langchain.document_loaders.blob_loaders.schema
"""Schema for Blobs and Blob Loaders.
The goal is to facilitate decoupling of content loading from content parsing code.
In addition, content loading code should provide a lazy loading interface by default.
"""
from __future__ import annotations
import cont... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/blob_loaders/schema.html |
52bea49e6b46-1 | return str(self.path) if self.path else None
[docs] @root_validator(pre=True)
def check_blob_is_valid(cls, values: Mapping[str, Any]) -> Mapping[str, Any]:
"""Verify that either data or path is provided."""
if "data" not in values and "path" not in values:
raise ValueError("Either dat... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/blob_loaders/schema.html |
52bea49e6b46-2 | yield f
else:
raise NotImplementedError(f"Unable to convert blob {self}")
[docs] @classmethod
def from_path(
cls,
path: PathLike,
*,
encoding: str = "utf-8",
mime_type: Optional[str] = None,
guess_type: bool = True,
) -> Blob:
"""Loa... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/blob_loaders/schema.html |
52bea49e6b46-3 | mime_type: if provided, will be set as the mime-type of the data
path: if provided, will be set as the source from which the data came
Returns:
Blob instance
"""
return cls(data=data, mimetype=mime_type, encoding=encoding, path=path)
def __repr__(self) -> str:
... | https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/blob_loaders/schema.html |
c83fb8943a93-0 | Source code for langchain.prompts.few_shot_with_templates
"""Prompt template that contains few shot examples."""
from typing import Any, Dict, List, Optional
from pydantic import Extra, root_validator
from langchain.prompts.base import DEFAULT_FORMATTER_MAPPING, StringPromptTemplate
from langchain.prompts.example_selec... | https://api.python.langchain.com/en/latest/_modules/langchain/prompts/few_shot_with_templates.html |
c83fb8943a93-1 | examples = values.get("examples", None)
example_selector = values.get("example_selector", None)
if examples and example_selector:
raise ValueError(
"Only one of 'examples' and 'example_selector' should be provided"
)
if examples is None and example_selecto... | https://api.python.langchain.com/en/latest/_modules/langchain/prompts/few_shot_with_templates.html |
c83fb8943a93-2 | """Format the prompt with the inputs.
Args:
kwargs: Any arguments to be passed to the prompt template.
Returns:
A formatted string.
Example:
.. code-block:: python
prompt.format(variable1="foo")
"""
kwargs = self._merge_partial_and_user... | https://api.python.langchain.com/en/latest/_modules/langchain/prompts/few_shot_with_templates.html |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.