id
stringlengths
14
15
text
stringlengths
35
2.51k
source
stringlengths
61
154
159ac61bed59-3
data = res.json() text = "\n".join( [ data[key] for key in ["title", "description", "contents_raw"] if key in data ] ).strip() metadata = {"source": self.web_path, "title": data["title"]} documents.append(Document(pa...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/ifixit.html
159ac61bed59-4
else: for part in data["parts"]: doc_parts.append("\n - " + part["text"]) for row in data["steps"]: doc_parts.append( "\n\n## " + ( row["title"] if row["title"] != "" else "Step {}...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/ifixit.html
c31fe87c2f6b-0
Source code for langchain.document_loaders.url """Loader that uses unstructured to load HTML files.""" import logging from typing import Any, List from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader logger = logging.getLogger(__name__) [docs]class UnstructuredURLLoade...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/url.html
c31fe87c2f6b-1
self.unstructured_kwargs = unstructured_kwargs self.show_progress_bar = show_progress_bar def _validate_mode(self, mode: str) -> None: _valid_modes = {"single", "elements"} if mode not in _valid_modes: raise ValueError( f"Got {mode} for `mode`, but should be one o...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/url.html
c31fe87c2f6b-2
"Please install with 'pip install tqdm' or set " "show_progress_bar=False." ) from e urls = tqdm(self.urls) else: urls = self.urls for url in urls: try: if self.__is_non_html_available(): if self....
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/url.html
e1b0f462dec1-0
Source code for langchain.document_loaders.docugami """Loader that loads processed documents from Docugami.""" import io import logging import os import re from pathlib import Path from typing import Any, Dict, List, Mapping, Optional, Sequence, Union import requests from pydantic import BaseModel, root_validator from ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/docugami.html
e1b0f462dec1-1
if values.get("file_paths") and values.get("docset_id"): raise ValueError("Cannot specify both file_paths and remote API docset_id") if not values.get("file_paths") and not values.get("docset_id"): raise ValueError("Must specify either file_paths or remote API docset_id") if valu...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/docugami.html
e1b0f462dec1-2
ancestor_chain = chunk.xpath("ancestor-or-self::*") return "/" + "/".join(_xpath_qname_for_chunk(x) for x in ancestor_chain) def _structure_value(node: Any) -> str: """Get the structure value for a node.""" structure = ( "table" if node.tag == ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/docugami.html
e1b0f462dec1-3
"""Create a Document from a node and text.""" metadata = { XPATH_KEY: _xpath_for_chunk(node), DOCUMENT_ID_KEY: document["id"], DOCUMENT_NAME_KEY: document["name"], STRUCTURE_KEY: node.attrib.get("structure", ""), TAG_KEY: re.sub...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/docugami.html
e1b0f462dec1-4
while url: response = requests.get( url, headers={"Authorization": f"Bearer {self.access_token}"}, ) if response.ok: data = response.json() all_documents.extend(data["documents"]) url = data.get("next", N...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/docugami.html
e1b0f462dec1-5
data={}, ) if response.ok: data = response.json() all_artifacts.extend(data["artifacts"]) url = data.get("next", None) else: raise Exception( f"Failed to download {url} (status: {response.status_code}...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/docugami.html
e1b0f462dec1-6
per_file_metadata[doc_id] = metadata else: raise Exception( f"Failed to download {artifact_url}/content " + "(status: {response.status_code})" ) return per_file_metadata def _load_chunks_for_document( ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/docugami.html
e1b0f462dec1-7
for project in _project_details: metadata = self._metadata_for_project(project) combined_project_metadata.update(metadata) for doc in _document_details: doc_metadata = combined_project_metadata.get(doc["id"]) chunks += self._load_chunks...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/docugami.html
d902a1612c4a-0
Source code for langchain.document_loaders.gitbook """Loader that loads GitBook.""" from typing import Any, List, Optional from urllib.parse import urljoin, urlparse from langchain.docstore.document import Document from langchain.document_loaders.web_base import WebBaseLoader [docs]class GitbookLoader(WebBaseLoader): ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/gitbook.html
d902a1612c4a-1
[docs] def load(self) -> List[Document]: """Fetch text from one single GitBook page.""" if self.load_all_paths: soup_info = self.scrape() relative_paths = self._get_paths(soup_info) documents = [] for path in relative_paths: url = urljoi...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/gitbook.html
f923c56c4337-0
Source code for langchain.document_loaders.psychic """Loader that loads documents from Psychic.dev.""" from typing import List, Optional from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader [docs]class PsychicLoader(BaseLoader): """Loader that loads documents from ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/psychic.html
a422b6e7d949-0
Source code for langchain.document_loaders.web_base """Web base loader class.""" import asyncio import logging import warnings from typing import Any, Dict, Iterator, List, Optional, Union import aiohttp import requests from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoa...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/web_base.html
a422b6e7d949-1
requests_kwargs: Dict[str, Any] = {} """kwargs for requests""" raise_for_status: bool = False """Raise an exception if http status code denotes an error.""" bs_get_text_kwargs: Dict[str, Any] = {} """kwargs for beatifulsoup4 get_text""" def __init__( self, web_path: Union[str, Li...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/web_base.html
a422b6e7d949-2
if proxies: self.session.proxies.update(proxies) @property def web_path(self) -> str: if len(self.web_paths) > 1: raise ValueError("Multiple webpaths found.") return self.web_paths[0] async def _fetch( self, url: str, retries: int = 3, cooldown: int = 2, backo...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/web_base.html
a422b6e7d949-3
tasks = [] for url in urls: task = asyncio.ensure_future(self._fetch_with_rate_limit(url, semaphore)) tasks.append(task) try: from tqdm.asyncio import tqdm_asyncio return await tqdm_asyncio.gather( *tasks, desc="Fetching pages", ascii=True,...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/web_base.html
a422b6e7d949-4
if parser is None: if url.endswith(".xml"): parser = "xml" else: parser = self.default_parser self._check_parser(parser) html_doc = self.session.get(url, verify=self.verify, **self.requests_kwargs) if self.raise_for_status: html...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/web_base.html
aab3917220d1-0
Source code for langchain.document_loaders.conllu """Load CoNLL-U files.""" import csv from typing import List from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader [docs]class CoNLLULoader(BaseLoader): """Load CoNLL-U files.""" def __init__(self, file_path: str...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/conllu.html
d58ad04281eb-0
Source code for langchain.document_loaders.slack_directory """Loader for documents from a Slack export.""" import json import zipfile from pathlib import Path from typing import Dict, List, Optional from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader [docs]class Slack...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/slack_directory.html
d58ad04281eb-1
if not channel_name: continue if channel_path.endswith(".json"): messages = self._read_json(zip_file, channel_path) for message in messages: document = self._convert_message_to_document( messa...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/slack_directory.html
d58ad04281eb-2
"timestamp": timestamp, "user": user, } def _get_message_source(self, channel_name: str, user: str, timestamp: str) -> str: """ Get the message source as a string. Args: channel_name (str): The name of the channel the message belongs to. user (str)...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/slack_directory.html
23c0c681609a-0
Source code for langchain.document_loaders.base """Abstract interface for document loader implementations.""" from abc import ABC, abstractmethod from typing import Iterator, List, Optional from langchain.document_loaders.blob_loaders import Blob from langchain.schema import Document from langchain.text_splitter import...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/base.html
23c0c681609a-1
f"{self.__class__.__name__} does not implement lazy_load()" ) [docs]class BaseBlobParser(ABC): """Abstract interface for blob parsers. A blob parser is provides a way to parse raw data stored in a blob into one or more documents. The parser can be composed with blob loaders, making it easy to re...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/base.html
c49f1d4ca379-0
Source code for langchain.document_loaders.whatsapp_chat import re from pathlib import Path from typing import List from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader [docs]def concatenate_rows(date: str, sender: str, text: str) -> str: """Combine message informa...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/whatsapp_chat.html
c49f1d4ca379-1
) if result: date, sender, text = result.groups() if text not in ignore_lines: text_content += concatenate_rows(date, sender, text) metadata = {"source": str(p)} return [Document(page_content=text_content, metadata=metadata)]
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/whatsapp_chat.html
9438a334eea5-0
Source code for langchain.document_loaders.html_bs """Loader that uses bs4 to load HTML files, enriching metadata with page title.""" import logging from typing import Dict, List, Union from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader logger = logging.getLogger(__n...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/html_bs.html
9438a334eea5-1
title = "" metadata: Dict[str, Union[str, None]] = { "source": self.file_path, "title": title, } return [Document(page_content=text, metadata=metadata)]
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/html_bs.html
159fdfcd17f9-0
Source code for langchain.document_loaders.word_document """Loader that loads word documents.""" import os import tempfile from abc import ABC from typing import List from urllib.parse import urlparse import requests from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/word_document.html
159fdfcd17f9-1
if hasattr(self, "temp_file"): self.temp_file.close() [docs] def load(self) -> List[Document]: """Load given path as single page.""" import docx2txt return [ Document( page_content=docx2txt.process(self.file_path), metadata={"source": se...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/word_document.html
159fdfcd17f9-2
f"You are on unstructured version {__unstructured_version__}. " "Partitioning .doc files is only supported in unstructured>=0.4.11. " "Please upgrade the unstructured package and try again." ) if is_doc: from unstructured.partition.doc import partition_doc...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/word_document.html
6a26d14f29d6-0
Source code for langchain.document_loaders.recursive_url_loader from typing import Iterator, List, Optional, Set from urllib.parse import urlparse import requests from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader [docs]class RecursiveUrlLoader(BaseLoader): """Lo...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/recursive_url_loader.html
6a26d14f29d6-1
): return visited # Get all links that are relative to the root of the website response = requests.get(url) soup = BeautifulSoup(response.text, "html.parser") all_links = [link.get("href") for link in soup.find_all("a")] # Extract only the links that are children of t...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/recursive_url_loader.html
580d9a90f133-0
Source code for langchain.document_loaders.onedrive_file from __future__ import annotations import tempfile from typing import TYPE_CHECKING, List from pydantic import BaseModel, Field from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader from langchain.document_loaders...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/onedrive_file.html
c74b082b9d77-0
Source code for langchain.document_loaders.mediawikidump """Load Data from a MediaWiki dump xml.""" from typing import List, Optional from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader [docs]class MWDumpLoader(BaseLoader): """ Load MediaWiki dump from XML fil...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/mediawikidump.html
c74b082b9d77-1
) metadata = {"source": page.title} docs.append(Document(page_content=text, metadata=metadata)) return docs
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/mediawikidump.html
d1546fa0098e-0
Source code for langchain.document_loaders.helpers """Document loader helpers.""" import concurrent.futures from typing import List, NamedTuple, Optional, cast [docs]class FileEncoding(NamedTuple): encoding: Optional[str] confidence: float language: Optional[str] [docs]def detect_file_encodings(file_path: s...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/helpers.html
c147353bc066-0
Source code for langchain.document_loaders.bilibili import json import re import warnings from typing import List, Tuple import requests from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader [docs]class BiliBiliLoader(BaseLoader): """Loader that loads bilibili trans...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/bilibili.html
c147353bc066-1
video_info = sync(v.get_info()) video_info.update({"url": url}) # Get subtitle url subtitle = video_info.pop("subtitle") sub_list = subtitle["list"] if sub_list: sub_url = sub_list[0]["subtitle_url"] result = requests.get(sub_url) raw_sub_title...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/bilibili.html
70c2fe86e04e-0
Source code for langchain.document_loaders.xml """Loader that loads Microsoft Excel files.""" from typing import Any, List from langchain.document_loaders.unstructured import ( UnstructuredFileLoader, validate_unstructured_version, ) [docs]class UnstructuredXMLLoader(UnstructuredFileLoader): """Loader that ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/xml.html
7e4dd43c4336-0
Source code for langchain.document_loaders.powerpoint """Loader that loads powerpoint files.""" import os from typing import List from langchain.document_loaders.unstructured import UnstructuredFileLoader [docs]class UnstructuredPowerPointLoader(UnstructuredFileLoader): """Loader that uses unstructured to load powe...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/powerpoint.html
567e70744c87-0
Source code for langchain.document_loaders.chatgpt """Load conversations from ChatGPT data export""" import datetime import json from typing import List from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader [docs]def concatenate_rows(message: dict, title: str) -> str: ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/chatgpt.html
567e70744c87-1
if not ( idx == 0 and messages[key]["message"]["author"]["role"] == "system" ) ] ) metadata = {"source": str(self.log_file)} documents.append(Document(page_content=text, metadata=metadata)) re...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/chatgpt.html
3458a7a9d094-0
Source code for langchain.document_loaders.odt """Loader that loads Open Office ODT files.""" from typing import Any, List from langchain.document_loaders.unstructured import ( UnstructuredFileLoader, validate_unstructured_version, ) [docs]class UnstructuredODTLoader(UnstructuredFileLoader): """Loader that ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/odt.html
9bc93c2fa5f7-0
Source code for langchain.document_loaders.python import tokenize from langchain.document_loaders.text import TextLoader [docs]class PythonLoader(TextLoader): """ Load Python files, respecting any non-default encoding if specified. """ def __init__(self, file_path: str): with open(file_path, "rb...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/python.html
6acf9601d050-0
Source code for langchain.document_loaders.telegram """Loader that loads Telegram chat json dump.""" from __future__ import annotations import asyncio import json from pathlib import Path from typing import TYPE_CHECKING, Dict, List, Optional, Union from langchain.docstore.document import Document from langchain.docume...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/telegram.html
6acf9601d050-1
if isinstance(text, str): # Take a single string as one page text = [text] page_docs = [Document(page_content=page) for page in text] # Add page numbers as metadata for i, doc in enumerate(page_docs): doc.metadata["page"] = i + 1 # Split pages into chunks doc_chunks = [] ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/telegram.html
6acf9601d050-2
[docs] async def fetch_data_from_telegram(self) -> None: """Fetch data from Telegram API and save it as a JSON file.""" from telethon.sync import TelegramClient data = [] async with TelegramClient(self.username, self.api_id, self.api_hash) as client: async for message in c...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/telegram.html
6acf9601d050-3
Args: parent_id (int): The parent message ID. reply_data (pd.DataFrame): A DataFrame containing reply messages. Returns: list: A list of message IDs that are replies to the parent message ID. """ # Find direct replies to the parent mess...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/telegram.html
6acf9601d050-4
message_threads (dict): A dictionary where the key is the parent message \ ID and the value is a list of message IDs in ascending order. data (pd.DataFrame): A DataFrame containing the conversation data: - message.sender_id - text - date ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/telegram.html
6acf9601d050-5
please install with `pip install pandas` """ ) normalized_messages = pd.json_normalize(d) df = pd.DataFrame(normalized_messages) message_threads = self._get_message_threads(df) combined_texts = self._combine_message_texts(message_threads, df) return te...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/telegram.html
53e27423a128-0
Source code for langchain.document_loaders.url_selenium """Loader that uses Selenium to load a page, then uses unstructured to load the html. """ import logging from typing import TYPE_CHECKING, List, Literal, Optional, Union if TYPE_CHECKING: from selenium.webdriver import Chrome, Firefox from langchain.docstore.d...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/url_selenium.html
53e27423a128-1
raise ImportError( "selenium package not found, please install it with " "`pip install selenium`" ) try: import unstructured # noqa:F401 except ImportError: raise ImportError( "unstructured package not found, please ins...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/url_selenium.html
53e27423a128-2
for arg in self.arguments: firefox_options.add_argument(arg) if self.headless: firefox_options.add_argument("--headless") if self.binary_location is not None: firefox_options.binary_location = self.binary_location if self.executable_pat...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/url_selenium.html
297e03b416f4-0
Source code for langchain.document_loaders.s3_file """Loading logic for loading documents from an s3 file.""" import os import tempfile from typing import List from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader from langchain.document_loaders.unstructured import Unst...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/s3_file.html
c9fbfedaeec2-0
Source code for langchain.document_loaders.gcs_file """Loading logic for loading documents from a GCS file.""" import os import tempfile from typing import List from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader from langchain.document_loaders.unstructured import Uns...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/gcs_file.html
177b7fc24be5-0
Source code for langchain.document_loaders.tomarkdown """Loader that loads HTML to markdown using 2markdown.""" from __future__ import annotations from typing import Iterator, List import requests from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader [docs]class ToMarkd...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/tomarkdown.html
2f16b59bc02c-0
Source code for langchain.document_loaders.pyspark_dataframe """Load from a Spark Dataframe object""" import itertools import logging import sys from typing import TYPE_CHECKING, Any, Iterator, List, Optional, Tuple from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/pyspark_dataframe.html
2f16b59bc02c-1
"""Gets the amount of "feasible" rows for the DataFrame""" try: import psutil except ImportError as e: raise ImportError( "psutil not installed. Please install it with `pip install psutil`." ) from e row = self.df.limit(1).collect()[0] ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/pyspark_dataframe.html
8c428a873dcd-0
Source code for langchain.document_loaders.evernote """Load documents from Evernote. https://gist.github.com/foxmask/7b29c43a161e001ff04afdb2f181e31c """ import hashlib import logging from base64 import b64decode from time import strptime from typing import Any, Dict, Iterator, List, Optional from langchain.docstore.do...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/evernote.html
8c428a873dcd-1
self.file_path = file_path self.load_single_document = load_single_document [docs] def load(self) -> List[Document]: """Load documents from EverNote export file.""" documents = [ Document( page_content=note["content"], metadata={ ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/evernote.html
8c428a873dcd-2
rsc_dict["hash"] = hashlib.md5(rsc_dict[elem.tag]).hexdigest() else: rsc_dict[elem.tag] = elem.text return rsc_dict @staticmethod def _parse_note(note: List, prefix: Optional[str] = None) -> dict: note_dict: Dict[str, Any] = {} resources = [] def add_p...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/evernote.html
8c428a873dcd-3
# Without huge_tree set to True, parser may complain about huge text node # Try to recover, because there may be " ", which will cause # "XMLSyntaxError: Entity 'nbsp' not defined" try: from lxml import etree except ImportError as e: logging.error( ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/evernote.html
1fe2193066c3-0
Source code for langchain.document_loaders.tencent_cos_file """Loading logic for loading documents from Tencent Cloud COS file.""" import os import tempfile from typing import Any, Iterator, List from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader from langchain.docum...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/tencent_cos_file.html
1fe2193066c3-1
) loader = UnstructuredFileLoader(file_path) # UnstructuredFileLoader not implement lazy_load yet return iter(loader.load())
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/tencent_cos_file.html
3bbcc0a92cff-0
Source code for langchain.document_loaders.parsers.grobid from typing import Dict, Iterator, List, Union import requests from langchain.docstore.document import Document from langchain.document_loaders.base import BaseBlobParser from langchain.document_loaders.blob_loaders import Blob [docs]class ServerUnavailableExcep...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/grobid.html
3bbcc0a92cff-1
for i, paragraph in enumerate(section.find_all("p")): chunk_bboxes = [] paragraph_text = [] for i, sentence in enumerate(paragraph.find_all("s")): paragraph_text.append(sentence.text) sbboxes = [] ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/grobid.html
3bbcc0a92cff-2
Document( page_content=chunk["text"], metadata=dict( { "text": str(chunk["text"]), "para": str(chunk["para"]), "bboxes": str(chunk["bboxes"]), "pages": str(chunk["pages"]),...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/grobid.html
71f60526a274-0
Source code for langchain.document_loaders.parsers.audio from typing import Iterator from langchain.document_loaders.base import BaseBlobParser from langchain.document_loaders.blob_loaders import Blob from langchain.schema import Document [docs]class OpenAIWhisperParser(BaseBlobParser): """Transcribe and parse audi...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/audio.html
71f60526a274-1
# Transcribe print(f"Transcribing part {split_number+1}!") transcript = openai.Audio.transcribe("whisper-1", file_obj) yield Document( page_content=transcript.text, metadata={"source": blob.source, "chunk": split_number}, )
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/audio.html
b155063f625d-0
Source code for langchain.document_loaders.parsers.pdf """Module contains common parsers for PDFs.""" from typing import Any, Iterator, Mapping, Optional, Union from langchain.document_loaders.base import BaseBlobParser from langchain.document_loaders.blob_loaders import Blob from langchain.schema import Document [docs...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/pdf.html
b155063f625d-1
"""Initialize the parser. Args: text_kwargs: Keyword arguments to pass to ``fitz.Page.get_text()``. """ self.text_kwargs = text_kwargs or {} [docs] def lazy_parse(self, blob: Blob) -> Iterator[Document]: """Lazily parse the blob.""" import fitz with blob.as...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/pdf.html
b155063f625d-2
# if done incorrectly creates seg faults. with blob.as_bytes_io() as file_path: pdf_reader = pypdfium2.PdfDocument(file_path, autoclose=True) try: for page_number, page in enumerate(pdf_reader): text_page = page.get_textpage() conte...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/pdf.html
b155063f625d-3
}, ), ) for page in doc.pages ]
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/pdf.html
4cd92ac7dba1-0
Source code for langchain.document_loaders.parsers.generic """Code for generic / auxiliary parsers. This module contains some logic to help assemble more sophisticated parsers. """ from typing import Iterator, Mapping, Optional from langchain.document_loaders.base import BaseBlobParser from langchain.document_loaders.b...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/generic.html
4cd92ac7dba1-1
""" self.handlers = handlers self.fallback_parser = fallback_parser [docs] def lazy_parse(self, blob: Blob) -> Iterator[Document]: """Load documents from a blob.""" mimetype = blob.mimetype if mimetype is None: raise ValueError(f"{blob} does not have a mimetype.") ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/generic.html
7e7265154bc9-0
Source code for langchain.document_loaders.parsers.txt """Module for parsing text files..""" from typing import Iterator from langchain.document_loaders.base import BaseBlobParser from langchain.document_loaders.blob_loaders import Blob from langchain.schema import Document [docs]class TextParser(BaseBlobParser): "...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/txt.html
a5d0a177b784-0
Source code for langchain.document_loaders.parsers.registry """Module includes a registry of default parser configurations.""" from langchain.document_loaders.base import BaseBlobParser from langchain.document_loaders.parsers.generic import MimeTypeBasedParser from langchain.document_loaders.parsers.pdf import PyMuPDFP...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/registry.html
c6126d0c8db0-0
Source code for langchain.document_loaders.parsers.html.bs4 """Loader that uses bs4 to load HTML files, enriching metadata with page title.""" import logging from typing import Any, Dict, Iterator, Union from langchain.docstore.document import Document from langchain.document_loaders.base import BaseBlobParser from lan...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/html/bs4.html
62ed8298bced-0
Source code for langchain.document_loaders.parsers.language.code_segmenter from abc import ABC, abstractmethod from typing import List [docs]class CodeSegmenter(ABC): def __init__(self, code: str): self.code = code [docs] def is_valid(self) -> bool: return True [docs] @abstractmethod def s...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/language/code_segmenter.html
bec8ced10a2f-0
Source code for langchain.document_loaders.parsers.language.javascript from typing import Any, List from langchain.document_loaders.parsers.language.code_segmenter import CodeSegmenter [docs]class JavaScriptSegmenter(CodeSegmenter): def __init__(self, code: str): super().__init__(code) self.source_l...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/language/javascript.html
bec8ced10a2f-1
for node in tree.body: if isinstance( node, (esprima.nodes.FunctionDeclaration, esprima.nodes.ClassDeclaration), ): start = node.loc.start.line - 1 simplified_lines[start] = f"// Code for: {simplified_lines[start]}" ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/language/javascript.html
1a219eef5d7d-0
Source code for langchain.document_loaders.parsers.language.language_parser from typing import Any, Dict, Iterator, Optional from langchain.docstore.document import Document from langchain.document_loaders.base import BaseBlobParser from langchain.document_loaders.blob_loaders import Blob from langchain.document_loader...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/language/language_parser.html
1a219eef5d7d-1
docs = loader.load() Example instantiations to manually select the language: ... code-block:: python from langchain.text_splitter import Language loader = GenericLoader.from_filesystem( "./code", glob="**/*", suffixes=[".py"], ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/language/language_parser.html
1a219eef5d7d-2
"language": language, }, ) return self.Segmenter = LANGUAGE_SEGMENTERS[language] segmenter = self.Segmenter(blob.as_string()) if not segmenter.is_valid(): yield Document( page_content=code, metadata={ ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/language/language_parser.html
fc6956191354-0
Source code for langchain.document_loaders.parsers.language.python import ast from typing import Any, List from langchain.document_loaders.parsers.language.code_segmenter import CodeSegmenter [docs]class PythonSegmenter(CodeSegmenter): def __init__(self, code: str): super().__init__(code) self.sourc...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/parsers/language/python.html
d02f9d1cad25-0
Source code for langchain.document_loaders.blob_loaders.file_system """Use to load blobs from the local file system.""" from pathlib import Path from typing import Callable, Iterable, Iterator, Optional, Sequence, TypeVar, Union from langchain.document_loaders.blob_loaders.schema import Blob, BlobLoader T = TypeVar("T"...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/blob_loaders/file_system.html
d02f9d1cad25-1
*, glob: str = "**/[!.]*", suffixes: Optional[Sequence[str]] = None, show_progress: bool = False, ) -> None: """Initialize with path to directory and how to glob over it. Args: path: Path to directory to load from glob: Glob pattern relative to the spe...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/blob_loaders/file_system.html
d02f9d1cad25-2
self, ) -> Iterable[Blob]: """Yield blobs that match the requested pattern.""" iterator = _make_iterator( length_func=self.count_matching_files, show_progress=self.show_progress ) for path in iterator(self._yield_paths()): yield Blob.from_path(path) def _y...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/blob_loaders/file_system.html
d14be4bb58da-0
Source code for langchain.document_loaders.blob_loaders.youtube_audio from typing import Iterable, List from langchain.document_loaders.blob_loaders import FileSystemBlobLoader from langchain.document_loaders.blob_loaders.schema import Blob, BlobLoader [docs]class YoutubeAudioLoader(BlobLoader): """Load YouTube url...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/blob_loaders/youtube_audio.html
52bea49e6b46-0
Source code for langchain.document_loaders.blob_loaders.schema """Schema for Blobs and Blob Loaders. The goal is to facilitate decoupling of content loading from content parsing code. In addition, content loading code should provide a lazy loading interface by default. """ from __future__ import annotations import cont...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/blob_loaders/schema.html
52bea49e6b46-1
return str(self.path) if self.path else None [docs] @root_validator(pre=True) def check_blob_is_valid(cls, values: Mapping[str, Any]) -> Mapping[str, Any]: """Verify that either data or path is provided.""" if "data" not in values and "path" not in values: raise ValueError("Either dat...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/blob_loaders/schema.html
52bea49e6b46-2
yield f else: raise NotImplementedError(f"Unable to convert blob {self}") [docs] @classmethod def from_path( cls, path: PathLike, *, encoding: str = "utf-8", mime_type: Optional[str] = None, guess_type: bool = True, ) -> Blob: """Loa...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/blob_loaders/schema.html
52bea49e6b46-3
mime_type: if provided, will be set as the mime-type of the data path: if provided, will be set as the source from which the data came Returns: Blob instance """ return cls(data=data, mimetype=mime_type, encoding=encoding, path=path) def __repr__(self) -> str: ...
https://api.python.langchain.com/en/latest/_modules/langchain/document_loaders/blob_loaders/schema.html
c83fb8943a93-0
Source code for langchain.prompts.few_shot_with_templates """Prompt template that contains few shot examples.""" from typing import Any, Dict, List, Optional from pydantic import Extra, root_validator from langchain.prompts.base import DEFAULT_FORMATTER_MAPPING, StringPromptTemplate from langchain.prompts.example_selec...
https://api.python.langchain.com/en/latest/_modules/langchain/prompts/few_shot_with_templates.html
c83fb8943a93-1
examples = values.get("examples", None) example_selector = values.get("example_selector", None) if examples and example_selector: raise ValueError( "Only one of 'examples' and 'example_selector' should be provided" ) if examples is None and example_selecto...
https://api.python.langchain.com/en/latest/_modules/langchain/prompts/few_shot_with_templates.html
c83fb8943a93-2
"""Format the prompt with the inputs. Args: kwargs: Any arguments to be passed to the prompt template. Returns: A formatted string. Example: .. code-block:: python prompt.format(variable1="foo") """ kwargs = self._merge_partial_and_user...
https://api.python.langchain.com/en/latest/_modules/langchain/prompts/few_shot_with_templates.html