| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| from dataclasses import dataclass |
| from urllib.parse import urlparse |
|
|
| import asyncio |
| import aiohttp |
| from bs4 import BeautifulSoup |
| from llms_txt_rs import parse_llms_txt |
| from datasets import Dataset |
| from tqdm.asyncio import tqdm_asyncio |
| import json |
|
|
| |
| ParsedTxt = dict[str, str | None | dict[str, list[dict[str, str | None]]]] |
|
|
|
|
| @dataclass |
| class Page: |
| name: str |
| url: str | None |
| url_full: str | None |
| |
| llms_txt: str | None = None |
| llms_txt_parsed: ParsedTxt | None = None |
| llms_txt_full: str | None = None |
|
|
| @property |
| def base_url(self) -> str: |
| parsed = urlparse(self.url) |
| return f"{parsed.scheme}://{parsed.netloc}" |
|
|
| def add_text(self, txt: str, parsed_txt: ParsedTxt, is_full: bool = False, url: str | None = None) -> None: |
| if url and not url.startswith(self.base_url): |
| return |
| if not is_full: |
| self.llms_txt = txt |
| self.llms_txt_parsed = parsed_txt |
| else: |
| self.llms_txt_full = txt |
|
|
| def record(self) -> dict: |
| |
| |
| |
| return { |
| "name": self.name, |
| "url": self.url, |
| "url_full": self.url_full, |
| "llms_txt": self.llms_txt, |
| "llms_txt_parsed": json.dumps(self.llms_txt_parsed), |
| "llms_txt_full": self.llms_txt_full, |
| } |
|
|
|
|
| async def download_directory_llmstxt(url: str = "https://directory.llmstxt.cloud/") -> list[Page]: |
| async with aiohttp.ClientSession() as session: |
| async with session.get(url) as response: |
| content = await response.read() |
| soup = BeautifulSoup(content, 'html.parser') |
|
|
| pages = [parse_row(row) for row in soup.find_all('li')] |
|
|
| return pages |
|
|
|
|
| |
| def parse_row(row: list[str]) -> Page: |
| a_elems = row.find_all("a") |
| name = a_elems[0].find("h3").text |
| url_full = None |
| if len(a_elems) > 2: |
| for elem in a_elems[1:]: |
| url = elem.get("href") |
| if "llms-full.txt" in url: |
| url_full = url |
| else: |
| url = a_elems[1].get("href") |
|
|
| return Page( |
| name=name, |
| url=url, |
| url_full=url_full |
| ) |
|
|
|
|
| async def download_url(session: aiohttp.ClientSession, page: Page, url: str) -> Page: |
| try: |
| async with session.get(url) as response: |
| if response.status == 200: |
| txt = await response.text() |
| try: |
| parsed = parse_llms_txt(txt) |
| if "llms.txt" in url: |
| page.add_text(txt, parsed, is_full=False, url=url) |
| elif "llms-full.txt" in url: |
| page.add_text(txt, None, is_full=True, url=url) |
| else: |
| print(f"Unexpected url downloaded from the directory: {url}") |
| return page |
|
|
| except Exception as e: |
| |
| print(f"Error parsing '{url}': {e}") |
|
|
| except Exception as e: |
| print(f"Error downloading url '{url}': {e}") |
|
|
| return page |
|
|
|
|
| async def process_page(session: aiohttp.ClientSession, page: Page) -> list[Page]: |
| if url := page.url: |
| page = await download_url(session, page, url) |
| if url := page.url_full: |
| page = await download_url(session, page, url) |
|
|
| return page |
|
|
|
|
| async def download_pages(pages: list[Page]) -> list[Page]: |
| async with aiohttp.ClientSession() as session: |
| tasks = [process_page(session, page) for page in pages] |
| return await tqdm_asyncio.gather(*tasks, total=len(pages)) |
|
|
|
|
| async def main(): |
| print("Downloading directory...") |
| pages = await download_directory_llmstxt() |
|
|
| print("Downloading pages...") |
| pages = await download_pages(pages) |
|
|
| print("Preparing dataset...") |
| ds = Dataset.from_list([page.record() for page in pages]) |
| |
| ds.push_to_hub("plaguss/llms-txt", commit_message="Initial commit") |
|
|
|
|
| if __name__ == "__main__": |
| asyncio.run(main()) |
|
|