| | """ |
| | :mod:`pandas.io.html` is a module containing functionality for dealing with |
| | HTML IO. |
| | |
| | """ |
| |
|
| | from __future__ import annotations |
| |
|
| | from collections import abc |
| | import numbers |
| | import re |
| | from re import Pattern |
| | from typing import ( |
| | TYPE_CHECKING, |
| | Literal, |
| | cast, |
| | ) |
| | import warnings |
| |
|
| | from pandas._libs import lib |
| | from pandas.compat._optional import import_optional_dependency |
| | from pandas.errors import ( |
| | AbstractMethodError, |
| | EmptyDataError, |
| | ) |
| | from pandas.util._decorators import doc |
| | from pandas.util._exceptions import find_stack_level |
| | from pandas.util._validators import check_dtype_backend |
| |
|
| | from pandas.core.dtypes.common import is_list_like |
| |
|
| | from pandas import isna |
| | from pandas.core.indexes.base import Index |
| | from pandas.core.indexes.multi import MultiIndex |
| | from pandas.core.series import Series |
| | from pandas.core.shared_docs import _shared_docs |
| |
|
| | from pandas.io.common import ( |
| | file_exists, |
| | get_handle, |
| | is_file_like, |
| | is_fsspec_url, |
| | is_url, |
| | stringify_path, |
| | validate_header_arg, |
| | ) |
| | from pandas.io.formats.printing import pprint_thing |
| | from pandas.io.parsers import TextParser |
| |
|
| | if TYPE_CHECKING: |
| | from collections.abc import ( |
| | Iterable, |
| | Sequence, |
| | ) |
| |
|
| | from pandas._typing import ( |
| | BaseBuffer, |
| | DtypeBackend, |
| | FilePath, |
| | HTMLFlavors, |
| | ReadBuffer, |
| | StorageOptions, |
| | ) |
| |
|
| | from pandas import DataFrame |
| |
|
| | |
| | |
| | |
| | _RE_WHITESPACE = re.compile(r"[\r\n]+|\s{2,}") |
| |
|
| |
|
| | def _remove_whitespace(s: str, regex: Pattern = _RE_WHITESPACE) -> str: |
| | """ |
| | Replace extra whitespace inside of a string with a single space. |
| | |
| | Parameters |
| | ---------- |
| | s : str or unicode |
| | The string from which to remove extra whitespace. |
| | regex : re.Pattern |
| | The regular expression to use to remove extra whitespace. |
| | |
| | Returns |
| | ------- |
| | subd : str or unicode |
| | `s` with all extra whitespace replaced with a single space. |
| | """ |
| | return regex.sub(" ", s.strip()) |
| |
|
| |
|
| | def _get_skiprows(skiprows: int | Sequence[int] | slice | None) -> int | Sequence[int]: |
| | """ |
| | Get an iterator given an integer, slice or container. |
| | |
| | Parameters |
| | ---------- |
| | skiprows : int, slice, container |
| | The iterator to use to skip rows; can also be a slice. |
| | |
| | Raises |
| | ------ |
| | TypeError |
| | * If `skiprows` is not a slice, integer, or Container |
| | |
| | Returns |
| | ------- |
| | it : iterable |
| | A proper iterator to use to skip rows of a DataFrame. |
| | """ |
| | if isinstance(skiprows, slice): |
| | start, step = skiprows.start or 0, skiprows.step or 1 |
| | return list(range(start, skiprows.stop, step)) |
| | elif isinstance(skiprows, numbers.Integral) or is_list_like(skiprows): |
| | return cast("int | Sequence[int]", skiprows) |
| | elif skiprows is None: |
| | return 0 |
| | raise TypeError(f"{type(skiprows).__name__} is not a valid type for skipping rows") |
| |
|
| |
|
| | def _read( |
| | obj: FilePath | BaseBuffer, |
| | encoding: str | None, |
| | storage_options: StorageOptions | None, |
| | ) -> str | bytes: |
| | """ |
| | Try to read from a url, file or string. |
| | |
| | Parameters |
| | ---------- |
| | obj : str, unicode, path object, or file-like object |
| | |
| | Returns |
| | ------- |
| | raw_text : str |
| | """ |
| | text: str | bytes |
| | if ( |
| | is_url(obj) |
| | or hasattr(obj, "read") |
| | or (isinstance(obj, str) and file_exists(obj)) |
| | ): |
| | with get_handle( |
| | obj, "r", encoding=encoding, storage_options=storage_options |
| | ) as handles: |
| | text = handles.handle.read() |
| | elif isinstance(obj, (str, bytes)): |
| | text = obj |
| | else: |
| | raise TypeError(f"Cannot read object of type '{type(obj).__name__}'") |
| | return text |
| |
|
| |
|
| | class _HtmlFrameParser: |
| | """ |
| | Base class for parsers that parse HTML into DataFrames. |
| | |
| | Parameters |
| | ---------- |
| | io : str or file-like |
| | This can be either a string of raw HTML, a valid URL using the HTTP, |
| | FTP, or FILE protocols or a file-like object. |
| | |
| | match : str or regex |
| | The text to match in the document. |
| | |
| | attrs : dict |
| | List of HTML <table> element attributes to match. |
| | |
| | encoding : str |
| | Encoding to be used by parser |
| | |
| | displayed_only : bool |
| | Whether or not items with "display:none" should be ignored |
| | |
| | extract_links : {None, "all", "header", "body", "footer"} |
| | Table elements in the specified section(s) with <a> tags will have their |
| | href extracted. |
| | |
| | .. versionadded:: 1.5.0 |
| | |
| | Attributes |
| | ---------- |
| | io : str or file-like |
| | raw HTML, URL, or file-like object |
| | |
| | match : regex |
| | The text to match in the raw HTML |
| | |
| | attrs : dict-like |
| | A dictionary of valid table attributes to use to search for table |
| | elements. |
| | |
| | encoding : str |
| | Encoding to be used by parser |
| | |
| | displayed_only : bool |
| | Whether or not items with "display:none" should be ignored |
| | |
| | extract_links : {None, "all", "header", "body", "footer"} |
| | Table elements in the specified section(s) with <a> tags will have their |
| | href extracted. |
| | |
| | .. versionadded:: 1.5.0 |
| | |
| | Notes |
| | ----- |
| | To subclass this class effectively you must override the following methods: |
| | * :func:`_build_doc` |
| | * :func:`_attr_getter` |
| | * :func:`_href_getter` |
| | * :func:`_text_getter` |
| | * :func:`_parse_td` |
| | * :func:`_parse_thead_tr` |
| | * :func:`_parse_tbody_tr` |
| | * :func:`_parse_tfoot_tr` |
| | * :func:`_parse_tables` |
| | * :func:`_equals_tag` |
| | See each method's respective documentation for details on their |
| | functionality. |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | io: FilePath | ReadBuffer[str] | ReadBuffer[bytes], |
| | match: str | Pattern, |
| | attrs: dict[str, str] | None, |
| | encoding: str, |
| | displayed_only: bool, |
| | extract_links: Literal[None, "header", "footer", "body", "all"], |
| | storage_options: StorageOptions = None, |
| | ) -> None: |
| | self.io = io |
| | self.match = match |
| | self.attrs = attrs |
| | self.encoding = encoding |
| | self.displayed_only = displayed_only |
| | self.extract_links = extract_links |
| | self.storage_options = storage_options |
| |
|
| | def parse_tables(self): |
| | """ |
| | Parse and return all tables from the DOM. |
| | |
| | Returns |
| | ------- |
| | list of parsed (header, body, footer) tuples from tables. |
| | """ |
| | tables = self._parse_tables(self._build_doc(), self.match, self.attrs) |
| | return (self._parse_thead_tbody_tfoot(table) for table in tables) |
| |
|
| | def _attr_getter(self, obj, attr): |
| | """ |
| | Return the attribute value of an individual DOM node. |
| | |
| | Parameters |
| | ---------- |
| | obj : node-like |
| | A DOM node. |
| | |
| | attr : str or unicode |
| | The attribute, such as "colspan" |
| | |
| | Returns |
| | ------- |
| | str or unicode |
| | The attribute value. |
| | """ |
| | |
| | return obj.get(attr) |
| |
|
| | def _href_getter(self, obj) -> str | None: |
| | """ |
| | Return a href if the DOM node contains a child <a> or None. |
| | |
| | Parameters |
| | ---------- |
| | obj : node-like |
| | A DOM node. |
| | |
| | Returns |
| | ------- |
| | href : str or unicode |
| | The href from the <a> child of the DOM node. |
| | """ |
| | raise AbstractMethodError(self) |
| |
|
| | def _text_getter(self, obj): |
| | """ |
| | Return the text of an individual DOM node. |
| | |
| | Parameters |
| | ---------- |
| | obj : node-like |
| | A DOM node. |
| | |
| | Returns |
| | ------- |
| | text : str or unicode |
| | The text from an individual DOM node. |
| | """ |
| | raise AbstractMethodError(self) |
| |
|
| | def _parse_td(self, obj): |
| | """ |
| | Return the td elements from a row element. |
| | |
| | Parameters |
| | ---------- |
| | obj : node-like |
| | A DOM <tr> node. |
| | |
| | Returns |
| | ------- |
| | list of node-like |
| | These are the elements of each row, i.e., the columns. |
| | """ |
| | raise AbstractMethodError(self) |
| |
|
| | def _parse_thead_tr(self, table): |
| | """ |
| | Return the list of thead row elements from the parsed table element. |
| | |
| | Parameters |
| | ---------- |
| | table : a table element that contains zero or more thead elements. |
| | |
| | Returns |
| | ------- |
| | list of node-like |
| | These are the <tr> row elements of a table. |
| | """ |
| | raise AbstractMethodError(self) |
| |
|
| | def _parse_tbody_tr(self, table): |
| | """ |
| | Return the list of tbody row elements from the parsed table element. |
| | |
| | HTML5 table bodies consist of either 0 or more <tbody> elements (which |
| | only contain <tr> elements) or 0 or more <tr> elements. This method |
| | checks for both structures. |
| | |
| | Parameters |
| | ---------- |
| | table : a table element that contains row elements. |
| | |
| | Returns |
| | ------- |
| | list of node-like |
| | These are the <tr> row elements of a table. |
| | """ |
| | raise AbstractMethodError(self) |
| |
|
| | def _parse_tfoot_tr(self, table): |
| | """ |
| | Return the list of tfoot row elements from the parsed table element. |
| | |
| | Parameters |
| | ---------- |
| | table : a table element that contains row elements. |
| | |
| | Returns |
| | ------- |
| | list of node-like |
| | These are the <tr> row elements of a table. |
| | """ |
| | raise AbstractMethodError(self) |
| |
|
| | def _parse_tables(self, document, match, attrs): |
| | """ |
| | Return all tables from the parsed DOM. |
| | |
| | Parameters |
| | ---------- |
| | document : the DOM from which to parse the table element. |
| | |
| | match : str or regular expression |
| | The text to search for in the DOM tree. |
| | |
| | attrs : dict |
| | A dictionary of table attributes that can be used to disambiguate |
| | multiple tables on a page. |
| | |
| | Raises |
| | ------ |
| | ValueError : `match` does not match any text in the document. |
| | |
| | Returns |
| | ------- |
| | list of node-like |
| | HTML <table> elements to be parsed into raw data. |
| | """ |
| | raise AbstractMethodError(self) |
| |
|
| | def _equals_tag(self, obj, tag) -> bool: |
| | """ |
| | Return whether an individual DOM node matches a tag |
| | |
| | Parameters |
| | ---------- |
| | obj : node-like |
| | A DOM node. |
| | |
| | tag : str |
| | Tag name to be checked for equality. |
| | |
| | Returns |
| | ------- |
| | boolean |
| | Whether `obj`'s tag name is `tag` |
| | """ |
| | raise AbstractMethodError(self) |
| |
|
| | def _build_doc(self): |
| | """ |
| | Return a tree-like object that can be used to iterate over the DOM. |
| | |
| | Returns |
| | ------- |
| | node-like |
| | The DOM from which to parse the table element. |
| | """ |
| | raise AbstractMethodError(self) |
| |
|
| | def _parse_thead_tbody_tfoot(self, table_html): |
| | """ |
| | Given a table, return parsed header, body, and foot. |
| | |
| | Parameters |
| | ---------- |
| | table_html : node-like |
| | |
| | Returns |
| | ------- |
| | tuple of (header, body, footer), each a list of list-of-text rows. |
| | |
| | Notes |
| | ----- |
| | Header and body are lists-of-lists. Top level list is a list of |
| | rows. Each row is a list of str text. |
| | |
| | Logic: Use <thead>, <tbody>, <tfoot> elements to identify |
| | header, body, and footer, otherwise: |
| | - Put all rows into body |
| | - Move rows from top of body to header only if |
| | all elements inside row are <th> |
| | - Move rows from bottom of body to footer only if |
| | all elements inside row are <th> |
| | """ |
| | header_rows = self._parse_thead_tr(table_html) |
| | body_rows = self._parse_tbody_tr(table_html) |
| | footer_rows = self._parse_tfoot_tr(table_html) |
| |
|
| | def row_is_all_th(row): |
| | return all(self._equals_tag(t, "th") for t in self._parse_td(row)) |
| |
|
| | if not header_rows: |
| | |
| | |
| | |
| | while body_rows and row_is_all_th(body_rows[0]): |
| | header_rows.append(body_rows.pop(0)) |
| |
|
| | header = self._expand_colspan_rowspan(header_rows, section="header") |
| | body = self._expand_colspan_rowspan(body_rows, section="body") |
| | footer = self._expand_colspan_rowspan(footer_rows, section="footer") |
| |
|
| | return header, body, footer |
| |
|
| | def _expand_colspan_rowspan( |
| | self, rows, section: Literal["header", "footer", "body"] |
| | ): |
| | """ |
| | Given a list of <tr>s, return a list of text rows. |
| | |
| | Parameters |
| | ---------- |
| | rows : list of node-like |
| | List of <tr>s |
| | section : the section that the rows belong to (header, body or footer). |
| | |
| | Returns |
| | ------- |
| | list of list |
| | Each returned row is a list of str text, or tuple (text, link) |
| | if extract_links is not None. |
| | |
| | Notes |
| | ----- |
| | Any cell with ``rowspan`` or ``colspan`` will have its contents copied |
| | to subsequent cells. |
| | """ |
| | all_texts = [] |
| | text: str | tuple |
| | remainder: list[ |
| | tuple[int, str | tuple, int] |
| | ] = [] |
| |
|
| | for tr in rows: |
| | texts = [] |
| | next_remainder = [] |
| |
|
| | index = 0 |
| | tds = self._parse_td(tr) |
| | for td in tds: |
| | |
| | |
| | while remainder and remainder[0][0] <= index: |
| | prev_i, prev_text, prev_rowspan = remainder.pop(0) |
| | texts.append(prev_text) |
| | if prev_rowspan > 1: |
| | next_remainder.append((prev_i, prev_text, prev_rowspan - 1)) |
| | index += 1 |
| |
|
| | |
| | text = _remove_whitespace(self._text_getter(td)) |
| | if self.extract_links in ("all", section): |
| | href = self._href_getter(td) |
| | text = (text, href) |
| | rowspan = int(self._attr_getter(td, "rowspan") or 1) |
| | colspan = int(self._attr_getter(td, "colspan") or 1) |
| |
|
| | for _ in range(colspan): |
| | texts.append(text) |
| | if rowspan > 1: |
| | next_remainder.append((index, text, rowspan - 1)) |
| | index += 1 |
| |
|
| | |
| | for prev_i, prev_text, prev_rowspan in remainder: |
| | texts.append(prev_text) |
| | if prev_rowspan > 1: |
| | next_remainder.append((prev_i, prev_text, prev_rowspan - 1)) |
| |
|
| | all_texts.append(texts) |
| | remainder = next_remainder |
| |
|
| | |
| | |
| | while remainder: |
| | next_remainder = [] |
| | texts = [] |
| | for prev_i, prev_text, prev_rowspan in remainder: |
| | texts.append(prev_text) |
| | if prev_rowspan > 1: |
| | next_remainder.append((prev_i, prev_text, prev_rowspan - 1)) |
| | all_texts.append(texts) |
| | remainder = next_remainder |
| |
|
| | return all_texts |
| |
|
| | def _handle_hidden_tables(self, tbl_list, attr_name: str): |
| | """ |
| | Return list of tables, potentially removing hidden elements |
| | |
| | Parameters |
| | ---------- |
| | tbl_list : list of node-like |
| | Type of list elements will vary depending upon parser used |
| | attr_name : str |
| | Name of the accessor for retrieving HTML attributes |
| | |
| | Returns |
| | ------- |
| | list of node-like |
| | Return type matches `tbl_list` |
| | """ |
| | if not self.displayed_only: |
| | return tbl_list |
| |
|
| | return [ |
| | x |
| | for x in tbl_list |
| | if "display:none" |
| | not in getattr(x, attr_name).get("style", "").replace(" ", "") |
| | ] |
| |
|
| |
|
| | class _BeautifulSoupHtml5LibFrameParser(_HtmlFrameParser): |
| | """ |
| | HTML to DataFrame parser that uses BeautifulSoup under the hood. |
| | |
| | See Also |
| | -------- |
| | pandas.io.html._HtmlFrameParser |
| | pandas.io.html._LxmlFrameParser |
| | |
| | Notes |
| | ----- |
| | Documentation strings for this class are in the base class |
| | :class:`pandas.io.html._HtmlFrameParser`. |
| | """ |
| |
|
| | def _parse_tables(self, document, match, attrs): |
| | element_name = "table" |
| | tables = document.find_all(element_name, attrs=attrs) |
| | if not tables: |
| | raise ValueError("No tables found") |
| |
|
| | result = [] |
| | unique_tables = set() |
| | tables = self._handle_hidden_tables(tables, "attrs") |
| |
|
| | for table in tables: |
| | if self.displayed_only: |
| | for elem in table.find_all("style"): |
| | elem.decompose() |
| |
|
| | for elem in table.find_all(style=re.compile(r"display:\s*none")): |
| | elem.decompose() |
| |
|
| | if table not in unique_tables and table.find(string=match) is not None: |
| | result.append(table) |
| | unique_tables.add(table) |
| | if not result: |
| | raise ValueError(f"No tables found matching pattern {repr(match.pattern)}") |
| | return result |
| |
|
| | def _href_getter(self, obj) -> str | None: |
| | a = obj.find("a", href=True) |
| | return None if not a else a["href"] |
| |
|
| | def _text_getter(self, obj): |
| | return obj.text |
| |
|
| | def _equals_tag(self, obj, tag) -> bool: |
| | return obj.name == tag |
| |
|
| | def _parse_td(self, row): |
| | return row.find_all(("td", "th"), recursive=False) |
| |
|
| | def _parse_thead_tr(self, table): |
| | return table.select("thead tr") |
| |
|
| | def _parse_tbody_tr(self, table): |
| | from_tbody = table.select("tbody tr") |
| | from_root = table.find_all("tr", recursive=False) |
| | |
| | return from_tbody + from_root |
| |
|
| | def _parse_tfoot_tr(self, table): |
| | return table.select("tfoot tr") |
| |
|
| | def _setup_build_doc(self): |
| | raw_text = _read(self.io, self.encoding, self.storage_options) |
| | if not raw_text: |
| | raise ValueError(f"No text parsed from document: {self.io}") |
| | return raw_text |
| |
|
| | def _build_doc(self): |
| | from bs4 import BeautifulSoup |
| |
|
| | bdoc = self._setup_build_doc() |
| | if isinstance(bdoc, bytes) and self.encoding is not None: |
| | udoc = bdoc.decode(self.encoding) |
| | from_encoding = None |
| | else: |
| | udoc = bdoc |
| | from_encoding = self.encoding |
| |
|
| | soup = BeautifulSoup(udoc, features="html5lib", from_encoding=from_encoding) |
| |
|
| | for br in soup.find_all("br"): |
| | br.replace_with("\n" + br.text) |
| |
|
| | return soup |
| |
|
| |
|
| | def _build_xpath_expr(attrs) -> str: |
| | """ |
| | Build an xpath expression to simulate bs4's ability to pass in kwargs to |
| | search for attributes when using the lxml parser. |
| | |
| | Parameters |
| | ---------- |
| | attrs : dict |
| | A dict of HTML attributes. These are NOT checked for validity. |
| | |
| | Returns |
| | ------- |
| | expr : unicode |
| | An XPath expression that checks for the given HTML attributes. |
| | """ |
| | |
| | if "class_" in attrs: |
| | attrs["class"] = attrs.pop("class_") |
| |
|
| | s = " and ".join([f"@{k}={repr(v)}" for k, v in attrs.items()]) |
| | return f"[{s}]" |
| |
|
| |
|
| | _re_namespace = {"re": "http://exslt.org/regular-expressions"} |
| |
|
| |
|
| | class _LxmlFrameParser(_HtmlFrameParser): |
| | """ |
| | HTML to DataFrame parser that uses lxml under the hood. |
| | |
| | Warning |
| | ------- |
| | This parser can only handle HTTP, FTP, and FILE urls. |
| | |
| | See Also |
| | -------- |
| | _HtmlFrameParser |
| | _BeautifulSoupLxmlFrameParser |
| | |
| | Notes |
| | ----- |
| | Documentation strings for this class are in the base class |
| | :class:`_HtmlFrameParser`. |
| | """ |
| |
|
| | def _href_getter(self, obj) -> str | None: |
| | href = obj.xpath(".//a/@href") |
| | return None if not href else href[0] |
| |
|
| | def _text_getter(self, obj): |
| | return obj.text_content() |
| |
|
| | def _parse_td(self, row): |
| | |
| | |
| | return row.xpath("./td|./th") |
| |
|
| | def _parse_tables(self, document, match, kwargs): |
| | pattern = match.pattern |
| |
|
| | |
| | |
| | xpath_expr = f"//table[.//text()[re:test(., {repr(pattern)})]]" |
| |
|
| | |
| | |
| | if kwargs: |
| | xpath_expr += _build_xpath_expr(kwargs) |
| |
|
| | tables = document.xpath(xpath_expr, namespaces=_re_namespace) |
| |
|
| | tables = self._handle_hidden_tables(tables, "attrib") |
| | if self.displayed_only: |
| | for table in tables: |
| | |
| | |
| | |
| | for elem in table.xpath(".//style"): |
| | elem.drop_tree() |
| | for elem in table.xpath(".//*[@style]"): |
| | if "display:none" in elem.attrib.get("style", "").replace(" ", ""): |
| | elem.drop_tree() |
| | if not tables: |
| | raise ValueError(f"No tables found matching regex {repr(pattern)}") |
| | return tables |
| |
|
| | def _equals_tag(self, obj, tag) -> bool: |
| | return obj.tag == tag |
| |
|
| | def _build_doc(self): |
| | """ |
| | Raises |
| | ------ |
| | ValueError |
| | * If a URL that lxml cannot parse is passed. |
| | |
| | Exception |
| | * Any other ``Exception`` thrown. For example, trying to parse a |
| | URL that is syntactically correct on a machine with no internet |
| | connection will fail. |
| | |
| | See Also |
| | -------- |
| | pandas.io.html._HtmlFrameParser._build_doc |
| | """ |
| | from lxml.etree import XMLSyntaxError |
| | from lxml.html import ( |
| | HTMLParser, |
| | fromstring, |
| | parse, |
| | ) |
| |
|
| | parser = HTMLParser(recover=True, encoding=self.encoding) |
| |
|
| | try: |
| | if is_url(self.io): |
| | with get_handle( |
| | self.io, "r", storage_options=self.storage_options |
| | ) as f: |
| | r = parse(f.handle, parser=parser) |
| | else: |
| | |
| | r = parse(self.io, parser=parser) |
| | try: |
| | r = r.getroot() |
| | except AttributeError: |
| | pass |
| | except (UnicodeDecodeError, OSError) as e: |
| | |
| | if not is_url(self.io): |
| | r = fromstring(self.io, parser=parser) |
| |
|
| | try: |
| | r = r.getroot() |
| | except AttributeError: |
| | pass |
| | else: |
| | raise e |
| | else: |
| | if not hasattr(r, "text_content"): |
| | raise XMLSyntaxError("no text parsed from document", 0, 0, 0) |
| |
|
| | for br in r.xpath("*//br"): |
| | br.tail = "\n" + (br.tail or "") |
| |
|
| | return r |
| |
|
| | def _parse_thead_tr(self, table): |
| | rows = [] |
| |
|
| | for thead in table.xpath(".//thead"): |
| | rows.extend(thead.xpath("./tr")) |
| |
|
| | |
| | |
| | |
| | |
| | |
| | |
| | elements_at_root = thead.xpath("./td|./th") |
| | if elements_at_root: |
| | rows.append(thead) |
| |
|
| | return rows |
| |
|
| | def _parse_tbody_tr(self, table): |
| | from_tbody = table.xpath(".//tbody//tr") |
| | from_root = table.xpath("./tr") |
| | |
| | return from_tbody + from_root |
| |
|
| | def _parse_tfoot_tr(self, table): |
| | return table.xpath(".//tfoot//tr") |
| |
|
| |
|
| | def _expand_elements(body) -> None: |
| | data = [len(elem) for elem in body] |
| | lens = Series(data) |
| | lens_max = lens.max() |
| | not_max = lens[lens != lens_max] |
| |
|
| | empty = [""] |
| | for ind, length in not_max.items(): |
| | body[ind] += empty * (lens_max - length) |
| |
|
| |
|
| | def _data_to_frame(**kwargs): |
| | head, body, foot = kwargs.pop("data") |
| | header = kwargs.pop("header") |
| | kwargs["skiprows"] = _get_skiprows(kwargs["skiprows"]) |
| | if head: |
| | body = head + body |
| |
|
| | |
| | if header is None: |
| | if len(head) == 1: |
| | header = 0 |
| | else: |
| | |
| | header = [i for i, row in enumerate(head) if any(text for text in row)] |
| |
|
| | if foot: |
| | body += foot |
| |
|
| | |
| | _expand_elements(body) |
| | with TextParser(body, header=header, **kwargs) as tp: |
| | return tp.read() |
| |
|
| |
|
| | _valid_parsers = { |
| | "lxml": _LxmlFrameParser, |
| | None: _LxmlFrameParser, |
| | "html5lib": _BeautifulSoupHtml5LibFrameParser, |
| | "bs4": _BeautifulSoupHtml5LibFrameParser, |
| | } |
| |
|
| |
|
| | def _parser_dispatch(flavor: HTMLFlavors | None) -> type[_HtmlFrameParser]: |
| | """ |
| | Choose the parser based on the input flavor. |
| | |
| | Parameters |
| | ---------- |
| | flavor : {{"lxml", "html5lib", "bs4"}} or None |
| | The type of parser to use. This must be a valid backend. |
| | |
| | Returns |
| | ------- |
| | cls : _HtmlFrameParser subclass |
| | The parser class based on the requested input flavor. |
| | |
| | Raises |
| | ------ |
| | ValueError |
| | * If `flavor` is not a valid backend. |
| | ImportError |
| | * If you do not have the requested `flavor` |
| | """ |
| | valid_parsers = list(_valid_parsers.keys()) |
| | if flavor not in valid_parsers: |
| | raise ValueError( |
| | f"{repr(flavor)} is not a valid flavor, valid flavors are {valid_parsers}" |
| | ) |
| |
|
| | if flavor in ("bs4", "html5lib"): |
| | import_optional_dependency("html5lib") |
| | import_optional_dependency("bs4") |
| | else: |
| | import_optional_dependency("lxml.etree") |
| | return _valid_parsers[flavor] |
| |
|
| |
|
| | def _print_as_set(s) -> str: |
| | arg = ", ".join([pprint_thing(el) for el in s]) |
| | return f"{{{arg}}}" |
| |
|
| |
|
| | def _validate_flavor(flavor): |
| | if flavor is None: |
| | flavor = "lxml", "bs4" |
| | elif isinstance(flavor, str): |
| | flavor = (flavor,) |
| | elif isinstance(flavor, abc.Iterable): |
| | if not all(isinstance(flav, str) for flav in flavor): |
| | raise TypeError( |
| | f"Object of type {repr(type(flavor).__name__)} " |
| | f"is not an iterable of strings" |
| | ) |
| | else: |
| | msg = repr(flavor) if isinstance(flavor, str) else str(flavor) |
| | msg += " is not a valid flavor" |
| | raise ValueError(msg) |
| |
|
| | flavor = tuple(flavor) |
| | valid_flavors = set(_valid_parsers) |
| | flavor_set = set(flavor) |
| |
|
| | if not flavor_set & valid_flavors: |
| | raise ValueError( |
| | f"{_print_as_set(flavor_set)} is not a valid set of flavors, valid " |
| | f"flavors are {_print_as_set(valid_flavors)}" |
| | ) |
| | return flavor |
| |
|
| |
|
| | def _parse( |
| | flavor, |
| | io, |
| | match, |
| | attrs, |
| | encoding, |
| | displayed_only, |
| | extract_links, |
| | storage_options, |
| | **kwargs, |
| | ): |
| | flavor = _validate_flavor(flavor) |
| | compiled_match = re.compile(match) |
| |
|
| | retained = None |
| | for flav in flavor: |
| | parser = _parser_dispatch(flav) |
| | p = parser( |
| | io, |
| | compiled_match, |
| | attrs, |
| | encoding, |
| | displayed_only, |
| | extract_links, |
| | storage_options, |
| | ) |
| |
|
| | try: |
| | tables = p.parse_tables() |
| | except ValueError as caught: |
| | |
| | |
| | if hasattr(io, "seekable") and io.seekable(): |
| | io.seek(0) |
| | elif hasattr(io, "seekable") and not io.seekable(): |
| | |
| | raise ValueError( |
| | f"The flavor {flav} failed to parse your input. " |
| | "Since you passed a non-rewindable file " |
| | "object, we can't rewind it to try " |
| | "another parser. Try read_html() with a different flavor." |
| | ) from caught |
| |
|
| | retained = caught |
| | else: |
| | break |
| | else: |
| | assert retained is not None |
| | raise retained |
| |
|
| | ret = [] |
| | for table in tables: |
| | try: |
| | df = _data_to_frame(data=table, **kwargs) |
| | |
| | |
| | |
| | if extract_links in ("all", "header") and isinstance( |
| | df.columns, MultiIndex |
| | ): |
| | df.columns = Index( |
| | ((col[0], None if isna(col[1]) else col[1]) for col in df.columns), |
| | tupleize_cols=False, |
| | ) |
| |
|
| | ret.append(df) |
| | except EmptyDataError: |
| | continue |
| | return ret |
| |
|
| |
|
| | @doc(storage_options=_shared_docs["storage_options"]) |
| | def read_html( |
| | io: FilePath | ReadBuffer[str], |
| | *, |
| | match: str | Pattern = ".+", |
| | flavor: HTMLFlavors | Sequence[HTMLFlavors] | None = None, |
| | header: int | Sequence[int] | None = None, |
| | index_col: int | Sequence[int] | None = None, |
| | skiprows: int | Sequence[int] | slice | None = None, |
| | attrs: dict[str, str] | None = None, |
| | parse_dates: bool = False, |
| | thousands: str | None = ",", |
| | encoding: str | None = None, |
| | decimal: str = ".", |
| | converters: dict | None = None, |
| | na_values: Iterable[object] | None = None, |
| | keep_default_na: bool = True, |
| | displayed_only: bool = True, |
| | extract_links: Literal[None, "header", "footer", "body", "all"] = None, |
| | dtype_backend: DtypeBackend | lib.NoDefault = lib.no_default, |
| | storage_options: StorageOptions = None, |
| | ) -> list[DataFrame]: |
| | r""" |
| | Read HTML tables into a ``list`` of ``DataFrame`` objects. |
| | |
| | Parameters |
| | ---------- |
| | io : str, path object, or file-like object |
| | String, path object (implementing ``os.PathLike[str]``), or file-like |
| | object implementing a string ``read()`` function. |
| | The string can represent a URL or the HTML itself. Note that |
| | lxml only accepts the http, ftp and file url protocols. If you have a |
| | URL that starts with ``'https'`` you might try removing the ``'s'``. |
| | |
| | .. deprecated:: 2.1.0 |
| | Passing html literal strings is deprecated. |
| | Wrap literal string/bytes input in ``io.StringIO``/``io.BytesIO`` instead. |
| | |
| | match : str or compiled regular expression, optional |
| | The set of tables containing text matching this regex or string will be |
| | returned. Unless the HTML is extremely simple you will probably need to |
| | pass a non-empty string here. Defaults to '.+' (match any non-empty |
| | string). The default value will return all tables contained on a page. |
| | This value is converted to a regular expression so that there is |
| | consistent behavior between Beautiful Soup and lxml. |
| | |
| | flavor : {{"lxml", "html5lib", "bs4"}} or list-like, optional |
| | The parsing engine (or list of parsing engines) to use. 'bs4' and |
| | 'html5lib' are synonymous with each other, they are both there for |
| | backwards compatibility. The default of ``None`` tries to use ``lxml`` |
| | to parse and if that fails it falls back on ``bs4`` + ``html5lib``. |
| | |
| | header : int or list-like, optional |
| | The row (or list of rows for a :class:`~pandas.MultiIndex`) to use to |
| | make the columns headers. |
| | |
| | index_col : int or list-like, optional |
| | The column (or list of columns) to use to create the index. |
| | |
| | skiprows : int, list-like or slice, optional |
| | Number of rows to skip after parsing the column integer. 0-based. If a |
| | sequence of integers or a slice is given, will skip the rows indexed by |
| | that sequence. Note that a single element sequence means 'skip the nth |
| | row' whereas an integer means 'skip n rows'. |
| | |
| | attrs : dict, optional |
| | This is a dictionary of attributes that you can pass to use to identify |
| | the table in the HTML. These are not checked for validity before being |
| | passed to lxml or Beautiful Soup. However, these attributes must be |
| | valid HTML table attributes to work correctly. For example, :: |
| | |
| | attrs = {{'id': 'table'}} |
| | |
| | is a valid attribute dictionary because the 'id' HTML tag attribute is |
| | a valid HTML attribute for *any* HTML tag as per `this document |
| | <https://html.spec.whatwg.org/multipage/dom.html#global-attributes>`__. :: |
| | |
| | attrs = {{'asdf': 'table'}} |
| | |
| | is *not* a valid attribute dictionary because 'asdf' is not a valid |
| | HTML attribute even if it is a valid XML attribute. Valid HTML 4.01 |
| | table attributes can be found `here |
| | <http://www.w3.org/TR/REC-html40/struct/tables.html#h-11.2>`__. A |
| | working draft of the HTML 5 spec can be found `here |
| | <https://html.spec.whatwg.org/multipage/tables.html>`__. It contains the |
| | latest information on table attributes for the modern web. |
| | |
| | parse_dates : bool, optional |
| | See :func:`~read_csv` for more details. |
| | |
| | thousands : str, optional |
| | Separator to use to parse thousands. Defaults to ``','``. |
| | |
| | encoding : str, optional |
| | The encoding used to decode the web page. Defaults to ``None``.``None`` |
| | preserves the previous encoding behavior, which depends on the |
| | underlying parser library (e.g., the parser library will try to use |
| | the encoding provided by the document). |
| | |
| | decimal : str, default '.' |
| | Character to recognize as decimal point (e.g. use ',' for European |
| | data). |
| | |
| | converters : dict, default None |
| | Dict of functions for converting values in certain columns. Keys can |
| | either be integers or column labels, values are functions that take one |
| | input argument, the cell (not column) content, and return the |
| | transformed content. |
| | |
| | na_values : iterable, default None |
| | Custom NA values. |
| | |
| | keep_default_na : bool, default True |
| | If na_values are specified and keep_default_na is False the default NaN |
| | values are overridden, otherwise they're appended to. |
| | |
| | displayed_only : bool, default True |
| | Whether elements with "display: none" should be parsed. |
| | |
| | extract_links : {{None, "all", "header", "body", "footer"}} |
| | Table elements in the specified section(s) with <a> tags will have their |
| | href extracted. |
| | |
| | .. versionadded:: 1.5.0 |
| | |
| | dtype_backend : {{'numpy_nullable', 'pyarrow'}}, default 'numpy_nullable' |
| | Back-end data type applied to the resultant :class:`DataFrame` |
| | (still experimental). Behaviour is as follows: |
| | |
| | * ``"numpy_nullable"``: returns nullable-dtype-backed :class:`DataFrame` |
| | (default). |
| | * ``"pyarrow"``: returns pyarrow-backed nullable :class:`ArrowDtype` |
| | DataFrame. |
| | |
| | .. versionadded:: 2.0 |
| | |
| | {storage_options} |
| | |
| | .. versionadded:: 2.1.0 |
| | |
| | Returns |
| | ------- |
| | dfs |
| | A list of DataFrames. |
| | |
| | See Also |
| | -------- |
| | read_csv : Read a comma-separated values (csv) file into DataFrame. |
| | |
| | Notes |
| | ----- |
| | Before using this function you should read the :ref:`gotchas about the |
| | HTML parsing libraries <io.html.gotchas>`. |
| | |
| | Expect to do some cleanup after you call this function. For example, you |
| | might need to manually assign column names if the column names are |
| | converted to NaN when you pass the `header=0` argument. We try to assume as |
| | little as possible about the structure of the table and push the |
| | idiosyncrasies of the HTML contained in the table to the user. |
| | |
| | This function searches for ``<table>`` elements and only for ``<tr>`` |
| | and ``<th>`` rows and ``<td>`` elements within each ``<tr>`` or ``<th>`` |
| | element in the table. ``<td>`` stands for "table data". This function |
| | attempts to properly handle ``colspan`` and ``rowspan`` attributes. |
| | If the function has a ``<thead>`` argument, it is used to construct |
| | the header, otherwise the function attempts to find the header within |
| | the body (by putting rows with only ``<th>`` elements into the header). |
| | |
| | Similar to :func:`~read_csv` the `header` argument is applied |
| | **after** `skiprows` is applied. |
| | |
| | This function will *always* return a list of :class:`DataFrame` *or* |
| | it will fail, e.g., it will *not* return an empty list. |
| | |
| | Examples |
| | -------- |
| | See the :ref:`read_html documentation in the IO section of the docs |
| | <io.read_html>` for some examples of reading in HTML tables. |
| | """ |
| | |
| | |
| | if isinstance(skiprows, numbers.Integral) and skiprows < 0: |
| | raise ValueError( |
| | "cannot skip rows starting from the end of the " |
| | "data (you passed a negative value)" |
| | ) |
| | if extract_links not in [None, "header", "footer", "body", "all"]: |
| | raise ValueError( |
| | "`extract_links` must be one of " |
| | '{None, "header", "footer", "body", "all"}, got ' |
| | f'"{extract_links}"' |
| | ) |
| |
|
| | validate_header_arg(header) |
| | check_dtype_backend(dtype_backend) |
| |
|
| | io = stringify_path(io) |
| |
|
| | if isinstance(io, str) and not any( |
| | [ |
| | is_file_like(io), |
| | file_exists(io), |
| | is_url(io), |
| | is_fsspec_url(io), |
| | ] |
| | ): |
| | warnings.warn( |
| | "Passing literal html to 'read_html' is deprecated and " |
| | "will be removed in a future version. To read from a " |
| | "literal string, wrap it in a 'StringIO' object.", |
| | FutureWarning, |
| | stacklevel=find_stack_level(), |
| | ) |
| |
|
| | return _parse( |
| | flavor=flavor, |
| | io=io, |
| | match=match, |
| | header=header, |
| | index_col=index_col, |
| | skiprows=skiprows, |
| | parse_dates=parse_dates, |
| | thousands=thousands, |
| | attrs=attrs, |
| | encoding=encoding, |
| | decimal=decimal, |
| | converters=converters, |
| | na_values=na_values, |
| | keep_default_na=keep_default_na, |
| | displayed_only=displayed_only, |
| | extract_links=extract_links, |
| | dtype_backend=dtype_backend, |
| | storage_options=storage_options, |
| | ) |
| |
|