| """ parquet compat """ |
| from __future__ import annotations |
|
|
| import io |
| import json |
| import os |
| from typing import ( |
| TYPE_CHECKING, |
| Any, |
| Literal, |
| ) |
| import warnings |
| from warnings import catch_warnings |
|
|
| from pandas._config import using_pyarrow_string_dtype |
| from pandas._config.config import _get_option |
|
|
| from pandas._libs import lib |
| from pandas.compat._optional import import_optional_dependency |
| from pandas.errors import AbstractMethodError |
| from pandas.util._decorators import doc |
| from pandas.util._exceptions import find_stack_level |
| from pandas.util._validators import check_dtype_backend |
|
|
| import pandas as pd |
| from pandas import ( |
| DataFrame, |
| get_option, |
| ) |
| from pandas.core.shared_docs import _shared_docs |
|
|
| from pandas.io._util import arrow_string_types_mapper |
| from pandas.io.common import ( |
| IOHandles, |
| get_handle, |
| is_fsspec_url, |
| is_url, |
| stringify_path, |
| ) |
|
|
| if TYPE_CHECKING: |
| from pandas._typing import ( |
| DtypeBackend, |
| FilePath, |
| ReadBuffer, |
| StorageOptions, |
| WriteBuffer, |
| ) |
|
|
|
|
| def get_engine(engine: str) -> BaseImpl: |
| """return our implementation""" |
| if engine == "auto": |
| engine = get_option("io.parquet.engine") |
|
|
| if engine == "auto": |
| |
| engine_classes = [PyArrowImpl, FastParquetImpl] |
|
|
| error_msgs = "" |
| for engine_class in engine_classes: |
| try: |
| return engine_class() |
| except ImportError as err: |
| error_msgs += "\n - " + str(err) |
|
|
| raise ImportError( |
| "Unable to find a usable engine; " |
| "tried using: 'pyarrow', 'fastparquet'.\n" |
| "A suitable version of " |
| "pyarrow or fastparquet is required for parquet " |
| "support.\n" |
| "Trying to import the above resulted in these errors:" |
| f"{error_msgs}" |
| ) |
|
|
| if engine == "pyarrow": |
| return PyArrowImpl() |
| elif engine == "fastparquet": |
| return FastParquetImpl() |
|
|
| raise ValueError("engine must be one of 'pyarrow', 'fastparquet'") |
|
|
|
|
| def _get_path_or_handle( |
| path: FilePath | ReadBuffer[bytes] | WriteBuffer[bytes], |
| fs: Any, |
| storage_options: StorageOptions | None = None, |
| mode: str = "rb", |
| is_dir: bool = False, |
| ) -> tuple[ |
| FilePath | ReadBuffer[bytes] | WriteBuffer[bytes], IOHandles[bytes] | None, Any |
| ]: |
| """File handling for PyArrow.""" |
| path_or_handle = stringify_path(path) |
| if fs is not None: |
| pa_fs = import_optional_dependency("pyarrow.fs", errors="ignore") |
| fsspec = import_optional_dependency("fsspec", errors="ignore") |
| if pa_fs is not None and isinstance(fs, pa_fs.FileSystem): |
| if storage_options: |
| raise NotImplementedError( |
| "storage_options not supported with a pyarrow FileSystem." |
| ) |
| elif fsspec is not None and isinstance(fs, fsspec.spec.AbstractFileSystem): |
| pass |
| else: |
| raise ValueError( |
| f"filesystem must be a pyarrow or fsspec FileSystem, " |
| f"not a {type(fs).__name__}" |
| ) |
| if is_fsspec_url(path_or_handle) and fs is None: |
| if storage_options is None: |
| pa = import_optional_dependency("pyarrow") |
| pa_fs = import_optional_dependency("pyarrow.fs") |
|
|
| try: |
| fs, path_or_handle = pa_fs.FileSystem.from_uri(path) |
| except (TypeError, pa.ArrowInvalid): |
| pass |
| if fs is None: |
| fsspec = import_optional_dependency("fsspec") |
| fs, path_or_handle = fsspec.core.url_to_fs( |
| path_or_handle, **(storage_options or {}) |
| ) |
| elif storage_options and (not is_url(path_or_handle) or mode != "rb"): |
| |
| |
| raise ValueError("storage_options passed with buffer, or non-supported URL") |
|
|
| handles = None |
| if ( |
| not fs |
| and not is_dir |
| and isinstance(path_or_handle, str) |
| and not os.path.isdir(path_or_handle) |
| ): |
| |
| |
| |
| handles = get_handle( |
| path_or_handle, mode, is_text=False, storage_options=storage_options |
| ) |
| fs = None |
| path_or_handle = handles.handle |
| return path_or_handle, handles, fs |
|
|
|
|
| class BaseImpl: |
| @staticmethod |
| def validate_dataframe(df: DataFrame) -> None: |
| if not isinstance(df, DataFrame): |
| raise ValueError("to_parquet only supports IO with DataFrames") |
|
|
| def write(self, df: DataFrame, path, compression, **kwargs): |
| raise AbstractMethodError(self) |
|
|
| def read(self, path, columns=None, **kwargs) -> DataFrame: |
| raise AbstractMethodError(self) |
|
|
|
|
| class PyArrowImpl(BaseImpl): |
| def __init__(self) -> None: |
| import_optional_dependency( |
| "pyarrow", extra="pyarrow is required for parquet support." |
| ) |
| import pyarrow.parquet |
|
|
| |
| import pandas.core.arrays.arrow.extension_types |
|
|
| self.api = pyarrow |
|
|
| def write( |
| self, |
| df: DataFrame, |
| path: FilePath | WriteBuffer[bytes], |
| compression: str | None = "snappy", |
| index: bool | None = None, |
| storage_options: StorageOptions | None = None, |
| partition_cols: list[str] | None = None, |
| filesystem=None, |
| **kwargs, |
| ) -> None: |
| self.validate_dataframe(df) |
|
|
| from_pandas_kwargs: dict[str, Any] = {"schema": kwargs.pop("schema", None)} |
| if index is not None: |
| from_pandas_kwargs["preserve_index"] = index |
|
|
| table = self.api.Table.from_pandas(df, **from_pandas_kwargs) |
|
|
| if df.attrs: |
| df_metadata = {"PANDAS_ATTRS": json.dumps(df.attrs)} |
| existing_metadata = table.schema.metadata |
| merged_metadata = {**existing_metadata, **df_metadata} |
| table = table.replace_schema_metadata(merged_metadata) |
|
|
| path_or_handle, handles, filesystem = _get_path_or_handle( |
| path, |
| filesystem, |
| storage_options=storage_options, |
| mode="wb", |
| is_dir=partition_cols is not None, |
| ) |
| if ( |
| isinstance(path_or_handle, io.BufferedWriter) |
| and hasattr(path_or_handle, "name") |
| and isinstance(path_or_handle.name, (str, bytes)) |
| ): |
| if isinstance(path_or_handle.name, bytes): |
| path_or_handle = path_or_handle.name.decode() |
| else: |
| path_or_handle = path_or_handle.name |
|
|
| try: |
| if partition_cols is not None: |
| |
| self.api.parquet.write_to_dataset( |
| table, |
| path_or_handle, |
| compression=compression, |
| partition_cols=partition_cols, |
| filesystem=filesystem, |
| **kwargs, |
| ) |
| else: |
| |
| self.api.parquet.write_table( |
| table, |
| path_or_handle, |
| compression=compression, |
| filesystem=filesystem, |
| **kwargs, |
| ) |
| finally: |
| if handles is not None: |
| handles.close() |
|
|
| def read( |
| self, |
| path, |
| columns=None, |
| filters=None, |
| use_nullable_dtypes: bool = False, |
| dtype_backend: DtypeBackend | lib.NoDefault = lib.no_default, |
| storage_options: StorageOptions | None = None, |
| filesystem=None, |
| **kwargs, |
| ) -> DataFrame: |
| kwargs["use_pandas_metadata"] = True |
|
|
| to_pandas_kwargs = {} |
| if dtype_backend == "numpy_nullable": |
| from pandas.io._util import _arrow_dtype_mapping |
|
|
| mapping = _arrow_dtype_mapping() |
| to_pandas_kwargs["types_mapper"] = mapping.get |
| elif dtype_backend == "pyarrow": |
| to_pandas_kwargs["types_mapper"] = pd.ArrowDtype |
| elif using_pyarrow_string_dtype(): |
| to_pandas_kwargs["types_mapper"] = arrow_string_types_mapper() |
|
|
| manager = _get_option("mode.data_manager", silent=True) |
| if manager == "array": |
| to_pandas_kwargs["split_blocks"] = True |
|
|
| path_or_handle, handles, filesystem = _get_path_or_handle( |
| path, |
| filesystem, |
| storage_options=storage_options, |
| mode="rb", |
| ) |
| try: |
| pa_table = self.api.parquet.read_table( |
| path_or_handle, |
| columns=columns, |
| filesystem=filesystem, |
| filters=filters, |
| **kwargs, |
| ) |
| result = pa_table.to_pandas(**to_pandas_kwargs) |
|
|
| if manager == "array": |
| result = result._as_manager("array", copy=False) |
|
|
| if pa_table.schema.metadata: |
| if b"PANDAS_ATTRS" in pa_table.schema.metadata: |
| df_metadata = pa_table.schema.metadata[b"PANDAS_ATTRS"] |
| result.attrs = json.loads(df_metadata) |
| return result |
| finally: |
| if handles is not None: |
| handles.close() |
|
|
|
|
| class FastParquetImpl(BaseImpl): |
| def __init__(self) -> None: |
| |
| |
| fastparquet = import_optional_dependency( |
| "fastparquet", extra="fastparquet is required for parquet support." |
| ) |
| self.api = fastparquet |
|
|
| def write( |
| self, |
| df: DataFrame, |
| path, |
| compression: Literal["snappy", "gzip", "brotli"] | None = "snappy", |
| index=None, |
| partition_cols=None, |
| storage_options: StorageOptions | None = None, |
| filesystem=None, |
| **kwargs, |
| ) -> None: |
| self.validate_dataframe(df) |
|
|
| if "partition_on" in kwargs and partition_cols is not None: |
| raise ValueError( |
| "Cannot use both partition_on and " |
| "partition_cols. Use partition_cols for partitioning data" |
| ) |
| if "partition_on" in kwargs: |
| partition_cols = kwargs.pop("partition_on") |
|
|
| if partition_cols is not None: |
| kwargs["file_scheme"] = "hive" |
|
|
| if filesystem is not None: |
| raise NotImplementedError( |
| "filesystem is not implemented for the fastparquet engine." |
| ) |
|
|
| |
| path = stringify_path(path) |
| if is_fsspec_url(path): |
| fsspec = import_optional_dependency("fsspec") |
|
|
| |
| kwargs["open_with"] = lambda path, _: fsspec.open( |
| path, "wb", **(storage_options or {}) |
| ).open() |
| elif storage_options: |
| raise ValueError( |
| "storage_options passed with file object or non-fsspec file path" |
| ) |
|
|
| with catch_warnings(record=True): |
| self.api.write( |
| path, |
| df, |
| compression=compression, |
| write_index=index, |
| partition_on=partition_cols, |
| **kwargs, |
| ) |
|
|
| def read( |
| self, |
| path, |
| columns=None, |
| filters=None, |
| storage_options: StorageOptions | None = None, |
| filesystem=None, |
| **kwargs, |
| ) -> DataFrame: |
| parquet_kwargs: dict[str, Any] = {} |
| use_nullable_dtypes = kwargs.pop("use_nullable_dtypes", False) |
| dtype_backend = kwargs.pop("dtype_backend", lib.no_default) |
| |
| parquet_kwargs["pandas_nulls"] = False |
| if use_nullable_dtypes: |
| raise ValueError( |
| "The 'use_nullable_dtypes' argument is not supported for the " |
| "fastparquet engine" |
| ) |
| if dtype_backend is not lib.no_default: |
| raise ValueError( |
| "The 'dtype_backend' argument is not supported for the " |
| "fastparquet engine" |
| ) |
| if filesystem is not None: |
| raise NotImplementedError( |
| "filesystem is not implemented for the fastparquet engine." |
| ) |
| path = stringify_path(path) |
| handles = None |
| if is_fsspec_url(path): |
| fsspec = import_optional_dependency("fsspec") |
|
|
| parquet_kwargs["fs"] = fsspec.open(path, "rb", **(storage_options or {})).fs |
| elif isinstance(path, str) and not os.path.isdir(path): |
| |
| |
| |
| handles = get_handle( |
| path, "rb", is_text=False, storage_options=storage_options |
| ) |
| path = handles.handle |
|
|
| try: |
| parquet_file = self.api.ParquetFile(path, **parquet_kwargs) |
| return parquet_file.to_pandas(columns=columns, filters=filters, **kwargs) |
| finally: |
| if handles is not None: |
| handles.close() |
|
|
|
|
| @doc(storage_options=_shared_docs["storage_options"]) |
| def to_parquet( |
| df: DataFrame, |
| path: FilePath | WriteBuffer[bytes] | None = None, |
| engine: str = "auto", |
| compression: str | None = "snappy", |
| index: bool | None = None, |
| storage_options: StorageOptions | None = None, |
| partition_cols: list[str] | None = None, |
| filesystem: Any = None, |
| **kwargs, |
| ) -> bytes | None: |
| """ |
| Write a DataFrame to the parquet format. |
| |
| Parameters |
| ---------- |
| df : DataFrame |
| path : str, path object, file-like object, or None, default None |
| String, path object (implementing ``os.PathLike[str]``), or file-like |
| object implementing a binary ``write()`` function. If None, the result is |
| returned as bytes. If a string, it will be used as Root Directory path |
| when writing a partitioned dataset. The engine fastparquet does not |
| accept file-like objects. |
| engine : {{'auto', 'pyarrow', 'fastparquet'}}, default 'auto' |
| Parquet library to use. If 'auto', then the option |
| ``io.parquet.engine`` is used. The default ``io.parquet.engine`` |
| behavior is to try 'pyarrow', falling back to 'fastparquet' if |
| 'pyarrow' is unavailable. |
| |
| When using the ``'pyarrow'`` engine and no storage options are provided |
| and a filesystem is implemented by both ``pyarrow.fs`` and ``fsspec`` |
| (e.g. "s3://"), then the ``pyarrow.fs`` filesystem is attempted first. |
| Use the filesystem keyword with an instantiated fsspec filesystem |
| if you wish to use its implementation. |
| compression : {{'snappy', 'gzip', 'brotli', 'lz4', 'zstd', None}}, |
| default 'snappy'. Name of the compression to use. Use ``None`` |
| for no compression. |
| index : bool, default None |
| If ``True``, include the dataframe's index(es) in the file output. If |
| ``False``, they will not be written to the file. |
| If ``None``, similar to ``True`` the dataframe's index(es) |
| will be saved. However, instead of being saved as values, |
| the RangeIndex will be stored as a range in the metadata so it |
| doesn't require much space and is faster. Other indexes will |
| be included as columns in the file output. |
| partition_cols : str or list, optional, default None |
| Column names by which to partition the dataset. |
| Columns are partitioned in the order they are given. |
| Must be None if path is not a string. |
| {storage_options} |
| |
| filesystem : fsspec or pyarrow filesystem, default None |
| Filesystem object to use when reading the parquet file. Only implemented |
| for ``engine="pyarrow"``. |
| |
| .. versionadded:: 2.1.0 |
| |
| kwargs |
| Additional keyword arguments passed to the engine |
| |
| Returns |
| ------- |
| bytes if no path argument is provided else None |
| """ |
| if isinstance(partition_cols, str): |
| partition_cols = [partition_cols] |
| impl = get_engine(engine) |
|
|
| path_or_buf: FilePath | WriteBuffer[bytes] = io.BytesIO() if path is None else path |
|
|
| impl.write( |
| df, |
| path_or_buf, |
| compression=compression, |
| index=index, |
| partition_cols=partition_cols, |
| storage_options=storage_options, |
| filesystem=filesystem, |
| **kwargs, |
| ) |
|
|
| if path is None: |
| assert isinstance(path_or_buf, io.BytesIO) |
| return path_or_buf.getvalue() |
| else: |
| return None |
|
|
|
|
| @doc(storage_options=_shared_docs["storage_options"]) |
| def read_parquet( |
| path: FilePath | ReadBuffer[bytes], |
| engine: str = "auto", |
| columns: list[str] | None = None, |
| storage_options: StorageOptions | None = None, |
| use_nullable_dtypes: bool | lib.NoDefault = lib.no_default, |
| dtype_backend: DtypeBackend | lib.NoDefault = lib.no_default, |
| filesystem: Any = None, |
| filters: list[tuple] | list[list[tuple]] | None = None, |
| **kwargs, |
| ) -> DataFrame: |
| """ |
| Load a parquet object from the file path, returning a DataFrame. |
| |
| Parameters |
| ---------- |
| path : str, path object or file-like object |
| String, path object (implementing ``os.PathLike[str]``), or file-like |
| object implementing a binary ``read()`` function. |
| The string could be a URL. Valid URL schemes include http, ftp, s3, |
| gs, and file. For file URLs, a host is expected. A local file could be: |
| ``file://localhost/path/to/table.parquet``. |
| A file URL can also be a path to a directory that contains multiple |
| partitioned parquet files. Both pyarrow and fastparquet support |
| paths to directories as well as file URLs. A directory path could be: |
| ``file://localhost/path/to/tables`` or ``s3://bucket/partition_dir``. |
| engine : {{'auto', 'pyarrow', 'fastparquet'}}, default 'auto' |
| Parquet library to use. If 'auto', then the option |
| ``io.parquet.engine`` is used. The default ``io.parquet.engine`` |
| behavior is to try 'pyarrow', falling back to 'fastparquet' if |
| 'pyarrow' is unavailable. |
| |
| When using the ``'pyarrow'`` engine and no storage options are provided |
| and a filesystem is implemented by both ``pyarrow.fs`` and ``fsspec`` |
| (e.g. "s3://"), then the ``pyarrow.fs`` filesystem is attempted first. |
| Use the filesystem keyword with an instantiated fsspec filesystem |
| if you wish to use its implementation. |
| columns : list, default=None |
| If not None, only these columns will be read from the file. |
| {storage_options} |
| |
| .. versionadded:: 1.3.0 |
| |
| use_nullable_dtypes : bool, default False |
| If True, use dtypes that use ``pd.NA`` as missing value indicator |
| for the resulting DataFrame. (only applicable for the ``pyarrow`` |
| engine) |
| As new dtypes are added that support ``pd.NA`` in the future, the |
| output with this option will change to use those dtypes. |
| Note: this is an experimental option, and behaviour (e.g. additional |
| support dtypes) may change without notice. |
| |
| .. deprecated:: 2.0 |
| |
| dtype_backend : {{'numpy_nullable', 'pyarrow'}}, default 'numpy_nullable' |
| Back-end data type applied to the resultant :class:`DataFrame` |
| (still experimental). Behaviour is as follows: |
| |
| * ``"numpy_nullable"``: returns nullable-dtype-backed :class:`DataFrame` |
| (default). |
| * ``"pyarrow"``: returns pyarrow-backed nullable :class:`ArrowDtype` |
| DataFrame. |
| |
| .. versionadded:: 2.0 |
| |
| filesystem : fsspec or pyarrow filesystem, default None |
| Filesystem object to use when reading the parquet file. Only implemented |
| for ``engine="pyarrow"``. |
| |
| .. versionadded:: 2.1.0 |
| |
| filters : List[Tuple] or List[List[Tuple]], default None |
| To filter out data. |
| Filter syntax: [[(column, op, val), ...],...] |
| where op is [==, =, >, >=, <, <=, !=, in, not in] |
| The innermost tuples are transposed into a set of filters applied |
| through an `AND` operation. |
| The outer list combines these sets of filters through an `OR` |
| operation. |
| A single list of tuples can also be used, meaning that no `OR` |
| operation between set of filters is to be conducted. |
| |
| Using this argument will NOT result in row-wise filtering of the final |
| partitions unless ``engine="pyarrow"`` is also specified. For |
| other engines, filtering is only performed at the partition level, that is, |
| to prevent the loading of some row-groups and/or files. |
| |
| .. versionadded:: 2.1.0 |
| |
| **kwargs |
| Any additional kwargs are passed to the engine. |
| |
| Returns |
| ------- |
| DataFrame |
| |
| See Also |
| -------- |
| DataFrame.to_parquet : Create a parquet object that serializes a DataFrame. |
| |
| Examples |
| -------- |
| >>> original_df = pd.DataFrame( |
| ... {{"foo": range(5), "bar": range(5, 10)}} |
| ... ) |
| >>> original_df |
| foo bar |
| 0 0 5 |
| 1 1 6 |
| 2 2 7 |
| 3 3 8 |
| 4 4 9 |
| >>> df_parquet_bytes = original_df.to_parquet() |
| >>> from io import BytesIO |
| >>> restored_df = pd.read_parquet(BytesIO(df_parquet_bytes)) |
| >>> restored_df |
| foo bar |
| 0 0 5 |
| 1 1 6 |
| 2 2 7 |
| 3 3 8 |
| 4 4 9 |
| >>> restored_df.equals(original_df) |
| True |
| >>> restored_bar = pd.read_parquet(BytesIO(df_parquet_bytes), columns=["bar"]) |
| >>> restored_bar |
| bar |
| 0 5 |
| 1 6 |
| 2 7 |
| 3 8 |
| 4 9 |
| >>> restored_bar.equals(original_df[['bar']]) |
| True |
| |
| The function uses `kwargs` that are passed directly to the engine. |
| In the following example, we use the `filters` argument of the pyarrow |
| engine to filter the rows of the DataFrame. |
| |
| Since `pyarrow` is the default engine, we can omit the `engine` argument. |
| Note that the `filters` argument is implemented by the `pyarrow` engine, |
| which can benefit from multithreading and also potentially be more |
| economical in terms of memory. |
| |
| >>> sel = [("foo", ">", 2)] |
| >>> restored_part = pd.read_parquet(BytesIO(df_parquet_bytes), filters=sel) |
| >>> restored_part |
| foo bar |
| 0 3 8 |
| 1 4 9 |
| """ |
|
|
| impl = get_engine(engine) |
|
|
| if use_nullable_dtypes is not lib.no_default: |
| msg = ( |
| "The argument 'use_nullable_dtypes' is deprecated and will be removed " |
| "in a future version." |
| ) |
| if use_nullable_dtypes is True: |
| msg += ( |
| "Use dtype_backend='numpy_nullable' instead of use_nullable_dtype=True." |
| ) |
| warnings.warn(msg, FutureWarning, stacklevel=find_stack_level()) |
| else: |
| use_nullable_dtypes = False |
| check_dtype_backend(dtype_backend) |
|
|
| return impl.read( |
| path, |
| columns=columns, |
| filters=filters, |
| storage_options=storage_options, |
| use_nullable_dtypes=use_nullable_dtypes, |
| dtype_backend=dtype_backend, |
| filesystem=filesystem, |
| **kwargs, |
| ) |
|
|