Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes. Β
See raw diff
- .gitattributes +1 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/__init__.py +70 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/arrow_dataset.py +0 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/arrow_reader.py +661 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/arrow_writer.py +745 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/builder.py +0 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/config.py +259 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/data_files.py +806 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/dataset_dict.py +0 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/distributed.py +39 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/exceptions.py +85 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/filesystems/__init__.py +86 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/filesystems/__pycache__/__init__.cpython-310.pyc +0 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/filesystems/__pycache__/s3filesystem.cpython-310.pyc +0 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/filesystems/compression.py +178 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/filesystems/s3filesystem.py +116 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/info.py +592 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/inspect.py +581 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/io/__pycache__/csv.cpython-310.pyc +0 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/io/__pycache__/generator.cpython-310.pyc +0 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/io/__pycache__/json.cpython-310.pyc +0 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/io/__pycache__/parquet.cpython-310.pyc +0 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/io/__pycache__/sql.cpython-310.pyc +0 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/io/__pycache__/text.cpython-310.pyc +0 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/io/abc.py +53 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/io/generator.py +58 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/io/json.py +166 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/io/parquet.py +156 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/io/spark.py +57 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/io/sql.py +125 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/io/text.py +61 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/iterable_dataset.py +0 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/keyhash.py +104 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/metric.py +652 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/naming.py +84 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/search.py +779 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/splits.py +635 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/streaming.py +140 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/table.py +2360 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/utils/_datasets_server.py +96 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/utils/download_manager.py +1 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/utils/logging.py +179 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/utils/metadata.py +320 -0
- evalkit_tf437/lib/python3.10/site-packages/datasets/utils/track.py +49 -0
- evalkit_tf437/lib/python3.10/site-packages/fontTools/ttLib/tables/__pycache__/otData.cpython-310.pyc +3 -0
- evalkit_tf437/lib/python3.10/site-packages/latex2mathml-3.77.0.dist-info/INSTALLER +1 -0
- evalkit_tf437/lib/python3.10/site-packages/latex2mathml-3.77.0.dist-info/LICENSE +21 -0
- evalkit_tf437/lib/python3.10/site-packages/latex2mathml-3.77.0.dist-info/METADATA +172 -0
- evalkit_tf437/lib/python3.10/site-packages/latex2mathml-3.77.0.dist-info/RECORD +25 -0
- evalkit_tf437/lib/python3.10/site-packages/latex2mathml-3.77.0.dist-info/REQUESTED +0 -0
.gitattributes
CHANGED
|
@@ -330,3 +330,4 @@ evalkit_tf437/lib/python3.10/site-packages/opencv_python.libs/libxcb-xkb-9ba31ab
|
|
| 330 |
evalkit_tf437/lib/python3.10/site-packages/opencv_python.libs/libxkbcommon-71ae2972.so.0.0.0 filter=lfs diff=lfs merge=lfs -text
|
| 331 |
evalkit_tf437/lib/python3.10/site-packages/regex/_regex.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
|
| 332 |
evalkit_tf437/lib/python3.10/site-packages/scikit_learn.libs/libgomp-a34b3233.so.1.0.0 filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 330 |
evalkit_tf437/lib/python3.10/site-packages/opencv_python.libs/libxkbcommon-71ae2972.so.0.0.0 filter=lfs diff=lfs merge=lfs -text
|
| 331 |
evalkit_tf437/lib/python3.10/site-packages/regex/_regex.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
|
| 332 |
evalkit_tf437/lib/python3.10/site-packages/scikit_learn.libs/libgomp-a34b3233.so.1.0.0 filter=lfs diff=lfs merge=lfs -text
|
| 333 |
+
evalkit_tf437/lib/python3.10/site-packages/fontTools/ttLib/tables/__pycache__/otData.cpython-310.pyc filter=lfs diff=lfs merge=lfs -text
|
evalkit_tf437/lib/python3.10/site-packages/datasets/__init__.py
ADDED
|
@@ -0,0 +1,70 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# ruff: noqa
|
| 2 |
+
# Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors.
|
| 3 |
+
#
|
| 4 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
| 5 |
+
# you may not use this file except in compliance with the License.
|
| 6 |
+
# You may obtain a copy of the License at
|
| 7 |
+
#
|
| 8 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
| 9 |
+
#
|
| 10 |
+
# Unless required by applicable law or agreed to in writing, software
|
| 11 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
| 12 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| 13 |
+
# See the License for the specific language governing permissions and
|
| 14 |
+
# limitations under the License.
|
| 15 |
+
|
| 16 |
+
__version__ = "2.18.0"
|
| 17 |
+
|
| 18 |
+
from .arrow_dataset import Dataset
|
| 19 |
+
from .arrow_reader import ReadInstruction
|
| 20 |
+
from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder
|
| 21 |
+
from .combine import concatenate_datasets, interleave_datasets
|
| 22 |
+
from .dataset_dict import DatasetDict, IterableDatasetDict
|
| 23 |
+
from .download import *
|
| 24 |
+
from .features import *
|
| 25 |
+
from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled
|
| 26 |
+
from .info import DatasetInfo, MetricInfo
|
| 27 |
+
from .inspect import (
|
| 28 |
+
get_dataset_config_info,
|
| 29 |
+
get_dataset_config_names,
|
| 30 |
+
get_dataset_default_config_name,
|
| 31 |
+
get_dataset_infos,
|
| 32 |
+
get_dataset_split_names,
|
| 33 |
+
inspect_dataset,
|
| 34 |
+
inspect_metric,
|
| 35 |
+
list_datasets,
|
| 36 |
+
list_metrics,
|
| 37 |
+
)
|
| 38 |
+
from .iterable_dataset import IterableDataset
|
| 39 |
+
from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric
|
| 40 |
+
from .metric import Metric
|
| 41 |
+
from .splits import (
|
| 42 |
+
NamedSplit,
|
| 43 |
+
NamedSplitAll,
|
| 44 |
+
Split,
|
| 45 |
+
SplitBase,
|
| 46 |
+
SplitDict,
|
| 47 |
+
SplitGenerator,
|
| 48 |
+
SplitInfo,
|
| 49 |
+
SubSplitInfo,
|
| 50 |
+
percent,
|
| 51 |
+
)
|
| 52 |
+
from .tasks import *
|
| 53 |
+
from .utils import *
|
| 54 |
+
from .utils import logging
|
| 55 |
+
|
| 56 |
+
|
| 57 |
+
# deprecated modules
|
| 58 |
+
from datasets import arrow_dataset as _arrow_dataset # isort:skip
|
| 59 |
+
from datasets import utils as _utils # isort:skip
|
| 60 |
+
from datasets.utils import download_manager as _deprecated_download_manager # isort:skip
|
| 61 |
+
|
| 62 |
+
_arrow_dataset.concatenate_datasets = concatenate_datasets
|
| 63 |
+
_utils.DownloadConfig = DownloadConfig
|
| 64 |
+
_utils.DownloadManager = DownloadManager
|
| 65 |
+
_utils.DownloadMode = DownloadMode
|
| 66 |
+
_deprecated_download_manager.DownloadConfig = DownloadConfig
|
| 67 |
+
_deprecated_download_manager.DownloadMode = DownloadMode
|
| 68 |
+
_deprecated_download_manager.DownloadManager = DownloadManager
|
| 69 |
+
|
| 70 |
+
del _arrow_dataset, _utils, _deprecated_download_manager
|
evalkit_tf437/lib/python3.10/site-packages/datasets/arrow_dataset.py
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
evalkit_tf437/lib/python3.10/site-packages/datasets/arrow_reader.py
ADDED
|
@@ -0,0 +1,661 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors.
|
| 2 |
+
#
|
| 3 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
| 4 |
+
# you may not use this file except in compliance with the License.
|
| 5 |
+
# You may obtain a copy of the License at
|
| 6 |
+
#
|
| 7 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
| 8 |
+
#
|
| 9 |
+
# Unless required by applicable law or agreed to in writing, software
|
| 10 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
| 11 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| 12 |
+
# See the License for the specific language governing permissions and
|
| 13 |
+
# limitations under the License.
|
| 14 |
+
|
| 15 |
+
# Lint as: python3
|
| 16 |
+
"""Arrow ArrowReader."""
|
| 17 |
+
|
| 18 |
+
import copy
|
| 19 |
+
import math
|
| 20 |
+
import os
|
| 21 |
+
import re
|
| 22 |
+
import shutil
|
| 23 |
+
from dataclasses import dataclass
|
| 24 |
+
from functools import partial
|
| 25 |
+
from pathlib import Path
|
| 26 |
+
from typing import TYPE_CHECKING, List, Optional, Union
|
| 27 |
+
|
| 28 |
+
import pyarrow as pa
|
| 29 |
+
import pyarrow.parquet as pq
|
| 30 |
+
from tqdm.contrib.concurrent import thread_map
|
| 31 |
+
|
| 32 |
+
from .download.download_config import DownloadConfig
|
| 33 |
+
from .naming import _split_re, filenames_for_dataset_split
|
| 34 |
+
from .table import InMemoryTable, MemoryMappedTable, Table, concat_tables
|
| 35 |
+
from .utils import logging
|
| 36 |
+
from .utils import tqdm as hf_tqdm
|
| 37 |
+
from .utils.file_utils import cached_path
|
| 38 |
+
|
| 39 |
+
|
| 40 |
+
if TYPE_CHECKING:
|
| 41 |
+
from .info import DatasetInfo # noqa: F401
|
| 42 |
+
from .splits import Split, SplitInfo # noqa: F401
|
| 43 |
+
|
| 44 |
+
|
| 45 |
+
logger = logging.get_logger(__name__)
|
| 46 |
+
|
| 47 |
+
HF_GCP_BASE_URL = "https://storage.googleapis.com/huggingface-nlp/cache/datasets"
|
| 48 |
+
|
| 49 |
+
_SUB_SPEC_RE = re.compile(
|
| 50 |
+
rf"""
|
| 51 |
+
^
|
| 52 |
+
(?P<split>{_split_re[1:-1]})
|
| 53 |
+
(\[
|
| 54 |
+
((?P<from>-?\d+)
|
| 55 |
+
(?P<from_pct>%)?)?
|
| 56 |
+
:
|
| 57 |
+
((?P<to>-?\d+)
|
| 58 |
+
(?P<to_pct>%)?)?
|
| 59 |
+
\])?(\((?P<rounding>[^\)]*)\))?
|
| 60 |
+
$
|
| 61 |
+
""", # remove ^ and $
|
| 62 |
+
re.X,
|
| 63 |
+
)
|
| 64 |
+
|
| 65 |
+
_ADDITION_SEP_RE = re.compile(r"\s*\+\s*")
|
| 66 |
+
|
| 67 |
+
|
| 68 |
+
class DatasetNotOnHfGcsError(ConnectionError):
|
| 69 |
+
"""When you can't get the dataset from the Hf google cloud storage"""
|
| 70 |
+
|
| 71 |
+
pass
|
| 72 |
+
|
| 73 |
+
|
| 74 |
+
class MissingFilesOnHfGcsError(ConnectionError):
|
| 75 |
+
"""When some files are missing on the Hf oogle cloud storage"""
|
| 76 |
+
|
| 77 |
+
pass
|
| 78 |
+
|
| 79 |
+
|
| 80 |
+
@dataclass(frozen=True)
|
| 81 |
+
class FileInstructions:
|
| 82 |
+
"""The file instructions associated with a split ReadInstruction.
|
| 83 |
+
|
| 84 |
+
Attributes:
|
| 85 |
+
num_examples: `int`, The total number of examples
|
| 86 |
+
file_instructions: List[dict(filename, skip, take)], the files information.
|
| 87 |
+
The filenames contains the relative path, not absolute.
|
| 88 |
+
skip/take indicates which example read in the file: `ds.slice(skip, take)`
|
| 89 |
+
"""
|
| 90 |
+
|
| 91 |
+
num_examples: int
|
| 92 |
+
file_instructions: List[dict]
|
| 93 |
+
|
| 94 |
+
|
| 95 |
+
def make_file_instructions(
|
| 96 |
+
name: str,
|
| 97 |
+
split_infos: List["SplitInfo"],
|
| 98 |
+
instruction: Union[str, "ReadInstruction"],
|
| 99 |
+
filetype_suffix: Optional[str] = None,
|
| 100 |
+
prefix_path: Optional[str] = None,
|
| 101 |
+
) -> FileInstructions:
|
| 102 |
+
"""Returns instructions of the split dict.
|
| 103 |
+
|
| 104 |
+
Args:
|
| 105 |
+
name (`str`): Name of the dataset.
|
| 106 |
+
split_infos (`list` of `[SplitInfo]`): Dataset splits information.
|
| 107 |
+
instruction ([`ReadInstruction`] or `str`): Reading instruction for a dataset.
|
| 108 |
+
filetype_suffix (`str`, *optional*): Suffix of dataset files, e.g. 'arrow' or 'parquet'.
|
| 109 |
+
prefix_path (`str`, *optional*): Prefix of dataset files, e.g. directory name.
|
| 110 |
+
|
| 111 |
+
Returns:
|
| 112 |
+
[`FileInstructions`]
|
| 113 |
+
"""
|
| 114 |
+
if not isinstance(name, str):
|
| 115 |
+
raise TypeError(f"Expected str 'name', but got: {type(name).__name__}")
|
| 116 |
+
elif not name:
|
| 117 |
+
raise ValueError("Expected non-empty str 'name'")
|
| 118 |
+
name2len = {info.name: info.num_examples for info in split_infos}
|
| 119 |
+
name2shard_lengths = {info.name: info.shard_lengths for info in split_infos}
|
| 120 |
+
name2filenames = {
|
| 121 |
+
info.name: filenames_for_dataset_split(
|
| 122 |
+
path=prefix_path,
|
| 123 |
+
dataset_name=name,
|
| 124 |
+
split=info.name,
|
| 125 |
+
filetype_suffix=filetype_suffix,
|
| 126 |
+
shard_lengths=name2shard_lengths[info.name],
|
| 127 |
+
)
|
| 128 |
+
for info in split_infos
|
| 129 |
+
}
|
| 130 |
+
if not isinstance(instruction, ReadInstruction):
|
| 131 |
+
instruction = ReadInstruction.from_spec(instruction)
|
| 132 |
+
# Create the absolute instruction (per split)
|
| 133 |
+
absolute_instructions = instruction.to_absolute(name2len)
|
| 134 |
+
|
| 135 |
+
# For each split, return the files instruction (skip/take)
|
| 136 |
+
file_instructions = []
|
| 137 |
+
num_examples = 0
|
| 138 |
+
for abs_instr in absolute_instructions:
|
| 139 |
+
split_length = name2len[abs_instr.splitname]
|
| 140 |
+
filenames = name2filenames[abs_instr.splitname]
|
| 141 |
+
shard_lengths = name2shard_lengths[abs_instr.splitname]
|
| 142 |
+
from_ = 0 if abs_instr.from_ is None else abs_instr.from_
|
| 143 |
+
to = split_length if abs_instr.to is None else abs_instr.to
|
| 144 |
+
if shard_lengths is None: # not sharded
|
| 145 |
+
for filename in filenames:
|
| 146 |
+
take = to - from_
|
| 147 |
+
if take == 0:
|
| 148 |
+
continue
|
| 149 |
+
num_examples += take
|
| 150 |
+
file_instructions.append({"filename": filename, "skip": from_, "take": take})
|
| 151 |
+
else: # sharded
|
| 152 |
+
index_start = 0 # Beginning (included) of moving window.
|
| 153 |
+
index_end = 0 # End (excluded) of moving window.
|
| 154 |
+
for filename, shard_length in zip(filenames, shard_lengths):
|
| 155 |
+
index_end += shard_length
|
| 156 |
+
if from_ < index_end and to > index_start: # There is something to take.
|
| 157 |
+
skip = from_ - index_start if from_ > index_start else 0
|
| 158 |
+
take = to - index_start - skip if to < index_end else -1
|
| 159 |
+
if take == 0:
|
| 160 |
+
continue
|
| 161 |
+
file_instructions.append({"filename": filename, "skip": skip, "take": take})
|
| 162 |
+
num_examples += shard_length - skip if take == -1 else take
|
| 163 |
+
index_start += shard_length
|
| 164 |
+
return FileInstructions(
|
| 165 |
+
num_examples=num_examples,
|
| 166 |
+
file_instructions=file_instructions,
|
| 167 |
+
)
|
| 168 |
+
|
| 169 |
+
|
| 170 |
+
class BaseReader:
|
| 171 |
+
"""
|
| 172 |
+
Build a Dataset object out of Instruction instance(s).
|
| 173 |
+
"""
|
| 174 |
+
|
| 175 |
+
def __init__(self, path: str, info: Optional["DatasetInfo"]):
|
| 176 |
+
"""Initializes ArrowReader.
|
| 177 |
+
|
| 178 |
+
Args:
|
| 179 |
+
path (str): path where tfrecords are stored.
|
| 180 |
+
info (DatasetInfo): info about the dataset.
|
| 181 |
+
"""
|
| 182 |
+
self._path: str = path
|
| 183 |
+
self._info: Optional["DatasetInfo"] = info
|
| 184 |
+
self._filetype_suffix: Optional[str] = None
|
| 185 |
+
|
| 186 |
+
def _get_table_from_filename(self, filename_skip_take, in_memory=False) -> Table:
|
| 187 |
+
"""Returns a Dataset instance from given (filename, skip, take)."""
|
| 188 |
+
raise NotImplementedError
|
| 189 |
+
|
| 190 |
+
def _read_files(self, files, in_memory=False) -> Table:
|
| 191 |
+
"""Returns Dataset for given file instructions.
|
| 192 |
+
|
| 193 |
+
Args:
|
| 194 |
+
files: List[dict(filename, skip, take)], the files information.
|
| 195 |
+
The filenames contain the absolute path, not relative.
|
| 196 |
+
skip/take indicates which example read in the file: `ds.slice(skip, take)`
|
| 197 |
+
in_memory (bool, default False): Whether to copy the data in-memory.
|
| 198 |
+
"""
|
| 199 |
+
if len(files) == 0 or not all(isinstance(f, dict) for f in files):
|
| 200 |
+
raise ValueError("please provide valid file informations")
|
| 201 |
+
files = copy.deepcopy(files)
|
| 202 |
+
for f in files:
|
| 203 |
+
f["filename"] = os.path.join(self._path, f["filename"])
|
| 204 |
+
|
| 205 |
+
pa_tables = thread_map(
|
| 206 |
+
partial(self._get_table_from_filename, in_memory=in_memory),
|
| 207 |
+
files,
|
| 208 |
+
tqdm_class=hf_tqdm,
|
| 209 |
+
desc="Loading dataset shards",
|
| 210 |
+
# set `disable=None` rather than `disable=False` by default to disable progress bar when no TTY attached
|
| 211 |
+
disable=len(files) <= 16 or None,
|
| 212 |
+
)
|
| 213 |
+
pa_tables = [t for t in pa_tables if len(t) > 0]
|
| 214 |
+
if not pa_tables and (self._info is None or self._info.features is None):
|
| 215 |
+
raise ValueError(
|
| 216 |
+
"Tried to read an empty table. Please specify at least info.features to create an empty table with the right type."
|
| 217 |
+
)
|
| 218 |
+
pa_tables = pa_tables or [InMemoryTable.from_batches([], schema=pa.schema(self._info.features.type))]
|
| 219 |
+
pa_table = concat_tables(pa_tables) if len(pa_tables) != 1 else pa_tables[0]
|
| 220 |
+
return pa_table
|
| 221 |
+
|
| 222 |
+
def get_file_instructions(self, name, instruction, split_infos):
|
| 223 |
+
"""Return list of dict {'filename': str, 'skip': int, 'take': int}"""
|
| 224 |
+
file_instructions = make_file_instructions(
|
| 225 |
+
name, split_infos, instruction, filetype_suffix=self._filetype_suffix, prefix_path=self._path
|
| 226 |
+
)
|
| 227 |
+
files = file_instructions.file_instructions
|
| 228 |
+
return files
|
| 229 |
+
|
| 230 |
+
def read(
|
| 231 |
+
self,
|
| 232 |
+
name,
|
| 233 |
+
instructions,
|
| 234 |
+
split_infos,
|
| 235 |
+
in_memory=False,
|
| 236 |
+
):
|
| 237 |
+
"""Returns Dataset instance(s).
|
| 238 |
+
|
| 239 |
+
Args:
|
| 240 |
+
name (str): name of the dataset.
|
| 241 |
+
instructions (ReadInstruction): instructions to read.
|
| 242 |
+
Instruction can be string and will then be passed to the Instruction
|
| 243 |
+
constructor as it.
|
| 244 |
+
split_infos (list of SplitInfo proto): the available splits for dataset.
|
| 245 |
+
in_memory (bool, default False): Whether to copy the data in-memory.
|
| 246 |
+
|
| 247 |
+
Returns:
|
| 248 |
+
kwargs to build a single Dataset instance.
|
| 249 |
+
"""
|
| 250 |
+
|
| 251 |
+
files = self.get_file_instructions(name, instructions, split_infos)
|
| 252 |
+
if not files:
|
| 253 |
+
msg = f'Instruction "{instructions}" corresponds to no data!'
|
| 254 |
+
raise ValueError(msg)
|
| 255 |
+
return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory)
|
| 256 |
+
|
| 257 |
+
def read_files(
|
| 258 |
+
self,
|
| 259 |
+
files: List[dict],
|
| 260 |
+
original_instructions: Union[None, "ReadInstruction", "Split"] = None,
|
| 261 |
+
in_memory=False,
|
| 262 |
+
):
|
| 263 |
+
"""Returns single Dataset instance for the set of file instructions.
|
| 264 |
+
|
| 265 |
+
Args:
|
| 266 |
+
files: List[dict(filename, skip, take)], the files information.
|
| 267 |
+
The filenames contains the relative path, not absolute.
|
| 268 |
+
skip/take indicates which example read in the file: `ds.skip().take()`
|
| 269 |
+
original_instructions: store the original instructions used to build the dataset split in the dataset.
|
| 270 |
+
in_memory (bool, default False): Whether to copy the data in-memory.
|
| 271 |
+
|
| 272 |
+
Returns:
|
| 273 |
+
kwargs to build a Dataset instance.
|
| 274 |
+
"""
|
| 275 |
+
# Prepend path to filename
|
| 276 |
+
pa_table = self._read_files(files, in_memory=in_memory)
|
| 277 |
+
# If original_instructions is not None, convert it to a human-readable NamedSplit
|
| 278 |
+
if original_instructions is not None:
|
| 279 |
+
from .splits import Split # noqa
|
| 280 |
+
|
| 281 |
+
split = Split(str(original_instructions))
|
| 282 |
+
else:
|
| 283 |
+
split = None
|
| 284 |
+
dataset_kwargs = {"arrow_table": pa_table, "info": self._info, "split": split}
|
| 285 |
+
return dataset_kwargs
|
| 286 |
+
|
| 287 |
+
def download_from_hf_gcs(self, download_config: DownloadConfig, relative_data_dir):
|
| 288 |
+
"""
|
| 289 |
+
Download the dataset files from the Hf GCS
|
| 290 |
+
|
| 291 |
+
Args:
|
| 292 |
+
dl_cache_dir: `str`, the local cache directory used to download files
|
| 293 |
+
relative_data_dir: `str`, the relative directory of the remote files from
|
| 294 |
+
the `datasets` directory on GCS.
|
| 295 |
+
|
| 296 |
+
"""
|
| 297 |
+
remote_cache_dir = HF_GCP_BASE_URL + "/" + relative_data_dir.replace(os.sep, "/")
|
| 298 |
+
try:
|
| 299 |
+
remote_dataset_info = os.path.join(remote_cache_dir, "dataset_info.json")
|
| 300 |
+
downloaded_dataset_info = cached_path(
|
| 301 |
+
remote_dataset_info.replace(os.sep, "/"), download_config=download_config
|
| 302 |
+
)
|
| 303 |
+
shutil.move(downloaded_dataset_info, os.path.join(self._path, "dataset_info.json"))
|
| 304 |
+
if self._info is not None:
|
| 305 |
+
self._info.update(self._info.from_directory(self._path))
|
| 306 |
+
except FileNotFoundError as err:
|
| 307 |
+
raise DatasetNotOnHfGcsError(err) from None
|
| 308 |
+
try:
|
| 309 |
+
for split in self._info.splits:
|
| 310 |
+
file_instructions = self.get_file_instructions(
|
| 311 |
+
name=self._info.builder_name,
|
| 312 |
+
instruction=split,
|
| 313 |
+
split_infos=self._info.splits.values(),
|
| 314 |
+
)
|
| 315 |
+
for file_instruction in file_instructions:
|
| 316 |
+
file_to_download = str(Path(file_instruction["filename"]).relative_to(self._path))
|
| 317 |
+
remote_prepared_filename = os.path.join(remote_cache_dir, file_to_download)
|
| 318 |
+
downloaded_prepared_filename = cached_path(
|
| 319 |
+
remote_prepared_filename.replace(os.sep, "/"), download_config=download_config
|
| 320 |
+
)
|
| 321 |
+
shutil.move(downloaded_prepared_filename, file_instruction["filename"])
|
| 322 |
+
except FileNotFoundError as err:
|
| 323 |
+
raise MissingFilesOnHfGcsError(err) from None
|
| 324 |
+
|
| 325 |
+
|
| 326 |
+
class ArrowReader(BaseReader):
|
| 327 |
+
"""
|
| 328 |
+
Build a Dataset object out of Instruction instance(s).
|
| 329 |
+
This Reader uses either memory mapping or file descriptors (in-memory) on arrow files.
|
| 330 |
+
"""
|
| 331 |
+
|
| 332 |
+
def __init__(self, path: str, info: Optional["DatasetInfo"]):
|
| 333 |
+
"""Initializes ArrowReader.
|
| 334 |
+
|
| 335 |
+
Args:
|
| 336 |
+
path (str): path where Arrow files are stored.
|
| 337 |
+
info (DatasetInfo): info about the dataset.
|
| 338 |
+
"""
|
| 339 |
+
super().__init__(path, info)
|
| 340 |
+
self._filetype_suffix = "arrow"
|
| 341 |
+
|
| 342 |
+
def _get_table_from_filename(self, filename_skip_take, in_memory=False) -> Table:
|
| 343 |
+
"""Returns a Dataset instance from given (filename, skip, take)."""
|
| 344 |
+
filename, skip, take = (
|
| 345 |
+
filename_skip_take["filename"],
|
| 346 |
+
filename_skip_take["skip"] if "skip" in filename_skip_take else None,
|
| 347 |
+
filename_skip_take["take"] if "take" in filename_skip_take else None,
|
| 348 |
+
)
|
| 349 |
+
table = ArrowReader.read_table(filename, in_memory=in_memory)
|
| 350 |
+
if take == -1:
|
| 351 |
+
take = len(table) - skip
|
| 352 |
+
# here we don't want to slice an empty table, or it may segfault
|
| 353 |
+
if skip is not None and take is not None and not (skip == 0 and take == len(table)):
|
| 354 |
+
table = table.slice(skip, take)
|
| 355 |
+
return table
|
| 356 |
+
|
| 357 |
+
@staticmethod
|
| 358 |
+
def read_table(filename, in_memory=False) -> Table:
|
| 359 |
+
"""
|
| 360 |
+
Read table from file.
|
| 361 |
+
|
| 362 |
+
Args:
|
| 363 |
+
filename (str): File name of the table.
|
| 364 |
+
in_memory (bool, default=False): Whether to copy the data in-memory.
|
| 365 |
+
|
| 366 |
+
Returns:
|
| 367 |
+
pyarrow.Table
|
| 368 |
+
"""
|
| 369 |
+
table_cls = InMemoryTable if in_memory else MemoryMappedTable
|
| 370 |
+
return table_cls.from_file(filename)
|
| 371 |
+
|
| 372 |
+
|
| 373 |
+
class ParquetReader(BaseReader):
|
| 374 |
+
"""
|
| 375 |
+
Build a Dataset object out of Instruction instance(s).
|
| 376 |
+
This Reader uses memory mapping on parquet files.
|
| 377 |
+
"""
|
| 378 |
+
|
| 379 |
+
def __init__(self, path: str, info: Optional["DatasetInfo"]):
|
| 380 |
+
"""Initializes ParquetReader.
|
| 381 |
+
|
| 382 |
+
Args:
|
| 383 |
+
path (str): path where tfrecords are stored.
|
| 384 |
+
info (DatasetInfo): info about the dataset.
|
| 385 |
+
"""
|
| 386 |
+
super().__init__(path, info)
|
| 387 |
+
self._filetype_suffix = "parquet"
|
| 388 |
+
|
| 389 |
+
def _get_table_from_filename(self, filename_skip_take, **kwargs):
|
| 390 |
+
"""Returns a Dataset instance from given (filename, skip, take)."""
|
| 391 |
+
filename, skip, take = (
|
| 392 |
+
filename_skip_take["filename"],
|
| 393 |
+
filename_skip_take["skip"] if "skip" in filename_skip_take else None,
|
| 394 |
+
filename_skip_take["take"] if "take" in filename_skip_take else None,
|
| 395 |
+
)
|
| 396 |
+
# Parquet read_table always loads data in memory, independently of memory_map
|
| 397 |
+
pa_table = pq.read_table(filename, memory_map=True)
|
| 398 |
+
# here we don't want to slice an empty table, or it may segfault
|
| 399 |
+
if skip is not None and take is not None and not (skip == 0 and take == len(pa_table)):
|
| 400 |
+
pa_table = pa_table.slice(skip, take)
|
| 401 |
+
return pa_table
|
| 402 |
+
|
| 403 |
+
|
| 404 |
+
@dataclass(frozen=True)
|
| 405 |
+
class _AbsoluteInstruction:
|
| 406 |
+
"""A machine friendly slice: defined absolute positive boundaries."""
|
| 407 |
+
|
| 408 |
+
splitname: str
|
| 409 |
+
from_: int # uint (starting index).
|
| 410 |
+
to: int # uint (ending index).
|
| 411 |
+
|
| 412 |
+
|
| 413 |
+
@dataclass(frozen=True)
|
| 414 |
+
class _RelativeInstruction:
|
| 415 |
+
"""Represents a single parsed slicing instruction, can use % and negatives."""
|
| 416 |
+
|
| 417 |
+
splitname: str
|
| 418 |
+
from_: Optional[int] = None # int (starting index) or None if no lower boundary.
|
| 419 |
+
to: Optional[int] = None # int (ending index) or None if no upper boundary.
|
| 420 |
+
unit: Optional[str] = None
|
| 421 |
+
rounding: Optional[str] = None
|
| 422 |
+
|
| 423 |
+
def __post_init__(self):
|
| 424 |
+
if self.unit is not None and self.unit not in ["%", "abs"]:
|
| 425 |
+
raise ValueError("unit must be either % or abs")
|
| 426 |
+
if self.rounding is not None and self.rounding not in ["closest", "pct1_dropremainder"]:
|
| 427 |
+
raise ValueError("rounding must be either closest or pct1_dropremainder")
|
| 428 |
+
if self.unit != "%" and self.rounding is not None:
|
| 429 |
+
raise ValueError("It is forbidden to specify rounding if not using percent slicing.")
|
| 430 |
+
if self.unit == "%" and self.from_ is not None and abs(self.from_) > 100:
|
| 431 |
+
raise ValueError("Percent slice boundaries must be > -100 and < 100.")
|
| 432 |
+
if self.unit == "%" and self.to is not None and abs(self.to) > 100:
|
| 433 |
+
raise ValueError("Percent slice boundaries must be > -100 and < 100.")
|
| 434 |
+
# Update via __dict__ due to instance being "frozen"
|
| 435 |
+
self.__dict__["rounding"] = "closest" if self.rounding is None and self.unit == "%" else self.rounding
|
| 436 |
+
|
| 437 |
+
|
| 438 |
+
def _str_to_read_instruction(spec):
|
| 439 |
+
"""Returns ReadInstruction for given string."""
|
| 440 |
+
res = _SUB_SPEC_RE.match(spec)
|
| 441 |
+
if not res:
|
| 442 |
+
raise ValueError(f"Unrecognized instruction format: {spec}")
|
| 443 |
+
unit = "%" if res.group("from_pct") or res.group("to_pct") else "abs"
|
| 444 |
+
return ReadInstruction(
|
| 445 |
+
split_name=res.group("split"),
|
| 446 |
+
rounding=res.group("rounding"),
|
| 447 |
+
from_=int(res.group("from")) if res.group("from") else None,
|
| 448 |
+
to=int(res.group("to")) if res.group("to") else None,
|
| 449 |
+
unit=unit,
|
| 450 |
+
)
|
| 451 |
+
|
| 452 |
+
|
| 453 |
+
def _pct_to_abs_pct1(boundary, num_examples):
|
| 454 |
+
# Using math.trunc here, since -99.5% should give -99%, not -100%.
|
| 455 |
+
if num_examples < 100:
|
| 456 |
+
msg = (
|
| 457 |
+
'Using "pct1_dropremainder" rounding on a split with less than 100 '
|
| 458 |
+
"elements is forbidden: it always results in an empty dataset."
|
| 459 |
+
)
|
| 460 |
+
raise ValueError(msg)
|
| 461 |
+
return boundary * math.trunc(num_examples / 100.0)
|
| 462 |
+
|
| 463 |
+
|
| 464 |
+
def _pct_to_abs_closest(boundary, num_examples):
|
| 465 |
+
return int(round(boundary * num_examples / 100.0))
|
| 466 |
+
|
| 467 |
+
|
| 468 |
+
def _rel_to_abs_instr(rel_instr, name2len):
|
| 469 |
+
"""Returns _AbsoluteInstruction instance for given RelativeInstruction.
|
| 470 |
+
|
| 471 |
+
Args:
|
| 472 |
+
rel_instr: RelativeInstruction instance.
|
| 473 |
+
name2len: dict {split_name: num_examples}.
|
| 474 |
+
"""
|
| 475 |
+
pct_to_abs = _pct_to_abs_closest if rel_instr.rounding == "closest" else _pct_to_abs_pct1
|
| 476 |
+
split = rel_instr.splitname
|
| 477 |
+
if split not in name2len:
|
| 478 |
+
raise ValueError(f'Unknown split "{split}". Should be one of {list(name2len)}.')
|
| 479 |
+
num_examples = name2len[split]
|
| 480 |
+
from_ = rel_instr.from_
|
| 481 |
+
to = rel_instr.to
|
| 482 |
+
if rel_instr.unit == "%":
|
| 483 |
+
from_ = 0 if from_ is None else pct_to_abs(from_, num_examples)
|
| 484 |
+
to = num_examples if to is None else pct_to_abs(to, num_examples)
|
| 485 |
+
else:
|
| 486 |
+
from_ = 0 if from_ is None else from_
|
| 487 |
+
to = num_examples if to is None else to
|
| 488 |
+
if from_ < 0:
|
| 489 |
+
from_ = max(num_examples + from_, 0)
|
| 490 |
+
if to < 0:
|
| 491 |
+
to = max(num_examples + to, 0)
|
| 492 |
+
from_ = min(from_, num_examples)
|
| 493 |
+
to = min(to, num_examples)
|
| 494 |
+
return _AbsoluteInstruction(split, from_, to)
|
| 495 |
+
|
| 496 |
+
|
| 497 |
+
class ReadInstruction:
|
| 498 |
+
"""Reading instruction for a dataset.
|
| 499 |
+
|
| 500 |
+
Examples::
|
| 501 |
+
|
| 502 |
+
# The following lines are equivalent:
|
| 503 |
+
ds = datasets.load_dataset('mnist', split='test[:33%]')
|
| 504 |
+
ds = datasets.load_dataset('mnist', split=datasets.ReadInstruction.from_spec('test[:33%]'))
|
| 505 |
+
ds = datasets.load_dataset('mnist', split=datasets.ReadInstruction('test', to=33, unit='%'))
|
| 506 |
+
ds = datasets.load_dataset('mnist', split=datasets.ReadInstruction(
|
| 507 |
+
'test', from_=0, to=33, unit='%'))
|
| 508 |
+
|
| 509 |
+
# The following lines are equivalent:
|
| 510 |
+
ds = datasets.load_dataset('mnist', split='test[:33%]+train[1:-1]')
|
| 511 |
+
ds = datasets.load_dataset('mnist', split=datasets.ReadInstruction.from_spec(
|
| 512 |
+
'test[:33%]+train[1:-1]'))
|
| 513 |
+
ds = datasets.load_dataset('mnist', split=(
|
| 514 |
+
datasets.ReadInstruction('test', to=33, unit='%') +
|
| 515 |
+
datasets.ReadInstruction('train', from_=1, to=-1, unit='abs')))
|
| 516 |
+
|
| 517 |
+
# The following lines are equivalent:
|
| 518 |
+
ds = datasets.load_dataset('mnist', split='test[:33%](pct1_dropremainder)')
|
| 519 |
+
ds = datasets.load_dataset('mnist', split=datasets.ReadInstruction.from_spec(
|
| 520 |
+
'test[:33%](pct1_dropremainder)'))
|
| 521 |
+
ds = datasets.load_dataset('mnist', split=datasets.ReadInstruction(
|
| 522 |
+
'test', from_=0, to=33, unit='%', rounding="pct1_dropremainder"))
|
| 523 |
+
|
| 524 |
+
# 10-fold validation:
|
| 525 |
+
tests = datasets.load_dataset(
|
| 526 |
+
'mnist',
|
| 527 |
+
[datasets.ReadInstruction('train', from_=k, to=k+10, unit='%')
|
| 528 |
+
for k in range(0, 100, 10)])
|
| 529 |
+
trains = datasets.load_dataset(
|
| 530 |
+
'mnist',
|
| 531 |
+
[datasets.ReadInstruction('train', to=k, unit='%') + datasets.ReadInstruction('train', from_=k+10, unit='%')
|
| 532 |
+
for k in range(0, 100, 10)])
|
| 533 |
+
|
| 534 |
+
"""
|
| 535 |
+
|
| 536 |
+
def _init(self, relative_instructions):
|
| 537 |
+
# Private initializer.
|
| 538 |
+
self._relative_instructions = relative_instructions
|
| 539 |
+
|
| 540 |
+
@classmethod
|
| 541 |
+
def _read_instruction_from_relative_instructions(cls, relative_instructions):
|
| 542 |
+
"""Returns ReadInstruction obj initialized with relative_instructions."""
|
| 543 |
+
# Use __new__ to bypass __init__ used by public API and not conveniant here.
|
| 544 |
+
result = cls.__new__(cls)
|
| 545 |
+
result._init(relative_instructions) # pylint: disable=protected-access
|
| 546 |
+
return result
|
| 547 |
+
|
| 548 |
+
def __init__(self, split_name, rounding=None, from_=None, to=None, unit=None):
|
| 549 |
+
"""Initialize ReadInstruction.
|
| 550 |
+
|
| 551 |
+
Args:
|
| 552 |
+
split_name (str): name of the split to read. Eg: 'train'.
|
| 553 |
+
rounding (str, optional): The rounding behaviour to use when percent slicing is
|
| 554 |
+
used. Ignored when slicing with absolute indices.
|
| 555 |
+
Possible values:
|
| 556 |
+
- 'closest' (default): The specified percentages are rounded to the
|
| 557 |
+
closest value. Use this if you want specified percents to be as
|
| 558 |
+
much exact as possible.
|
| 559 |
+
- 'pct1_dropremainder': the specified percentages are treated as
|
| 560 |
+
multiple of 1%. Use this option if you want consistency. Eg:
|
| 561 |
+
len(5%) == 5 * len(1%).
|
| 562 |
+
Using this option, one might not be able to use the full set of
|
| 563 |
+
examples, if the number of those is not a multiple of 100.
|
| 564 |
+
from_ (int):
|
| 565 |
+
to (int): alternative way of specifying slicing boundaries. If any of
|
| 566 |
+
{from_, to, unit} argument is used, slicing cannot be specified as
|
| 567 |
+
string.
|
| 568 |
+
unit (str): optional, one of:
|
| 569 |
+
'%': to set the slicing unit as percents of the split size.
|
| 570 |
+
'abs': to set the slicing unit as absolute numbers.
|
| 571 |
+
"""
|
| 572 |
+
# This constructor is not always called. See factory method
|
| 573 |
+
# `_read_instruction_from_relative_instructions`. Common init instructions
|
| 574 |
+
# MUST be placed in the _init method.
|
| 575 |
+
self._init([_RelativeInstruction(split_name, from_, to, unit, rounding)])
|
| 576 |
+
|
| 577 |
+
@classmethod
|
| 578 |
+
def from_spec(cls, spec):
|
| 579 |
+
"""Creates a `ReadInstruction` instance out of a string spec.
|
| 580 |
+
|
| 581 |
+
Args:
|
| 582 |
+
spec (`str`):
|
| 583 |
+
Split(s) + optional slice(s) to read + optional rounding
|
| 584 |
+
if percents are used as the slicing unit. A slice can be specified,
|
| 585 |
+
using absolute numbers (`int`) or percentages (`int`).
|
| 586 |
+
|
| 587 |
+
Examples:
|
| 588 |
+
|
| 589 |
+
```
|
| 590 |
+
test: test split.
|
| 591 |
+
test + validation: test split + validation split.
|
| 592 |
+
test[10:]: test split, minus its first 10 records.
|
| 593 |
+
test[:10%]: first 10% records of test split.
|
| 594 |
+
test[:20%](pct1_dropremainder): first 10% records, rounded with the pct1_dropremainder rounding.
|
| 595 |
+
test[:-5%]+train[40%:60%]: first 95% of test + middle 20% of train.
|
| 596 |
+
```
|
| 597 |
+
|
| 598 |
+
Returns:
|
| 599 |
+
ReadInstruction instance.
|
| 600 |
+
"""
|
| 601 |
+
spec = str(spec) # Need to convert to str in case of NamedSplit instance.
|
| 602 |
+
subs = _ADDITION_SEP_RE.split(spec)
|
| 603 |
+
if not subs:
|
| 604 |
+
raise ValueError(f"No instructions could be built out of {spec}")
|
| 605 |
+
instruction = _str_to_read_instruction(subs[0])
|
| 606 |
+
return sum((_str_to_read_instruction(sub) for sub in subs[1:]), instruction)
|
| 607 |
+
|
| 608 |
+
def to_spec(self):
|
| 609 |
+
rel_instr_specs = []
|
| 610 |
+
for rel_instr in self._relative_instructions:
|
| 611 |
+
rel_instr_spec = rel_instr.splitname
|
| 612 |
+
if rel_instr.from_ is not None or rel_instr.to is not None:
|
| 613 |
+
from_ = rel_instr.from_
|
| 614 |
+
to = rel_instr.to
|
| 615 |
+
unit = rel_instr.unit
|
| 616 |
+
rounding = rel_instr.rounding
|
| 617 |
+
unit = unit if unit == "%" else ""
|
| 618 |
+
from_ = str(from_) + unit if from_ is not None else ""
|
| 619 |
+
to = str(to) + unit if to is not None else ""
|
| 620 |
+
slice_str = f"[{from_}:{to}]"
|
| 621 |
+
rounding_str = (
|
| 622 |
+
f"({rounding})" if unit == "%" and rounding is not None and rounding != "closest" else ""
|
| 623 |
+
)
|
| 624 |
+
rel_instr_spec += slice_str + rounding_str
|
| 625 |
+
rel_instr_specs.append(rel_instr_spec)
|
| 626 |
+
return "+".join(rel_instr_specs)
|
| 627 |
+
|
| 628 |
+
def __add__(self, other):
|
| 629 |
+
"""Returns a new ReadInstruction obj, result of appending other to self."""
|
| 630 |
+
if not isinstance(other, ReadInstruction):
|
| 631 |
+
msg = "ReadInstruction can only be added to another ReadInstruction obj."
|
| 632 |
+
raise TypeError(msg)
|
| 633 |
+
self_ris = self._relative_instructions
|
| 634 |
+
other_ris = other._relative_instructions # pylint: disable=protected-access
|
| 635 |
+
if (
|
| 636 |
+
self_ris[0].unit != "abs"
|
| 637 |
+
and other_ris[0].unit != "abs"
|
| 638 |
+
and self._relative_instructions[0].rounding != other_ris[0].rounding
|
| 639 |
+
):
|
| 640 |
+
raise ValueError("It is forbidden to sum ReadInstruction instances with different rounding values.")
|
| 641 |
+
return self._read_instruction_from_relative_instructions(self_ris + other_ris)
|
| 642 |
+
|
| 643 |
+
def __str__(self):
|
| 644 |
+
return self.to_spec()
|
| 645 |
+
|
| 646 |
+
def __repr__(self):
|
| 647 |
+
return f"ReadInstruction({self._relative_instructions})"
|
| 648 |
+
|
| 649 |
+
def to_absolute(self, name2len):
|
| 650 |
+
"""Translate instruction into a list of absolute instructions.
|
| 651 |
+
|
| 652 |
+
Those absolute instructions are then to be added together.
|
| 653 |
+
|
| 654 |
+
Args:
|
| 655 |
+
name2len (`dict`):
|
| 656 |
+
Associating split names to number of examples.
|
| 657 |
+
|
| 658 |
+
Returns:
|
| 659 |
+
list of _AbsoluteInstruction instances (corresponds to the + in spec).
|
| 660 |
+
"""
|
| 661 |
+
return [_rel_to_abs_instr(rel_instr, name2len) for rel_instr in self._relative_instructions]
|
evalkit_tf437/lib/python3.10/site-packages/datasets/arrow_writer.py
ADDED
|
@@ -0,0 +1,745 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors.
|
| 2 |
+
#
|
| 3 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
| 4 |
+
# you may not use this file except in compliance with the License.
|
| 5 |
+
# You may obtain a copy of the License at
|
| 6 |
+
#
|
| 7 |
+
# Unless required by applicable law or agreed to in writing, software
|
| 8 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
| 9 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| 10 |
+
# See the License for the specific language governing permissions and
|
| 11 |
+
# limitations under the License.
|
| 12 |
+
|
| 13 |
+
# Lint as: python3
|
| 14 |
+
"""To write records into Parquet files."""
|
| 15 |
+
|
| 16 |
+
import errno
|
| 17 |
+
import json
|
| 18 |
+
import os
|
| 19 |
+
import sys
|
| 20 |
+
from pathlib import Path
|
| 21 |
+
from typing import Any, Dict, Iterable, List, Optional, Tuple, Union
|
| 22 |
+
|
| 23 |
+
import fsspec
|
| 24 |
+
import numpy as np
|
| 25 |
+
import pyarrow as pa
|
| 26 |
+
import pyarrow.parquet as pq
|
| 27 |
+
|
| 28 |
+
from . import config
|
| 29 |
+
from .features import Features, Image, Value
|
| 30 |
+
from .features.features import (
|
| 31 |
+
FeatureType,
|
| 32 |
+
_ArrayXDExtensionType,
|
| 33 |
+
cast_to_python_objects,
|
| 34 |
+
generate_from_arrow_type,
|
| 35 |
+
get_nested_type,
|
| 36 |
+
list_of_np_array_to_pyarrow_listarray,
|
| 37 |
+
numpy_to_pyarrow_listarray,
|
| 38 |
+
to_pyarrow_listarray,
|
| 39 |
+
)
|
| 40 |
+
from .filesystems import is_remote_filesystem
|
| 41 |
+
from .info import DatasetInfo
|
| 42 |
+
from .keyhash import DuplicatedKeysError, KeyHasher
|
| 43 |
+
from .table import array_cast, cast_array_to_feature, embed_table_storage, table_cast
|
| 44 |
+
from .utils import logging
|
| 45 |
+
from .utils import tqdm as hf_tqdm
|
| 46 |
+
from .utils.file_utils import hash_url_to_filename
|
| 47 |
+
from .utils.py_utils import asdict, first_non_null_value
|
| 48 |
+
|
| 49 |
+
|
| 50 |
+
logger = logging.get_logger(__name__)
|
| 51 |
+
|
| 52 |
+
type_ = type # keep python's type function
|
| 53 |
+
|
| 54 |
+
|
| 55 |
+
class SchemaInferenceError(ValueError):
|
| 56 |
+
pass
|
| 57 |
+
|
| 58 |
+
|
| 59 |
+
class TypedSequence:
|
| 60 |
+
"""
|
| 61 |
+
This data container generalizes the typing when instantiating pyarrow arrays, tables or batches.
|
| 62 |
+
|
| 63 |
+
More specifically it adds several features:
|
| 64 |
+
- Support extension types like ``datasets.features.Array2DExtensionType``:
|
| 65 |
+
By default pyarrow arrays don't return extension arrays. One has to call
|
| 66 |
+
``pa.ExtensionArray.from_storage(type, pa.array(data, type.storage_type))``
|
| 67 |
+
in order to get an extension array.
|
| 68 |
+
- Support for ``try_type`` parameter that can be used instead of ``type``:
|
| 69 |
+
When an array is transformed, we like to keep the same type as before if possible.
|
| 70 |
+
For example when calling :func:`datasets.Dataset.map`, we don't want to change the type
|
| 71 |
+
of each column by default.
|
| 72 |
+
- Better error message when a pyarrow array overflows.
|
| 73 |
+
|
| 74 |
+
Example::
|
| 75 |
+
|
| 76 |
+
from datasets.features import Array2D, Array2DExtensionType, Value
|
| 77 |
+
from datasets.arrow_writer import TypedSequence
|
| 78 |
+
import pyarrow as pa
|
| 79 |
+
|
| 80 |
+
arr = pa.array(TypedSequence([1, 2, 3], type=Value("int32")))
|
| 81 |
+
assert arr.type == pa.int32()
|
| 82 |
+
|
| 83 |
+
arr = pa.array(TypedSequence([1, 2, 3], try_type=Value("int32")))
|
| 84 |
+
assert arr.type == pa.int32()
|
| 85 |
+
|
| 86 |
+
arr = pa.array(TypedSequence(["foo", "bar"], try_type=Value("int32")))
|
| 87 |
+
assert arr.type == pa.string()
|
| 88 |
+
|
| 89 |
+
arr = pa.array(TypedSequence([[[1, 2, 3]]], type=Array2D((1, 3), "int64")))
|
| 90 |
+
assert arr.type == Array2DExtensionType((1, 3), "int64")
|
| 91 |
+
|
| 92 |
+
table = pa.Table.from_pydict({
|
| 93 |
+
"image": TypedSequence([[[1, 2, 3]]], type=Array2D((1, 3), "int64"))
|
| 94 |
+
})
|
| 95 |
+
assert table["image"].type == Array2DExtensionType((1, 3), "int64")
|
| 96 |
+
|
| 97 |
+
"""
|
| 98 |
+
|
| 99 |
+
def __init__(
|
| 100 |
+
self,
|
| 101 |
+
data: Iterable,
|
| 102 |
+
type: Optional[FeatureType] = None,
|
| 103 |
+
try_type: Optional[FeatureType] = None,
|
| 104 |
+
optimized_int_type: Optional[FeatureType] = None,
|
| 105 |
+
):
|
| 106 |
+
# assert type is None or try_type is None,
|
| 107 |
+
if type is not None and try_type is not None:
|
| 108 |
+
raise ValueError("You cannot specify both type and try_type")
|
| 109 |
+
# set attributes
|
| 110 |
+
self.data = data
|
| 111 |
+
self.type = type
|
| 112 |
+
self.try_type = try_type # is ignored if it doesn't match the data
|
| 113 |
+
self.optimized_int_type = optimized_int_type
|
| 114 |
+
# when trying a type (is ignored if data is not compatible)
|
| 115 |
+
self.trying_type = self.try_type is not None
|
| 116 |
+
self.trying_int_optimization = optimized_int_type is not None and type is None and try_type is None
|
| 117 |
+
# used to get back the inferred type after __arrow_array__() is called once
|
| 118 |
+
self._inferred_type = None
|
| 119 |
+
|
| 120 |
+
def get_inferred_type(self) -> FeatureType:
|
| 121 |
+
"""Return the inferred feature type.
|
| 122 |
+
This is done by converting the sequence to an Arrow array, and getting the corresponding
|
| 123 |
+
feature type.
|
| 124 |
+
|
| 125 |
+
Since building the Arrow array can be expensive, the value of the inferred type is cached
|
| 126 |
+
as soon as pa.array is called on the typed sequence.
|
| 127 |
+
|
| 128 |
+
Returns:
|
| 129 |
+
FeatureType: inferred feature type of the sequence.
|
| 130 |
+
"""
|
| 131 |
+
if self._inferred_type is None:
|
| 132 |
+
self._inferred_type = generate_from_arrow_type(pa.array(self).type)
|
| 133 |
+
return self._inferred_type
|
| 134 |
+
|
| 135 |
+
@staticmethod
|
| 136 |
+
def _infer_custom_type_and_encode(data: Iterable) -> Tuple[Iterable, Optional[FeatureType]]:
|
| 137 |
+
"""Implement type inference for custom objects like PIL.Image.Image -> Image type.
|
| 138 |
+
|
| 139 |
+
This function is only used for custom python objects that can't be direclty passed to build
|
| 140 |
+
an Arrow array. In such cases is infers the feature type to use, and it encodes the data so
|
| 141 |
+
that they can be passed to an Arrow array.
|
| 142 |
+
|
| 143 |
+
Args:
|
| 144 |
+
data (Iterable): array of data to infer the type, e.g. a list of PIL images.
|
| 145 |
+
|
| 146 |
+
Returns:
|
| 147 |
+
Tuple[Iterable, Optional[FeatureType]]: a tuple with:
|
| 148 |
+
- the (possibly encoded) array, if the inferred feature type requires encoding
|
| 149 |
+
- the inferred feature type if the array is made of supported custom objects like
|
| 150 |
+
PIL images, else None.
|
| 151 |
+
"""
|
| 152 |
+
if config.PIL_AVAILABLE and "PIL" in sys.modules:
|
| 153 |
+
import PIL.Image
|
| 154 |
+
|
| 155 |
+
non_null_idx, non_null_value = first_non_null_value(data)
|
| 156 |
+
if isinstance(non_null_value, PIL.Image.Image):
|
| 157 |
+
return [Image().encode_example(value) if value is not None else None for value in data], Image()
|
| 158 |
+
return data, None
|
| 159 |
+
|
| 160 |
+
def __arrow_array__(self, type: Optional[pa.DataType] = None):
|
| 161 |
+
"""This function is called when calling pa.array(typed_sequence)"""
|
| 162 |
+
|
| 163 |
+
if type is not None:
|
| 164 |
+
raise ValueError("TypedSequence is supposed to be used with pa.array(typed_sequence, type=None)")
|
| 165 |
+
del type # make sure we don't use it
|
| 166 |
+
data = self.data
|
| 167 |
+
# automatic type inference for custom objects
|
| 168 |
+
if self.type is None and self.try_type is None:
|
| 169 |
+
data, self._inferred_type = self._infer_custom_type_and_encode(data)
|
| 170 |
+
if self._inferred_type is None:
|
| 171 |
+
type = self.try_type if self.trying_type else self.type
|
| 172 |
+
else:
|
| 173 |
+
type = self._inferred_type
|
| 174 |
+
pa_type = get_nested_type(type) if type is not None else None
|
| 175 |
+
optimized_int_pa_type = (
|
| 176 |
+
get_nested_type(self.optimized_int_type) if self.optimized_int_type is not None else None
|
| 177 |
+
)
|
| 178 |
+
trying_cast_to_python_objects = False
|
| 179 |
+
try:
|
| 180 |
+
# custom pyarrow types
|
| 181 |
+
if isinstance(pa_type, _ArrayXDExtensionType):
|
| 182 |
+
storage = to_pyarrow_listarray(data, pa_type)
|
| 183 |
+
return pa.ExtensionArray.from_storage(pa_type, storage)
|
| 184 |
+
|
| 185 |
+
# efficient np array to pyarrow array
|
| 186 |
+
if isinstance(data, np.ndarray):
|
| 187 |
+
out = numpy_to_pyarrow_listarray(data)
|
| 188 |
+
elif isinstance(data, list) and data and isinstance(first_non_null_value(data)[1], np.ndarray):
|
| 189 |
+
out = list_of_np_array_to_pyarrow_listarray(data)
|
| 190 |
+
else:
|
| 191 |
+
trying_cast_to_python_objects = True
|
| 192 |
+
out = pa.array(cast_to_python_objects(data, only_1d_for_numpy=True))
|
| 193 |
+
# use smaller integer precisions if possible
|
| 194 |
+
if self.trying_int_optimization:
|
| 195 |
+
if pa.types.is_int64(out.type):
|
| 196 |
+
out = out.cast(optimized_int_pa_type)
|
| 197 |
+
elif pa.types.is_list(out.type):
|
| 198 |
+
if pa.types.is_int64(out.type.value_type):
|
| 199 |
+
out = array_cast(out, pa.list_(optimized_int_pa_type))
|
| 200 |
+
elif pa.types.is_list(out.type.value_type) and pa.types.is_int64(out.type.value_type.value_type):
|
| 201 |
+
out = array_cast(out, pa.list_(pa.list_(optimized_int_pa_type)))
|
| 202 |
+
# otherwise we can finally use the user's type
|
| 203 |
+
elif type is not None:
|
| 204 |
+
# We use cast_array_to_feature to support casting to custom types like Audio and Image
|
| 205 |
+
# Also, when trying type "string", we don't want to convert integers or floats to "string".
|
| 206 |
+
# We only do it if trying_type is False - since this is what the user asks for.
|
| 207 |
+
out = cast_array_to_feature(out, type, allow_number_to_str=not self.trying_type)
|
| 208 |
+
return out
|
| 209 |
+
except (
|
| 210 |
+
TypeError,
|
| 211 |
+
pa.lib.ArrowInvalid,
|
| 212 |
+
pa.lib.ArrowNotImplementedError,
|
| 213 |
+
) as e: # handle type errors and overflows
|
| 214 |
+
# Ignore ArrowNotImplementedError caused by trying type, otherwise re-raise
|
| 215 |
+
if not self.trying_type and isinstance(e, pa.lib.ArrowNotImplementedError):
|
| 216 |
+
raise
|
| 217 |
+
|
| 218 |
+
if self.trying_type:
|
| 219 |
+
try: # second chance
|
| 220 |
+
if isinstance(data, np.ndarray):
|
| 221 |
+
return numpy_to_pyarrow_listarray(data)
|
| 222 |
+
elif isinstance(data, list) and data and any(isinstance(value, np.ndarray) for value in data):
|
| 223 |
+
return list_of_np_array_to_pyarrow_listarray(data)
|
| 224 |
+
else:
|
| 225 |
+
trying_cast_to_python_objects = True
|
| 226 |
+
return pa.array(cast_to_python_objects(data, only_1d_for_numpy=True))
|
| 227 |
+
except pa.lib.ArrowInvalid as e:
|
| 228 |
+
if "overflow" in str(e):
|
| 229 |
+
raise OverflowError(
|
| 230 |
+
f"There was an overflow with type {type_(data)}. Try to reduce writer_batch_size to have batches smaller than 2GB.\n({e})"
|
| 231 |
+
) from None
|
| 232 |
+
elif self.trying_int_optimization and "not in range" in str(e):
|
| 233 |
+
optimized_int_pa_type_str = np.dtype(optimized_int_pa_type.to_pandas_dtype()).name
|
| 234 |
+
logger.info(
|
| 235 |
+
f"Failed to cast a sequence to {optimized_int_pa_type_str}. Falling back to int64."
|
| 236 |
+
)
|
| 237 |
+
return out
|
| 238 |
+
elif trying_cast_to_python_objects and "Could not convert" in str(e):
|
| 239 |
+
out = pa.array(
|
| 240 |
+
cast_to_python_objects(data, only_1d_for_numpy=True, optimize_list_casting=False)
|
| 241 |
+
)
|
| 242 |
+
if type is not None:
|
| 243 |
+
out = cast_array_to_feature(out, type, allow_number_to_str=True)
|
| 244 |
+
return out
|
| 245 |
+
else:
|
| 246 |
+
raise
|
| 247 |
+
elif "overflow" in str(e):
|
| 248 |
+
raise OverflowError(
|
| 249 |
+
f"There was an overflow with type {type_(data)}. Try to reduce writer_batch_size to have batches smaller than 2GB.\n({e})"
|
| 250 |
+
) from None
|
| 251 |
+
elif self.trying_int_optimization and "not in range" in str(e):
|
| 252 |
+
optimized_int_pa_type_str = np.dtype(optimized_int_pa_type.to_pandas_dtype()).name
|
| 253 |
+
logger.info(f"Failed to cast a sequence to {optimized_int_pa_type_str}. Falling back to int64.")
|
| 254 |
+
return out
|
| 255 |
+
elif trying_cast_to_python_objects and "Could not convert" in str(e):
|
| 256 |
+
out = pa.array(cast_to_python_objects(data, only_1d_for_numpy=True, optimize_list_casting=False))
|
| 257 |
+
if type is not None:
|
| 258 |
+
out = cast_array_to_feature(out, type, allow_number_to_str=True)
|
| 259 |
+
return out
|
| 260 |
+
else:
|
| 261 |
+
raise
|
| 262 |
+
|
| 263 |
+
|
| 264 |
+
class OptimizedTypedSequence(TypedSequence):
|
| 265 |
+
def __init__(
|
| 266 |
+
self,
|
| 267 |
+
data,
|
| 268 |
+
type: Optional[FeatureType] = None,
|
| 269 |
+
try_type: Optional[FeatureType] = None,
|
| 270 |
+
col: Optional[str] = None,
|
| 271 |
+
optimized_int_type: Optional[FeatureType] = None,
|
| 272 |
+
):
|
| 273 |
+
optimized_int_type_by_col = {
|
| 274 |
+
"attention_mask": Value("int8"), # binary tensor
|
| 275 |
+
"special_tokens_mask": Value("int8"),
|
| 276 |
+
"input_ids": Value("int32"), # typical vocab size: 0-50k (max ~500k, never > 1M)
|
| 277 |
+
"token_type_ids": Value(
|
| 278 |
+
"int8"
|
| 279 |
+
), # binary mask; some (XLNetModel) use an additional token represented by a 2
|
| 280 |
+
}
|
| 281 |
+
if type is None and try_type is None:
|
| 282 |
+
optimized_int_type = optimized_int_type_by_col.get(col, None)
|
| 283 |
+
super().__init__(data, type=type, try_type=try_type, optimized_int_type=optimized_int_type)
|
| 284 |
+
|
| 285 |
+
|
| 286 |
+
class ArrowWriter:
|
| 287 |
+
"""Shuffles and writes Examples to Arrow files."""
|
| 288 |
+
|
| 289 |
+
_WRITER_CLASS = pa.RecordBatchStreamWriter
|
| 290 |
+
|
| 291 |
+
def __init__(
|
| 292 |
+
self,
|
| 293 |
+
schema: Optional[pa.Schema] = None,
|
| 294 |
+
features: Optional[Features] = None,
|
| 295 |
+
path: Optional[str] = None,
|
| 296 |
+
stream: Optional[pa.NativeFile] = None,
|
| 297 |
+
fingerprint: Optional[str] = None,
|
| 298 |
+
writer_batch_size: Optional[int] = None,
|
| 299 |
+
hash_salt: Optional[str] = None,
|
| 300 |
+
check_duplicates: Optional[bool] = False,
|
| 301 |
+
disable_nullable: bool = False,
|
| 302 |
+
update_features: bool = False,
|
| 303 |
+
with_metadata: bool = True,
|
| 304 |
+
unit: str = "examples",
|
| 305 |
+
embed_local_files: bool = False,
|
| 306 |
+
storage_options: Optional[dict] = None,
|
| 307 |
+
):
|
| 308 |
+
if path is None and stream is None:
|
| 309 |
+
raise ValueError("At least one of path and stream must be provided.")
|
| 310 |
+
if features is not None:
|
| 311 |
+
self._features = features
|
| 312 |
+
self._schema = None
|
| 313 |
+
elif schema is not None:
|
| 314 |
+
self._schema: pa.Schema = schema
|
| 315 |
+
self._features = Features.from_arrow_schema(self._schema)
|
| 316 |
+
else:
|
| 317 |
+
self._features = None
|
| 318 |
+
self._schema = None
|
| 319 |
+
|
| 320 |
+
if hash_salt is not None:
|
| 321 |
+
# Create KeyHasher instance using split name as hash salt
|
| 322 |
+
self._hasher = KeyHasher(hash_salt)
|
| 323 |
+
else:
|
| 324 |
+
self._hasher = KeyHasher("")
|
| 325 |
+
|
| 326 |
+
self._check_duplicates = check_duplicates
|
| 327 |
+
self._disable_nullable = disable_nullable
|
| 328 |
+
|
| 329 |
+
if stream is None:
|
| 330 |
+
fs_token_paths = fsspec.get_fs_token_paths(path, storage_options=storage_options)
|
| 331 |
+
self._fs: fsspec.AbstractFileSystem = fs_token_paths[0]
|
| 332 |
+
self._path = (
|
| 333 |
+
fs_token_paths[2][0]
|
| 334 |
+
if not is_remote_filesystem(self._fs)
|
| 335 |
+
else self._fs.unstrip_protocol(fs_token_paths[2][0])
|
| 336 |
+
)
|
| 337 |
+
self.stream = self._fs.open(fs_token_paths[2][0], "wb")
|
| 338 |
+
self._closable_stream = True
|
| 339 |
+
else:
|
| 340 |
+
self._fs = None
|
| 341 |
+
self._path = None
|
| 342 |
+
self.stream = stream
|
| 343 |
+
self._closable_stream = False
|
| 344 |
+
|
| 345 |
+
self.fingerprint = fingerprint
|
| 346 |
+
self.disable_nullable = disable_nullable
|
| 347 |
+
self.writer_batch_size = writer_batch_size or config.DEFAULT_MAX_BATCH_SIZE
|
| 348 |
+
self.update_features = update_features
|
| 349 |
+
self.with_metadata = with_metadata
|
| 350 |
+
self.unit = unit
|
| 351 |
+
self.embed_local_files = embed_local_files
|
| 352 |
+
|
| 353 |
+
self._num_examples = 0
|
| 354 |
+
self._num_bytes = 0
|
| 355 |
+
self.current_examples: List[Tuple[Dict[str, Any], str]] = []
|
| 356 |
+
self.current_rows: List[pa.Table] = []
|
| 357 |
+
self.pa_writer: Optional[pa.RecordBatchStreamWriter] = None
|
| 358 |
+
self.hkey_record = []
|
| 359 |
+
|
| 360 |
+
def __len__(self):
|
| 361 |
+
"""Return the number of writed and staged examples"""
|
| 362 |
+
return self._num_examples + len(self.current_examples) + len(self.current_rows)
|
| 363 |
+
|
| 364 |
+
def __enter__(self):
|
| 365 |
+
return self
|
| 366 |
+
|
| 367 |
+
def __exit__(self, exc_type, exc_val, exc_tb):
|
| 368 |
+
self.close()
|
| 369 |
+
|
| 370 |
+
def close(self):
|
| 371 |
+
# Try closing if opened; if closed: pyarrow.lib.ArrowInvalid: Invalid operation on closed file
|
| 372 |
+
if self.pa_writer: # it might be None
|
| 373 |
+
try:
|
| 374 |
+
self.pa_writer.close()
|
| 375 |
+
except Exception: # pyarrow.lib.ArrowInvalid, OSError
|
| 376 |
+
pass
|
| 377 |
+
if self._closable_stream and not self.stream.closed:
|
| 378 |
+
self.stream.close() # This also closes self.pa_writer if it is opened
|
| 379 |
+
|
| 380 |
+
def _build_writer(self, inferred_schema: pa.Schema):
|
| 381 |
+
schema = self.schema
|
| 382 |
+
inferred_features = Features.from_arrow_schema(inferred_schema)
|
| 383 |
+
if self._features is not None:
|
| 384 |
+
if self.update_features: # keep original features it they match, or update them
|
| 385 |
+
fields = {field.name: field for field in self._features.type}
|
| 386 |
+
for inferred_field in inferred_features.type:
|
| 387 |
+
name = inferred_field.name
|
| 388 |
+
if name in fields:
|
| 389 |
+
if inferred_field == fields[name]:
|
| 390 |
+
inferred_features[name] = self._features[name]
|
| 391 |
+
self._features = inferred_features
|
| 392 |
+
schema: pa.Schema = inferred_schema
|
| 393 |
+
else:
|
| 394 |
+
self._features = inferred_features
|
| 395 |
+
schema: pa.Schema = inferred_features.arrow_schema
|
| 396 |
+
if self.disable_nullable:
|
| 397 |
+
schema = pa.schema(pa.field(field.name, field.type, nullable=False) for field in schema)
|
| 398 |
+
if self.with_metadata:
|
| 399 |
+
schema = schema.with_metadata(self._build_metadata(DatasetInfo(features=self._features), self.fingerprint))
|
| 400 |
+
else:
|
| 401 |
+
schema = schema.with_metadata({})
|
| 402 |
+
self._schema = schema
|
| 403 |
+
self.pa_writer = self._WRITER_CLASS(self.stream, schema)
|
| 404 |
+
|
| 405 |
+
@property
|
| 406 |
+
def schema(self):
|
| 407 |
+
_schema = (
|
| 408 |
+
self._schema
|
| 409 |
+
if self._schema is not None
|
| 410 |
+
else (pa.schema(self._features.type) if self._features is not None else None)
|
| 411 |
+
)
|
| 412 |
+
if self._disable_nullable and _schema is not None:
|
| 413 |
+
_schema = pa.schema(pa.field(field.name, field.type, nullable=False) for field in _schema)
|
| 414 |
+
return _schema if _schema is not None else []
|
| 415 |
+
|
| 416 |
+
@staticmethod
|
| 417 |
+
def _build_metadata(info: DatasetInfo, fingerprint: Optional[str] = None) -> Dict[str, str]:
|
| 418 |
+
info_keys = ["features"] # we can add support for more DatasetInfo keys in the future
|
| 419 |
+
info_as_dict = asdict(info)
|
| 420 |
+
metadata = {}
|
| 421 |
+
metadata["info"] = {key: info_as_dict[key] for key in info_keys}
|
| 422 |
+
if fingerprint is not None:
|
| 423 |
+
metadata["fingerprint"] = fingerprint
|
| 424 |
+
return {"huggingface": json.dumps(metadata)}
|
| 425 |
+
|
| 426 |
+
def write_examples_on_file(self):
|
| 427 |
+
"""Write stored examples from the write-pool of examples. It makes a table out of the examples and write it."""
|
| 428 |
+
if not self.current_examples:
|
| 429 |
+
return
|
| 430 |
+
# preserve the order the columns
|
| 431 |
+
if self.schema:
|
| 432 |
+
schema_cols = set(self.schema.names)
|
| 433 |
+
examples_cols = self.current_examples[0][0].keys() # .keys() preserves the order (unlike set)
|
| 434 |
+
common_cols = [col for col in self.schema.names if col in examples_cols]
|
| 435 |
+
extra_cols = [col for col in examples_cols if col not in schema_cols]
|
| 436 |
+
cols = common_cols + extra_cols
|
| 437 |
+
else:
|
| 438 |
+
cols = list(self.current_examples[0][0])
|
| 439 |
+
batch_examples = {}
|
| 440 |
+
for col in cols:
|
| 441 |
+
# We use row[0][col] since current_examples contains (example, key) tuples.
|
| 442 |
+
# Morever, examples could be Arrow arrays of 1 element.
|
| 443 |
+
# This can happen in `.map()` when we want to re-write the same Arrow data
|
| 444 |
+
if all(isinstance(row[0][col], (pa.Array, pa.ChunkedArray)) for row in self.current_examples):
|
| 445 |
+
arrays = [row[0][col] for row in self.current_examples]
|
| 446 |
+
arrays = [
|
| 447 |
+
chunk
|
| 448 |
+
for array in arrays
|
| 449 |
+
for chunk in (array.chunks if isinstance(array, pa.ChunkedArray) else [array])
|
| 450 |
+
]
|
| 451 |
+
batch_examples[col] = pa.concat_arrays(arrays)
|
| 452 |
+
else:
|
| 453 |
+
batch_examples[col] = [
|
| 454 |
+
row[0][col].to_pylist()[0] if isinstance(row[0][col], (pa.Array, pa.ChunkedArray)) else row[0][col]
|
| 455 |
+
for row in self.current_examples
|
| 456 |
+
]
|
| 457 |
+
self.write_batch(batch_examples=batch_examples)
|
| 458 |
+
self.current_examples = []
|
| 459 |
+
|
| 460 |
+
def write_rows_on_file(self):
|
| 461 |
+
"""Write stored rows from the write-pool of rows. It concatenates the single-row tables and it writes the resulting table."""
|
| 462 |
+
if not self.current_rows:
|
| 463 |
+
return
|
| 464 |
+
table = pa.concat_tables(self.current_rows)
|
| 465 |
+
self.write_table(table)
|
| 466 |
+
self.current_rows = []
|
| 467 |
+
|
| 468 |
+
def write(
|
| 469 |
+
self,
|
| 470 |
+
example: Dict[str, Any],
|
| 471 |
+
key: Optional[Union[str, int, bytes]] = None,
|
| 472 |
+
writer_batch_size: Optional[int] = None,
|
| 473 |
+
):
|
| 474 |
+
"""Add a given (Example,Key) pair to the write-pool of examples which is written to file.
|
| 475 |
+
|
| 476 |
+
Args:
|
| 477 |
+
example: the Example to add.
|
| 478 |
+
key: Optional, a unique identifier(str, int or bytes) associated with each example
|
| 479 |
+
"""
|
| 480 |
+
# Utilize the keys and duplicate checking when `self._check_duplicates` is passed True
|
| 481 |
+
if self._check_duplicates:
|
| 482 |
+
# Create unique hash from key and store as (key, example) pairs
|
| 483 |
+
hash = self._hasher.hash(key)
|
| 484 |
+
self.current_examples.append((example, hash))
|
| 485 |
+
# Maintain record of keys and their respective hashes for checking duplicates
|
| 486 |
+
self.hkey_record.append((hash, key))
|
| 487 |
+
else:
|
| 488 |
+
# Store example as a tuple so as to keep the structure of `self.current_examples` uniform
|
| 489 |
+
self.current_examples.append((example, ""))
|
| 490 |
+
|
| 491 |
+
if writer_batch_size is None:
|
| 492 |
+
writer_batch_size = self.writer_batch_size
|
| 493 |
+
if writer_batch_size is not None and len(self.current_examples) >= writer_batch_size:
|
| 494 |
+
if self._check_duplicates:
|
| 495 |
+
self.check_duplicate_keys()
|
| 496 |
+
# Re-intializing to empty list for next batch
|
| 497 |
+
self.hkey_record = []
|
| 498 |
+
|
| 499 |
+
self.write_examples_on_file()
|
| 500 |
+
|
| 501 |
+
def check_duplicate_keys(self):
|
| 502 |
+
"""Raises error if duplicates found in a batch"""
|
| 503 |
+
tmp_record = set()
|
| 504 |
+
for hash, key in self.hkey_record:
|
| 505 |
+
if hash in tmp_record:
|
| 506 |
+
duplicate_key_indices = [
|
| 507 |
+
str(self._num_examples + index)
|
| 508 |
+
for index, (duplicate_hash, _) in enumerate(self.hkey_record)
|
| 509 |
+
if duplicate_hash == hash
|
| 510 |
+
]
|
| 511 |
+
|
| 512 |
+
raise DuplicatedKeysError(key, duplicate_key_indices)
|
| 513 |
+
else:
|
| 514 |
+
tmp_record.add(hash)
|
| 515 |
+
|
| 516 |
+
def write_row(self, row: pa.Table, writer_batch_size: Optional[int] = None):
|
| 517 |
+
"""Add a given single-row Table to the write-pool of rows which is written to file.
|
| 518 |
+
|
| 519 |
+
Args:
|
| 520 |
+
row: the row to add.
|
| 521 |
+
"""
|
| 522 |
+
if len(row) != 1:
|
| 523 |
+
raise ValueError(f"Only single-row pyarrow tables are allowed but got table with {len(row)} rows.")
|
| 524 |
+
self.current_rows.append(row)
|
| 525 |
+
if writer_batch_size is None:
|
| 526 |
+
writer_batch_size = self.writer_batch_size
|
| 527 |
+
if writer_batch_size is not None and len(self.current_rows) >= writer_batch_size:
|
| 528 |
+
self.write_rows_on_file()
|
| 529 |
+
|
| 530 |
+
def write_batch(
|
| 531 |
+
self,
|
| 532 |
+
batch_examples: Dict[str, List],
|
| 533 |
+
writer_batch_size: Optional[int] = None,
|
| 534 |
+
):
|
| 535 |
+
"""Write a batch of Example to file.
|
| 536 |
+
Ignores the batch if it appears to be empty,
|
| 537 |
+
preventing a potential schema update of unknown types.
|
| 538 |
+
|
| 539 |
+
Args:
|
| 540 |
+
batch_examples: the batch of examples to add.
|
| 541 |
+
"""
|
| 542 |
+
if batch_examples and len(next(iter(batch_examples.values()))) == 0:
|
| 543 |
+
return
|
| 544 |
+
features = None if self.pa_writer is None and self.update_features else self._features
|
| 545 |
+
try_features = self._features if self.pa_writer is None and self.update_features else None
|
| 546 |
+
arrays = []
|
| 547 |
+
inferred_features = Features()
|
| 548 |
+
# preserve the order the columns
|
| 549 |
+
if self.schema:
|
| 550 |
+
schema_cols = set(self.schema.names)
|
| 551 |
+
batch_cols = batch_examples.keys() # .keys() preserves the order (unlike set)
|
| 552 |
+
common_cols = [col for col in self.schema.names if col in batch_cols]
|
| 553 |
+
extra_cols = [col for col in batch_cols if col not in schema_cols]
|
| 554 |
+
cols = common_cols + extra_cols
|
| 555 |
+
else:
|
| 556 |
+
cols = list(batch_examples)
|
| 557 |
+
for col in cols:
|
| 558 |
+
col_values = batch_examples[col]
|
| 559 |
+
col_type = features[col] if features else None
|
| 560 |
+
if isinstance(col_values, (pa.Array, pa.ChunkedArray)):
|
| 561 |
+
array = cast_array_to_feature(col_values, col_type) if col_type is not None else col_values
|
| 562 |
+
arrays.append(array)
|
| 563 |
+
inferred_features[col] = generate_from_arrow_type(col_values.type)
|
| 564 |
+
else:
|
| 565 |
+
col_try_type = try_features[col] if try_features is not None and col in try_features else None
|
| 566 |
+
typed_sequence = OptimizedTypedSequence(col_values, type=col_type, try_type=col_try_type, col=col)
|
| 567 |
+
arrays.append(pa.array(typed_sequence))
|
| 568 |
+
inferred_features[col] = typed_sequence.get_inferred_type()
|
| 569 |
+
schema = inferred_features.arrow_schema if self.pa_writer is None else self.schema
|
| 570 |
+
pa_table = pa.Table.from_arrays(arrays, schema=schema)
|
| 571 |
+
self.write_table(pa_table, writer_batch_size)
|
| 572 |
+
|
| 573 |
+
def write_table(self, pa_table: pa.Table, writer_batch_size: Optional[int] = None):
|
| 574 |
+
"""Write a Table to file.
|
| 575 |
+
|
| 576 |
+
Args:
|
| 577 |
+
example: the Table to add.
|
| 578 |
+
"""
|
| 579 |
+
if writer_batch_size is None:
|
| 580 |
+
writer_batch_size = self.writer_batch_size
|
| 581 |
+
if self.pa_writer is None:
|
| 582 |
+
self._build_writer(inferred_schema=pa_table.schema)
|
| 583 |
+
pa_table = pa_table.combine_chunks()
|
| 584 |
+
pa_table = table_cast(pa_table, self._schema)
|
| 585 |
+
if self.embed_local_files:
|
| 586 |
+
pa_table = embed_table_storage(pa_table)
|
| 587 |
+
self._num_bytes += pa_table.nbytes
|
| 588 |
+
self._num_examples += pa_table.num_rows
|
| 589 |
+
self.pa_writer.write_table(pa_table, writer_batch_size)
|
| 590 |
+
|
| 591 |
+
def finalize(self, close_stream=True):
|
| 592 |
+
self.write_rows_on_file()
|
| 593 |
+
# In case current_examples < writer_batch_size, but user uses finalize()
|
| 594 |
+
if self._check_duplicates:
|
| 595 |
+
self.check_duplicate_keys()
|
| 596 |
+
# Re-intializing to empty list for next batch
|
| 597 |
+
self.hkey_record = []
|
| 598 |
+
self.write_examples_on_file()
|
| 599 |
+
# If schema is known, infer features even if no examples were written
|
| 600 |
+
if self.pa_writer is None and self.schema:
|
| 601 |
+
self._build_writer(self.schema)
|
| 602 |
+
if self.pa_writer is not None:
|
| 603 |
+
self.pa_writer.close()
|
| 604 |
+
self.pa_writer = None
|
| 605 |
+
if close_stream:
|
| 606 |
+
self.stream.close()
|
| 607 |
+
else:
|
| 608 |
+
if close_stream:
|
| 609 |
+
self.stream.close()
|
| 610 |
+
raise SchemaInferenceError("Please pass `features` or at least one example when writing data")
|
| 611 |
+
logger.debug(
|
| 612 |
+
f"Done writing {self._num_examples} {self.unit} in {self._num_bytes} bytes {self._path if self._path else ''}."
|
| 613 |
+
)
|
| 614 |
+
return self._num_examples, self._num_bytes
|
| 615 |
+
|
| 616 |
+
|
| 617 |
+
class ParquetWriter(ArrowWriter):
|
| 618 |
+
_WRITER_CLASS = pq.ParquetWriter
|
| 619 |
+
|
| 620 |
+
|
| 621 |
+
class BeamWriter:
|
| 622 |
+
"""
|
| 623 |
+
Shuffles and writes Examples to Arrow files.
|
| 624 |
+
The Arrow files are converted from Parquet files that are the output of Apache Beam pipelines.
|
| 625 |
+
"""
|
| 626 |
+
|
| 627 |
+
def __init__(
|
| 628 |
+
self,
|
| 629 |
+
features: Optional[Features] = None,
|
| 630 |
+
schema: Optional[pa.Schema] = None,
|
| 631 |
+
path: Optional[str] = None,
|
| 632 |
+
namespace: Optional[str] = None,
|
| 633 |
+
cache_dir: Optional[str] = None,
|
| 634 |
+
):
|
| 635 |
+
if features is None and schema is None:
|
| 636 |
+
raise ValueError("At least one of features and schema must be provided.")
|
| 637 |
+
if path is None:
|
| 638 |
+
raise ValueError("Path must be provided.")
|
| 639 |
+
|
| 640 |
+
if features is not None:
|
| 641 |
+
self._features: Features = features
|
| 642 |
+
self._schema: pa.Schema = features.arrow_schema
|
| 643 |
+
else:
|
| 644 |
+
self._schema: pa.Schema = schema
|
| 645 |
+
self._features: Features = Features.from_arrow_schema(schema)
|
| 646 |
+
|
| 647 |
+
self._path = path
|
| 648 |
+
self._parquet_path = os.path.splitext(path)[0] # remove extension
|
| 649 |
+
self._namespace = namespace or "default"
|
| 650 |
+
self._num_examples = None
|
| 651 |
+
self._cache_dir = cache_dir or config.HF_DATASETS_CACHE
|
| 652 |
+
|
| 653 |
+
def write_from_pcollection(self, pcoll_examples):
|
| 654 |
+
"""Add the final steps of the beam pipeline: write to parquet files."""
|
| 655 |
+
import apache_beam as beam
|
| 656 |
+
|
| 657 |
+
def inc_num_examples(example):
|
| 658 |
+
beam.metrics.Metrics.counter(self._namespace, "num_examples").inc()
|
| 659 |
+
|
| 660 |
+
# count examples
|
| 661 |
+
_ = pcoll_examples | "Count N. Examples" >> beam.Map(inc_num_examples)
|
| 662 |
+
|
| 663 |
+
# save dataset
|
| 664 |
+
return (
|
| 665 |
+
pcoll_examples
|
| 666 |
+
| "Get values" >> beam.Values()
|
| 667 |
+
| "Save to parquet"
|
| 668 |
+
>> beam.io.parquetio.WriteToParquet(
|
| 669 |
+
self._parquet_path, self._schema, shard_name_template="-SSSSS-of-NNNNN.parquet"
|
| 670 |
+
)
|
| 671 |
+
)
|
| 672 |
+
|
| 673 |
+
def finalize(self, metrics_query_result: dict):
|
| 674 |
+
"""
|
| 675 |
+
Run after the pipeline has finished.
|
| 676 |
+
It converts the resulting parquet files to arrow and it completes the info from the pipeline metrics.
|
| 677 |
+
|
| 678 |
+
Args:
|
| 679 |
+
metrics_query_result: `dict` obtained from pipeline_results.metrics().query(m_filter). Make sure
|
| 680 |
+
that the filter keeps only the metrics for the considered split, under the namespace `split_name`.
|
| 681 |
+
"""
|
| 682 |
+
|
| 683 |
+
# Beam FileSystems require the system's path separator in the older versions
|
| 684 |
+
fs, _, [parquet_path] = fsspec.get_fs_token_paths(self._parquet_path)
|
| 685 |
+
parquet_path = str(Path(parquet_path)) if not is_remote_filesystem(fs) else fs.unstrip_protocol(parquet_path)
|
| 686 |
+
|
| 687 |
+
shards = fs.glob(parquet_path + "*.parquet")
|
| 688 |
+
num_bytes = sum(fs.sizes(shards))
|
| 689 |
+
shard_lengths = get_parquet_lengths(shards)
|
| 690 |
+
|
| 691 |
+
# Convert to arrow
|
| 692 |
+
if self._path.endswith(".arrow"):
|
| 693 |
+
logger.info(f"Converting parquet files {self._parquet_path} to arrow {self._path}")
|
| 694 |
+
try: # stream conversion
|
| 695 |
+
num_bytes = 0
|
| 696 |
+
for shard in hf_tqdm(shards, unit="shards"):
|
| 697 |
+
with fs.open(shard, "rb") as source:
|
| 698 |
+
with fs.open(shard.replace(".parquet", ".arrow"), "wb") as destination:
|
| 699 |
+
shard_num_bytes, _ = parquet_to_arrow(source, destination)
|
| 700 |
+
num_bytes += shard_num_bytes
|
| 701 |
+
except OSError as e: # broken pipe can happen if the connection is unstable, do local conversion instead
|
| 702 |
+
if e.errno != errno.EPIPE: # not a broken pipe
|
| 703 |
+
raise
|
| 704 |
+
logger.warning(
|
| 705 |
+
"Broken Pipe during stream conversion from parquet to arrow. Using local convert instead"
|
| 706 |
+
)
|
| 707 |
+
local_convert_dir = os.path.join(self._cache_dir, "beam_convert")
|
| 708 |
+
os.makedirs(local_convert_dir, exist_ok=True)
|
| 709 |
+
num_bytes = 0
|
| 710 |
+
for shard in hf_tqdm(shards, unit="shards"):
|
| 711 |
+
local_parquet_path = os.path.join(local_convert_dir, hash_url_to_filename(shard) + ".parquet")
|
| 712 |
+
fs.download(shard, local_parquet_path)
|
| 713 |
+
local_arrow_path = local_parquet_path.replace(".parquet", ".arrow")
|
| 714 |
+
shard_num_bytes, _ = parquet_to_arrow(local_parquet_path, local_arrow_path)
|
| 715 |
+
num_bytes += shard_num_bytes
|
| 716 |
+
remote_arrow_path = shard.replace(".parquet", ".arrow")
|
| 717 |
+
fs.upload(local_arrow_path, remote_arrow_path)
|
| 718 |
+
|
| 719 |
+
# Save metrics
|
| 720 |
+
counters_dict = {metric.key.metric.name: metric.result for metric in metrics_query_result["counters"]}
|
| 721 |
+
self._num_examples = counters_dict["num_examples"]
|
| 722 |
+
self._num_bytes = num_bytes
|
| 723 |
+
self._shard_lengths = shard_lengths
|
| 724 |
+
return self._num_examples, self._num_bytes
|
| 725 |
+
|
| 726 |
+
|
| 727 |
+
def get_parquet_lengths(sources) -> List[int]:
|
| 728 |
+
shard_lengths = []
|
| 729 |
+
for source in hf_tqdm(sources, unit="parquet files"):
|
| 730 |
+
parquet_file = pa.parquet.ParquetFile(source)
|
| 731 |
+
shard_lengths.append(parquet_file.metadata.num_rows)
|
| 732 |
+
return shard_lengths
|
| 733 |
+
|
| 734 |
+
|
| 735 |
+
def parquet_to_arrow(source, destination) -> List[int]:
|
| 736 |
+
"""Convert parquet file to arrow file. Inputs can be str paths or file-like objects"""
|
| 737 |
+
stream = None if isinstance(destination, str) else destination
|
| 738 |
+
parquet_file = pa.parquet.ParquetFile(source)
|
| 739 |
+
# Beam can create empty Parquet files, so we need to pass the source Parquet file's schema
|
| 740 |
+
with ArrowWriter(schema=parquet_file.schema_arrow, path=destination, stream=stream) as writer:
|
| 741 |
+
for record_batch in parquet_file.iter_batches():
|
| 742 |
+
pa_table = pa.Table.from_batches([record_batch])
|
| 743 |
+
writer.write_table(pa_table)
|
| 744 |
+
num_bytes, num_examples = writer.finalize()
|
| 745 |
+
return num_bytes, num_examples
|
evalkit_tf437/lib/python3.10/site-packages/datasets/builder.py
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
evalkit_tf437/lib/python3.10/site-packages/datasets/config.py
ADDED
|
@@ -0,0 +1,259 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import importlib
|
| 2 |
+
import importlib.metadata
|
| 3 |
+
import logging
|
| 4 |
+
import os
|
| 5 |
+
import platform
|
| 6 |
+
from pathlib import Path
|
| 7 |
+
from typing import Optional
|
| 8 |
+
|
| 9 |
+
from packaging import version
|
| 10 |
+
|
| 11 |
+
|
| 12 |
+
logger = logging.getLogger(__name__.split(".", 1)[0]) # to avoid circular import from .utils.logging
|
| 13 |
+
|
| 14 |
+
# Datasets
|
| 15 |
+
S3_DATASETS_BUCKET_PREFIX = "https://s3.amazonaws.com/datasets.huggingface.co/datasets/datasets"
|
| 16 |
+
CLOUDFRONT_DATASETS_DISTRIB_PREFIX = "https://cdn-datasets.huggingface.co/datasets/datasets"
|
| 17 |
+
REPO_DATASETS_URL = "https://raw.githubusercontent.com/huggingface/datasets/{revision}/datasets/{path}/{name}"
|
| 18 |
+
|
| 19 |
+
# Metrics
|
| 20 |
+
S3_METRICS_BUCKET_PREFIX = "https://s3.amazonaws.com/datasets.huggingface.co/datasets/metrics"
|
| 21 |
+
CLOUDFRONT_METRICS_DISTRIB_PREFIX = "https://cdn-datasets.huggingface.co/datasets/metric"
|
| 22 |
+
REPO_METRICS_URL = "https://raw.githubusercontent.com/huggingface/datasets/{revision}/metrics/{path}/{name}"
|
| 23 |
+
|
| 24 |
+
# Hub
|
| 25 |
+
HF_ENDPOINT = os.environ.get("HF_ENDPOINT", "https://huggingface.co")
|
| 26 |
+
HUB_DATASETS_URL = HF_ENDPOINT + "/datasets/{repo_id}/resolve/{revision}/{path}"
|
| 27 |
+
HUB_DATASETS_HFFS_URL = "hf://datasets/{repo_id}@{revision}/{path}"
|
| 28 |
+
HUB_DEFAULT_VERSION = "main"
|
| 29 |
+
|
| 30 |
+
PY_VERSION = version.parse(platform.python_version())
|
| 31 |
+
|
| 32 |
+
# General environment variables accepted values for booleans
|
| 33 |
+
ENV_VARS_TRUE_VALUES = {"1", "ON", "YES", "TRUE"}
|
| 34 |
+
ENV_VARS_FALSE_VALUES = {"0", "OFF", "NO", "FALSE"}
|
| 35 |
+
ENV_VARS_TRUE_AND_AUTO_VALUES = ENV_VARS_TRUE_VALUES.union({"AUTO"})
|
| 36 |
+
ENV_VARS_FALSE_AND_AUTO_VALUES = ENV_VARS_FALSE_VALUES.union({"AUTO"})
|
| 37 |
+
|
| 38 |
+
|
| 39 |
+
# Imports
|
| 40 |
+
DILL_VERSION = version.parse(importlib.metadata.version("dill"))
|
| 41 |
+
FSSPEC_VERSION = version.parse(importlib.metadata.version("fsspec"))
|
| 42 |
+
PANDAS_VERSION = version.parse(importlib.metadata.version("pandas"))
|
| 43 |
+
PYARROW_VERSION = version.parse(importlib.metadata.version("pyarrow"))
|
| 44 |
+
HF_HUB_VERSION = version.parse(importlib.metadata.version("huggingface_hub"))
|
| 45 |
+
|
| 46 |
+
USE_TF = os.environ.get("USE_TF", "AUTO").upper()
|
| 47 |
+
USE_TORCH = os.environ.get("USE_TORCH", "AUTO").upper()
|
| 48 |
+
USE_JAX = os.environ.get("USE_JAX", "AUTO").upper()
|
| 49 |
+
|
| 50 |
+
TORCH_VERSION = "N/A"
|
| 51 |
+
TORCH_AVAILABLE = False
|
| 52 |
+
|
| 53 |
+
if USE_TORCH in ENV_VARS_TRUE_AND_AUTO_VALUES and USE_TF not in ENV_VARS_TRUE_VALUES:
|
| 54 |
+
TORCH_AVAILABLE = importlib.util.find_spec("torch") is not None
|
| 55 |
+
if TORCH_AVAILABLE:
|
| 56 |
+
try:
|
| 57 |
+
TORCH_VERSION = version.parse(importlib.metadata.version("torch"))
|
| 58 |
+
logger.info(f"PyTorch version {TORCH_VERSION} available.")
|
| 59 |
+
except importlib.metadata.PackageNotFoundError:
|
| 60 |
+
pass
|
| 61 |
+
else:
|
| 62 |
+
logger.info("Disabling PyTorch because USE_TF is set")
|
| 63 |
+
|
| 64 |
+
TF_VERSION = "N/A"
|
| 65 |
+
TF_AVAILABLE = False
|
| 66 |
+
|
| 67 |
+
if USE_TF in ENV_VARS_TRUE_AND_AUTO_VALUES and USE_TORCH not in ENV_VARS_TRUE_VALUES:
|
| 68 |
+
TF_AVAILABLE = importlib.util.find_spec("tensorflow") is not None
|
| 69 |
+
if TF_AVAILABLE:
|
| 70 |
+
# For the metadata, we have to look for both tensorflow and tensorflow-cpu
|
| 71 |
+
for package in [
|
| 72 |
+
"tensorflow",
|
| 73 |
+
"tensorflow-cpu",
|
| 74 |
+
"tensorflow-gpu",
|
| 75 |
+
"tf-nightly",
|
| 76 |
+
"tf-nightly-cpu",
|
| 77 |
+
"tf-nightly-gpu",
|
| 78 |
+
"intel-tensorflow",
|
| 79 |
+
"tensorflow-rocm",
|
| 80 |
+
"tensorflow-macos",
|
| 81 |
+
]:
|
| 82 |
+
try:
|
| 83 |
+
TF_VERSION = version.parse(importlib.metadata.version(package))
|
| 84 |
+
except importlib.metadata.PackageNotFoundError:
|
| 85 |
+
continue
|
| 86 |
+
else:
|
| 87 |
+
break
|
| 88 |
+
else:
|
| 89 |
+
TF_AVAILABLE = False
|
| 90 |
+
if TF_AVAILABLE:
|
| 91 |
+
if TF_VERSION.major < 2:
|
| 92 |
+
logger.info(f"TensorFlow found but with version {TF_VERSION}. `datasets` requires version 2 minimum.")
|
| 93 |
+
TF_AVAILABLE = False
|
| 94 |
+
else:
|
| 95 |
+
logger.info(f"TensorFlow version {TF_VERSION} available.")
|
| 96 |
+
else:
|
| 97 |
+
logger.info("Disabling Tensorflow because USE_TORCH is set")
|
| 98 |
+
|
| 99 |
+
|
| 100 |
+
JAX_VERSION = "N/A"
|
| 101 |
+
JAX_AVAILABLE = False
|
| 102 |
+
|
| 103 |
+
if USE_JAX in ENV_VARS_TRUE_AND_AUTO_VALUES:
|
| 104 |
+
JAX_AVAILABLE = importlib.util.find_spec("jax") is not None and importlib.util.find_spec("jaxlib") is not None
|
| 105 |
+
if JAX_AVAILABLE:
|
| 106 |
+
try:
|
| 107 |
+
JAX_VERSION = version.parse(importlib.metadata.version("jax"))
|
| 108 |
+
logger.info(f"JAX version {JAX_VERSION} available.")
|
| 109 |
+
except importlib.metadata.PackageNotFoundError:
|
| 110 |
+
pass
|
| 111 |
+
else:
|
| 112 |
+
logger.info("Disabling JAX because USE_JAX is set to False")
|
| 113 |
+
|
| 114 |
+
|
| 115 |
+
USE_BEAM = os.environ.get("USE_BEAM", "AUTO").upper()
|
| 116 |
+
BEAM_VERSION = "N/A"
|
| 117 |
+
BEAM_AVAILABLE = False
|
| 118 |
+
if USE_BEAM in ENV_VARS_TRUE_AND_AUTO_VALUES:
|
| 119 |
+
try:
|
| 120 |
+
BEAM_VERSION = version.parse(importlib.metadata.version("apache_beam"))
|
| 121 |
+
BEAM_AVAILABLE = True
|
| 122 |
+
logger.info(f"Apache Beam version {BEAM_VERSION} available.")
|
| 123 |
+
except importlib.metadata.PackageNotFoundError:
|
| 124 |
+
pass
|
| 125 |
+
else:
|
| 126 |
+
logger.info("Disabling Apache Beam because USE_BEAM is set to False")
|
| 127 |
+
|
| 128 |
+
|
| 129 |
+
# Optional tools for data loading
|
| 130 |
+
SQLALCHEMY_AVAILABLE = importlib.util.find_spec("sqlalchemy") is not None
|
| 131 |
+
|
| 132 |
+
# Optional tools for feature decoding
|
| 133 |
+
PIL_AVAILABLE = importlib.util.find_spec("PIL") is not None
|
| 134 |
+
IS_OPUS_SUPPORTED = importlib.util.find_spec("soundfile") is not None and version.parse(
|
| 135 |
+
importlib.import_module("soundfile").__libsndfile_version__
|
| 136 |
+
) >= version.parse("1.0.31")
|
| 137 |
+
IS_MP3_SUPPORTED = importlib.util.find_spec("soundfile") is not None and version.parse(
|
| 138 |
+
importlib.import_module("soundfile").__libsndfile_version__
|
| 139 |
+
) >= version.parse("1.1.0")
|
| 140 |
+
|
| 141 |
+
# Optional compression tools
|
| 142 |
+
RARFILE_AVAILABLE = importlib.util.find_spec("rarfile") is not None
|
| 143 |
+
ZSTANDARD_AVAILABLE = importlib.util.find_spec("zstandard") is not None
|
| 144 |
+
LZ4_AVAILABLE = importlib.util.find_spec("lz4") is not None
|
| 145 |
+
PY7ZR_AVAILABLE = importlib.util.find_spec("py7zr") is not None
|
| 146 |
+
|
| 147 |
+
# Cache location
|
| 148 |
+
DEFAULT_XDG_CACHE_HOME = "~/.cache"
|
| 149 |
+
XDG_CACHE_HOME = os.getenv("XDG_CACHE_HOME", DEFAULT_XDG_CACHE_HOME)
|
| 150 |
+
DEFAULT_HF_CACHE_HOME = os.path.join(XDG_CACHE_HOME, "huggingface")
|
| 151 |
+
HF_CACHE_HOME = os.path.expanduser(os.getenv("HF_HOME", DEFAULT_HF_CACHE_HOME))
|
| 152 |
+
|
| 153 |
+
DEFAULT_HF_DATASETS_CACHE = os.path.join(HF_CACHE_HOME, "datasets")
|
| 154 |
+
HF_DATASETS_CACHE = Path(os.getenv("HF_DATASETS_CACHE", DEFAULT_HF_DATASETS_CACHE))
|
| 155 |
+
|
| 156 |
+
DEFAULT_HF_METRICS_CACHE = os.path.join(HF_CACHE_HOME, "metrics")
|
| 157 |
+
HF_METRICS_CACHE = Path(os.getenv("HF_METRICS_CACHE", DEFAULT_HF_METRICS_CACHE))
|
| 158 |
+
|
| 159 |
+
DEFAULT_HF_MODULES_CACHE = os.path.join(HF_CACHE_HOME, "modules")
|
| 160 |
+
HF_MODULES_CACHE = Path(os.getenv("HF_MODULES_CACHE", DEFAULT_HF_MODULES_CACHE))
|
| 161 |
+
|
| 162 |
+
DOWNLOADED_DATASETS_DIR = "downloads"
|
| 163 |
+
DEFAULT_DOWNLOADED_DATASETS_PATH = os.path.join(HF_DATASETS_CACHE, DOWNLOADED_DATASETS_DIR)
|
| 164 |
+
DOWNLOADED_DATASETS_PATH = Path(os.getenv("HF_DATASETS_DOWNLOADED_DATASETS_PATH", DEFAULT_DOWNLOADED_DATASETS_PATH))
|
| 165 |
+
|
| 166 |
+
EXTRACTED_DATASETS_DIR = "extracted"
|
| 167 |
+
DEFAULT_EXTRACTED_DATASETS_PATH = os.path.join(DEFAULT_DOWNLOADED_DATASETS_PATH, EXTRACTED_DATASETS_DIR)
|
| 168 |
+
EXTRACTED_DATASETS_PATH = Path(os.getenv("HF_DATASETS_EXTRACTED_DATASETS_PATH", DEFAULT_EXTRACTED_DATASETS_PATH))
|
| 169 |
+
|
| 170 |
+
# Download count for the website
|
| 171 |
+
HF_UPDATE_DOWNLOAD_COUNTS = (
|
| 172 |
+
os.environ.get("HF_UPDATE_DOWNLOAD_COUNTS", "AUTO").upper() in ENV_VARS_TRUE_AND_AUTO_VALUES
|
| 173 |
+
)
|
| 174 |
+
|
| 175 |
+
# Remote dataset scripts support
|
| 176 |
+
__HF_DATASETS_TRUST_REMOTE_CODE = os.environ.get("HF_DATASETS_TRUST_REMOTE_CODE", "1")
|
| 177 |
+
HF_DATASETS_TRUST_REMOTE_CODE: Optional[bool] = (
|
| 178 |
+
True
|
| 179 |
+
if __HF_DATASETS_TRUST_REMOTE_CODE.upper() in ENV_VARS_TRUE_VALUES
|
| 180 |
+
else False
|
| 181 |
+
if __HF_DATASETS_TRUST_REMOTE_CODE.upper() in ENV_VARS_FALSE_VALUES
|
| 182 |
+
else None
|
| 183 |
+
)
|
| 184 |
+
TIME_OUT_REMOTE_CODE = 15
|
| 185 |
+
|
| 186 |
+
# Datasets-server
|
| 187 |
+
USE_PARQUET_EXPORT = True
|
| 188 |
+
|
| 189 |
+
# Batch size constants. For more info, see:
|
| 190 |
+
# https://github.com/apache/arrow/blob/master/docs/source/cpp/arrays.rst#size-limitations-and-recommendations)
|
| 191 |
+
DEFAULT_MAX_BATCH_SIZE = 1000
|
| 192 |
+
|
| 193 |
+
# Size of the preloaded record batch in `Dataset.__iter__`
|
| 194 |
+
ARROW_READER_BATCH_SIZE_IN_DATASET_ITER = 10
|
| 195 |
+
|
| 196 |
+
# Max shard size in bytes (e.g. to shard parquet datasets in push_to_hub or download_and_prepare)
|
| 197 |
+
MAX_SHARD_SIZE = "500MB"
|
| 198 |
+
|
| 199 |
+
# Parquet configuration
|
| 200 |
+
PARQUET_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS = 100
|
| 201 |
+
PARQUET_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS = 100
|
| 202 |
+
PARQUET_ROW_GROUP_SIZE_FOR_BINARY_DATASETS = 100
|
| 203 |
+
|
| 204 |
+
# Offline mode
|
| 205 |
+
HF_DATASETS_OFFLINE = os.environ.get("HF_DATASETS_OFFLINE", "AUTO").upper() in ENV_VARS_TRUE_VALUES
|
| 206 |
+
|
| 207 |
+
# Here, `True` will disable progress bars globally without possibility of enabling it
|
| 208 |
+
# programmatically. `False` will enable them without possibility of disabling them.
|
| 209 |
+
# If environment variable is not set (None), then the user is free to enable/disable
|
| 210 |
+
# them programmatically.
|
| 211 |
+
# TL;DR: env variable has priority over code
|
| 212 |
+
__HF_DATASETS_DISABLE_PROGRESS_BARS = os.environ.get("HF_DATASETS_DISABLE_PROGRESS_BARS")
|
| 213 |
+
HF_DATASETS_DISABLE_PROGRESS_BARS: Optional[bool] = (
|
| 214 |
+
__HF_DATASETS_DISABLE_PROGRESS_BARS.upper() in ENV_VARS_TRUE_VALUES
|
| 215 |
+
if __HF_DATASETS_DISABLE_PROGRESS_BARS is not None
|
| 216 |
+
else None
|
| 217 |
+
)
|
| 218 |
+
|
| 219 |
+
# In-memory
|
| 220 |
+
DEFAULT_IN_MEMORY_MAX_SIZE = 0 # Disabled
|
| 221 |
+
IN_MEMORY_MAX_SIZE = float(os.environ.get("HF_DATASETS_IN_MEMORY_MAX_SIZE", DEFAULT_IN_MEMORY_MAX_SIZE))
|
| 222 |
+
|
| 223 |
+
# File names
|
| 224 |
+
DATASET_ARROW_FILENAME = "dataset.arrow"
|
| 225 |
+
DATASET_INDICES_FILENAME = "indices.arrow"
|
| 226 |
+
DATASET_STATE_JSON_FILENAME = "state.json"
|
| 227 |
+
DATASET_INFO_FILENAME = "dataset_info.json"
|
| 228 |
+
DATASETDICT_INFOS_FILENAME = "dataset_infos.json"
|
| 229 |
+
LICENSE_FILENAME = "LICENSE"
|
| 230 |
+
METRIC_INFO_FILENAME = "metric_info.json"
|
| 231 |
+
DATASETDICT_JSON_FILENAME = "dataset_dict.json"
|
| 232 |
+
METADATA_CONFIGS_FIELD = "configs"
|
| 233 |
+
REPOCARD_FILENAME = "README.md"
|
| 234 |
+
REPOYAML_FILENAME = ".huggingface.yaml"
|
| 235 |
+
|
| 236 |
+
MODULE_NAME_FOR_DYNAMIC_MODULES = "datasets_modules"
|
| 237 |
+
|
| 238 |
+
MAX_DATASET_CONFIG_ID_READABLE_LENGTH = 255
|
| 239 |
+
|
| 240 |
+
# Temporary cache directory prefix
|
| 241 |
+
TEMP_CACHE_DIR_PREFIX = "hf_datasets-"
|
| 242 |
+
|
| 243 |
+
# Streaming
|
| 244 |
+
STREAMING_READ_MAX_RETRIES = 20
|
| 245 |
+
STREAMING_READ_RETRY_INTERVAL = 5
|
| 246 |
+
|
| 247 |
+
# Datasets without script
|
| 248 |
+
DATA_FILES_MAX_NUMBER_FOR_MODULE_INFERENCE = 200
|
| 249 |
+
GLOBBED_DATA_FILES_MAX_NUMBER_FOR_MODULE_INFERENCE = 10
|
| 250 |
+
ARCHIVED_DATA_FILES_MAX_NUMBER_FOR_MODULE_INFERENCE = 200
|
| 251 |
+
|
| 252 |
+
# Progress bars
|
| 253 |
+
PBAR_REFRESH_TIME_INTERVAL = 0.05 # 20 progress updates per sec
|
| 254 |
+
|
| 255 |
+
# Maximum number of uploaded files per commit
|
| 256 |
+
UPLOADS_MAX_NUMBER_PER_COMMIT = 50
|
| 257 |
+
|
| 258 |
+
# Backward compatibiliy
|
| 259 |
+
MAX_TABLE_NBYTES_FOR_PICKLING = 4 << 30
|
evalkit_tf437/lib/python3.10/site-packages/datasets/data_files.py
ADDED
|
@@ -0,0 +1,806 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import os
|
| 2 |
+
import re
|
| 3 |
+
from functools import partial
|
| 4 |
+
from glob import has_magic
|
| 5 |
+
from pathlib import Path, PurePath
|
| 6 |
+
from typing import Callable, Dict, List, Optional, Set, Tuple, Union
|
| 7 |
+
|
| 8 |
+
import huggingface_hub
|
| 9 |
+
from fsspec import get_fs_token_paths
|
| 10 |
+
from fsspec.implementations.http import HTTPFileSystem
|
| 11 |
+
from huggingface_hub import HfFileSystem
|
| 12 |
+
from packaging import version
|
| 13 |
+
from tqdm.contrib.concurrent import thread_map
|
| 14 |
+
|
| 15 |
+
from . import config
|
| 16 |
+
from .download import DownloadConfig
|
| 17 |
+
from .download.streaming_download_manager import _prepare_path_and_storage_options, xbasename, xjoin
|
| 18 |
+
from .naming import _split_re
|
| 19 |
+
from .splits import Split
|
| 20 |
+
from .utils import logging
|
| 21 |
+
from .utils import tqdm as hf_tqdm
|
| 22 |
+
from .utils.file_utils import is_local_path, is_relative_path
|
| 23 |
+
from .utils.py_utils import glob_pattern_to_regex, string_to_dict
|
| 24 |
+
|
| 25 |
+
|
| 26 |
+
SANITIZED_DEFAULT_SPLIT = str(Split.TRAIN)
|
| 27 |
+
|
| 28 |
+
|
| 29 |
+
logger = logging.get_logger(__name__)
|
| 30 |
+
|
| 31 |
+
|
| 32 |
+
class Url(str):
|
| 33 |
+
pass
|
| 34 |
+
|
| 35 |
+
|
| 36 |
+
class EmptyDatasetError(FileNotFoundError):
|
| 37 |
+
pass
|
| 38 |
+
|
| 39 |
+
|
| 40 |
+
SPLIT_PATTERN_SHARDED = "data/{split}-[0-9][0-9][0-9][0-9][0-9]-of-[0-9][0-9][0-9][0-9][0-9]*.*"
|
| 41 |
+
|
| 42 |
+
SPLIT_KEYWORDS = {
|
| 43 |
+
Split.TRAIN: ["train", "training"],
|
| 44 |
+
Split.VALIDATION: ["validation", "valid", "dev", "val"],
|
| 45 |
+
Split.TEST: ["test", "testing", "eval", "evaluation"],
|
| 46 |
+
}
|
| 47 |
+
NON_WORDS_CHARS = "-._ 0-9"
|
| 48 |
+
if config.FSSPEC_VERSION < version.parse("2023.9.0"):
|
| 49 |
+
KEYWORDS_IN_PATH_NAME_BASE_PATTERNS = ["{keyword}[{sep}/]**", "**[{sep}/]{keyword}[{sep}/]**"]
|
| 50 |
+
elif config.FSSPEC_VERSION < version.parse("2023.12.0"):
|
| 51 |
+
KEYWORDS_IN_PATH_NAME_BASE_PATTERNS = ["{keyword}[{sep}/]**", "**/*[{sep}/]{keyword}[{sep}/]**"]
|
| 52 |
+
else:
|
| 53 |
+
KEYWORDS_IN_PATH_NAME_BASE_PATTERNS = [
|
| 54 |
+
"**/{keyword}[{sep}]*",
|
| 55 |
+
"**/{keyword}/**",
|
| 56 |
+
"**/*[{sep}]{keyword}[{sep}]*",
|
| 57 |
+
"**/*[{sep}]{keyword}[{sep}]*/**",
|
| 58 |
+
"**/{keyword}[{sep}]*/**",
|
| 59 |
+
"**/*[{sep}]{keyword}/**",
|
| 60 |
+
]
|
| 61 |
+
|
| 62 |
+
DEFAULT_SPLITS = [Split.TRAIN, Split.VALIDATION, Split.TEST]
|
| 63 |
+
DEFAULT_PATTERNS_SPLIT_IN_PATH_NAME = {
|
| 64 |
+
split: [
|
| 65 |
+
pattern.format(keyword=keyword, sep=NON_WORDS_CHARS)
|
| 66 |
+
for keyword in SPLIT_KEYWORDS[split]
|
| 67 |
+
for pattern in KEYWORDS_IN_PATH_NAME_BASE_PATTERNS
|
| 68 |
+
]
|
| 69 |
+
for split in DEFAULT_SPLITS
|
| 70 |
+
}
|
| 71 |
+
|
| 72 |
+
DEFAULT_PATTERNS_ALL = {
|
| 73 |
+
Split.TRAIN: ["**"],
|
| 74 |
+
}
|
| 75 |
+
|
| 76 |
+
ALL_SPLIT_PATTERNS = [SPLIT_PATTERN_SHARDED]
|
| 77 |
+
ALL_DEFAULT_PATTERNS = [
|
| 78 |
+
DEFAULT_PATTERNS_SPLIT_IN_PATH_NAME,
|
| 79 |
+
DEFAULT_PATTERNS_ALL,
|
| 80 |
+
]
|
| 81 |
+
if config.FSSPEC_VERSION < version.parse("2023.9.0"):
|
| 82 |
+
METADATA_PATTERNS = [
|
| 83 |
+
"metadata.csv",
|
| 84 |
+
"**/metadata.csv",
|
| 85 |
+
"metadata.jsonl",
|
| 86 |
+
"**/metadata.jsonl",
|
| 87 |
+
] # metadata file for ImageFolder and AudioFolder
|
| 88 |
+
else:
|
| 89 |
+
METADATA_PATTERNS = [
|
| 90 |
+
"**/metadata.csv",
|
| 91 |
+
"**/metadata.jsonl",
|
| 92 |
+
] # metadata file for ImageFolder and AudioFolder
|
| 93 |
+
WILDCARD_CHARACTERS = "*[]"
|
| 94 |
+
FILES_TO_IGNORE = [
|
| 95 |
+
"README.md",
|
| 96 |
+
"config.json",
|
| 97 |
+
"dataset_info.json",
|
| 98 |
+
"dataset_infos.json",
|
| 99 |
+
"dummy_data.zip",
|
| 100 |
+
"dataset_dict.json",
|
| 101 |
+
]
|
| 102 |
+
|
| 103 |
+
|
| 104 |
+
def contains_wildcards(pattern: str) -> bool:
|
| 105 |
+
return any(wilcard_character in pattern for wilcard_character in WILDCARD_CHARACTERS)
|
| 106 |
+
|
| 107 |
+
|
| 108 |
+
def sanitize_patterns(patterns: Union[Dict, List, str]) -> Dict[str, Union[List[str], "DataFilesList"]]:
|
| 109 |
+
"""
|
| 110 |
+
Take the data_files patterns from the user, and format them into a dictionary.
|
| 111 |
+
Each key is the name of the split, and each value is a list of data files patterns (paths or urls).
|
| 112 |
+
The default split is "train".
|
| 113 |
+
|
| 114 |
+
Returns:
|
| 115 |
+
patterns: dictionary of split_name -> list of patterns
|
| 116 |
+
"""
|
| 117 |
+
if isinstance(patterns, dict):
|
| 118 |
+
return {str(key): value if isinstance(value, list) else [value] for key, value in patterns.items()}
|
| 119 |
+
elif isinstance(patterns, str):
|
| 120 |
+
return {SANITIZED_DEFAULT_SPLIT: [patterns]}
|
| 121 |
+
elif isinstance(patterns, list):
|
| 122 |
+
if any(isinstance(pattern, dict) for pattern in patterns):
|
| 123 |
+
for pattern in patterns:
|
| 124 |
+
if not (
|
| 125 |
+
isinstance(pattern, dict)
|
| 126 |
+
and len(pattern) == 2
|
| 127 |
+
and "split" in pattern
|
| 128 |
+
and isinstance(pattern.get("path"), (str, list))
|
| 129 |
+
):
|
| 130 |
+
raise ValueError(
|
| 131 |
+
f"Expected each split to have a 'path' key which can be a string or a list of strings, but got {pattern}"
|
| 132 |
+
)
|
| 133 |
+
splits = [pattern["split"] for pattern in patterns]
|
| 134 |
+
if len(set(splits)) != len(splits):
|
| 135 |
+
raise ValueError(f"Some splits are duplicated in data_files: {splits}")
|
| 136 |
+
return {
|
| 137 |
+
str(pattern["split"]): pattern["path"] if isinstance(pattern["path"], list) else [pattern["path"]]
|
| 138 |
+
for pattern in patterns
|
| 139 |
+
}
|
| 140 |
+
else:
|
| 141 |
+
return {SANITIZED_DEFAULT_SPLIT: patterns}
|
| 142 |
+
else:
|
| 143 |
+
return sanitize_patterns(list(patterns))
|
| 144 |
+
|
| 145 |
+
|
| 146 |
+
def _is_inside_unrequested_special_dir(matched_rel_path: str, pattern: str) -> bool:
|
| 147 |
+
"""
|
| 148 |
+
When a path matches a pattern, we additionnally check if it's inside a special directory
|
| 149 |
+
we ignore by default (if it starts with a double underscore).
|
| 150 |
+
|
| 151 |
+
Users can still explicitly request a filepath inside such a directory if "__pycache__" is
|
| 152 |
+
mentioned explicitly in the requested pattern.
|
| 153 |
+
|
| 154 |
+
Some examples:
|
| 155 |
+
|
| 156 |
+
base directory:
|
| 157 |
+
|
| 158 |
+
./
|
| 159 |
+
βββ __pycache__
|
| 160 |
+
βββ b.txt
|
| 161 |
+
|
| 162 |
+
>>> _is_inside_unrequested_special_dir("__pycache__/b.txt", "**")
|
| 163 |
+
True
|
| 164 |
+
>>> _is_inside_unrequested_special_dir("__pycache__/b.txt", "*/b.txt")
|
| 165 |
+
True
|
| 166 |
+
>>> _is_inside_unrequested_special_dir("__pycache__/b.txt", "__pycache__/*")
|
| 167 |
+
False
|
| 168 |
+
>>> _is_inside_unrequested_special_dir("__pycache__/b.txt", "__*/*")
|
| 169 |
+
False
|
| 170 |
+
"""
|
| 171 |
+
# We just need to check if every special directories from the path is present explicly in the pattern.
|
| 172 |
+
# Since we assume that the path matches the pattern, it's equivalent to counting that both
|
| 173 |
+
# the parent path and the parent pattern have the same number of special directories.
|
| 174 |
+
data_dirs_to_ignore_in_path = [part for part in PurePath(matched_rel_path).parent.parts if part.startswith("__")]
|
| 175 |
+
data_dirs_to_ignore_in_pattern = [part for part in PurePath(pattern).parent.parts if part.startswith("__")]
|
| 176 |
+
return len(data_dirs_to_ignore_in_path) != len(data_dirs_to_ignore_in_pattern)
|
| 177 |
+
|
| 178 |
+
|
| 179 |
+
def _is_unrequested_hidden_file_or_is_inside_unrequested_hidden_dir(matched_rel_path: str, pattern: str) -> bool:
|
| 180 |
+
"""
|
| 181 |
+
When a path matches a pattern, we additionnally check if it's a hidden file or if it's inside
|
| 182 |
+
a hidden directory we ignore by default, i.e. if the file name or a parent directory name starts with a dot.
|
| 183 |
+
|
| 184 |
+
Users can still explicitly request a filepath that is hidden or is inside a hidden directory
|
| 185 |
+
if the hidden part is mentioned explicitly in the requested pattern.
|
| 186 |
+
|
| 187 |
+
Some examples:
|
| 188 |
+
|
| 189 |
+
base directory:
|
| 190 |
+
|
| 191 |
+
./
|
| 192 |
+
βββ .hidden_file.txt
|
| 193 |
+
|
| 194 |
+
>>> _is_unrequested_hidden_file_or_is_inside_unrequested_hidden_dir(".hidden_file.txt", "**")
|
| 195 |
+
True
|
| 196 |
+
>>> _is_unrequested_hidden_file_or_is_inside_unrequested_hidden_dir(".hidden_file.txt", ".*")
|
| 197 |
+
False
|
| 198 |
+
|
| 199 |
+
base directory:
|
| 200 |
+
|
| 201 |
+
./
|
| 202 |
+
βββ .hidden_dir
|
| 203 |
+
βββ a.txt
|
| 204 |
+
|
| 205 |
+
>>> _is_unrequested_hidden_file_or_is_inside_unrequested_hidden_dir(".hidden_dir/a.txt", "**")
|
| 206 |
+
True
|
| 207 |
+
>>> _is_unrequested_hidden_file_or_is_inside_unrequested_hidden_dir(".hidden_dir/a.txt", ".*/*")
|
| 208 |
+
False
|
| 209 |
+
>>> _is_unrequested_hidden_file_or_is_inside_unrequested_hidden_dir(".hidden_dir/a.txt", ".hidden_dir/*")
|
| 210 |
+
False
|
| 211 |
+
|
| 212 |
+
base directory:
|
| 213 |
+
|
| 214 |
+
./
|
| 215 |
+
βββ .hidden_dir
|
| 216 |
+
βββ .hidden_file.txt
|
| 217 |
+
|
| 218 |
+
>>> _is_unrequested_hidden_file_or_is_inside_unrequested_hidden_dir(".hidden_dir/.hidden_file.txt", "**")
|
| 219 |
+
True
|
| 220 |
+
>>> _is_unrequested_hidden_file_or_is_inside_unrequested_hidden_dir(".hidden_dir/.hidden_file.txt", ".*/*")
|
| 221 |
+
True
|
| 222 |
+
>>> _is_unrequested_hidden_file_or_is_inside_unrequested_hidden_dir(".hidden_dir/.hidden_file.txt", ".*/.*")
|
| 223 |
+
False
|
| 224 |
+
>>> _is_unrequested_hidden_file_or_is_inside_unrequested_hidden_dir(".hidden_dir/.hidden_file.txt", ".hidden_dir/*")
|
| 225 |
+
True
|
| 226 |
+
>>> _is_unrequested_hidden_file_or_is_inside_unrequested_hidden_dir(".hidden_dir/.hidden_file.txt", ".hidden_dir/.*")
|
| 227 |
+
False
|
| 228 |
+
"""
|
| 229 |
+
# We just need to check if every hidden part from the path is present explicly in the pattern.
|
| 230 |
+
# Since we assume that the path matches the pattern, it's equivalent to counting that both
|
| 231 |
+
# the path and the pattern have the same number of hidden parts.
|
| 232 |
+
hidden_directories_in_path = [
|
| 233 |
+
part for part in PurePath(matched_rel_path).parts if part.startswith(".") and not set(part) == {"."}
|
| 234 |
+
]
|
| 235 |
+
hidden_directories_in_pattern = [
|
| 236 |
+
part for part in PurePath(pattern).parts if part.startswith(".") and not set(part) == {"."}
|
| 237 |
+
]
|
| 238 |
+
return len(hidden_directories_in_path) != len(hidden_directories_in_pattern)
|
| 239 |
+
|
| 240 |
+
|
| 241 |
+
def _get_data_files_patterns(pattern_resolver: Callable[[str], List[str]]) -> Dict[str, List[str]]:
|
| 242 |
+
"""
|
| 243 |
+
Get the default pattern from a directory or repository by testing all the supported patterns.
|
| 244 |
+
The first patterns to return a non-empty list of data files is returned.
|
| 245 |
+
|
| 246 |
+
In order, it first tests if SPLIT_PATTERN_SHARDED works, otherwise it tests the patterns in ALL_DEFAULT_PATTERNS.
|
| 247 |
+
"""
|
| 248 |
+
# first check the split patterns like data/{split}-00000-of-00001.parquet
|
| 249 |
+
for split_pattern in ALL_SPLIT_PATTERNS:
|
| 250 |
+
pattern = split_pattern.replace("{split}", "*")
|
| 251 |
+
try:
|
| 252 |
+
data_files = pattern_resolver(pattern)
|
| 253 |
+
except FileNotFoundError:
|
| 254 |
+
continue
|
| 255 |
+
if len(data_files) > 0:
|
| 256 |
+
splits: Set[str] = {
|
| 257 |
+
string_to_dict(xbasename(p), glob_pattern_to_regex(xbasename(split_pattern)))["split"]
|
| 258 |
+
for p in data_files
|
| 259 |
+
}
|
| 260 |
+
if any(not re.match(_split_re, split) for split in splits):
|
| 261 |
+
raise ValueError(f"Split name should match '{_split_re}'' but got '{splits}'.")
|
| 262 |
+
sorted_splits = [str(split) for split in DEFAULT_SPLITS if split in splits] + sorted(
|
| 263 |
+
splits - set(DEFAULT_SPLITS)
|
| 264 |
+
)
|
| 265 |
+
return {split: [split_pattern.format(split=split)] for split in sorted_splits}
|
| 266 |
+
# then check the default patterns based on train/valid/test splits
|
| 267 |
+
for patterns_dict in ALL_DEFAULT_PATTERNS:
|
| 268 |
+
non_empty_splits = []
|
| 269 |
+
for split, patterns in patterns_dict.items():
|
| 270 |
+
for pattern in patterns:
|
| 271 |
+
try:
|
| 272 |
+
data_files = pattern_resolver(pattern)
|
| 273 |
+
except FileNotFoundError:
|
| 274 |
+
continue
|
| 275 |
+
if len(data_files) > 0:
|
| 276 |
+
non_empty_splits.append(split)
|
| 277 |
+
break
|
| 278 |
+
if non_empty_splits:
|
| 279 |
+
return {split: patterns_dict[split] for split in non_empty_splits}
|
| 280 |
+
raise FileNotFoundError(f"Couldn't resolve pattern {pattern} with resolver {pattern_resolver}")
|
| 281 |
+
|
| 282 |
+
|
| 283 |
+
def _get_metadata_files_patterns(pattern_resolver: Callable[[str], List[str]]) -> List[str]:
|
| 284 |
+
"""
|
| 285 |
+
Get the supported metadata patterns from a directory or repository.
|
| 286 |
+
"""
|
| 287 |
+
non_empty_patterns = []
|
| 288 |
+
for pattern in METADATA_PATTERNS:
|
| 289 |
+
try:
|
| 290 |
+
metadata_files = pattern_resolver(pattern)
|
| 291 |
+
if len(metadata_files) > 0:
|
| 292 |
+
non_empty_patterns.append(pattern)
|
| 293 |
+
except FileNotFoundError:
|
| 294 |
+
pass
|
| 295 |
+
if non_empty_patterns:
|
| 296 |
+
return non_empty_patterns
|
| 297 |
+
raise FileNotFoundError(f"Couldn't resolve pattern {pattern} with resolver {pattern_resolver}")
|
| 298 |
+
|
| 299 |
+
|
| 300 |
+
def resolve_pattern(
|
| 301 |
+
pattern: str,
|
| 302 |
+
base_path: str,
|
| 303 |
+
allowed_extensions: Optional[List[str]] = None,
|
| 304 |
+
download_config: Optional[DownloadConfig] = None,
|
| 305 |
+
) -> List[str]:
|
| 306 |
+
"""
|
| 307 |
+
Resolve the paths and URLs of the data files from the pattern passed by the user.
|
| 308 |
+
|
| 309 |
+
You can use patterns to resolve multiple local files. Here are a few examples:
|
| 310 |
+
- *.csv to match all the CSV files at the first level
|
| 311 |
+
- **.csv to match all the CSV files at any level
|
| 312 |
+
- data/* to match all the files inside "data"
|
| 313 |
+
- data/** to match all the files inside "data" and its subdirectories
|
| 314 |
+
|
| 315 |
+
The patterns are resolved using the fsspec glob. In fsspec>=2023.12.0 this is equivalent to
|
| 316 |
+
Python's glob.glob, Path.glob, Path.match and fnmatch where ** is unsupported with a prefix/suffix
|
| 317 |
+
other than a forward slash /.
|
| 318 |
+
|
| 319 |
+
More generally:
|
| 320 |
+
- '*' matches any character except a forward-slash (to match just the file or directory name)
|
| 321 |
+
- '**' matches any character including a forward-slash /
|
| 322 |
+
|
| 323 |
+
Hidden files and directories (i.e. whose names start with a dot) are ignored, unless they are explicitly requested.
|
| 324 |
+
The same applies to special directories that start with a double underscore like "__pycache__".
|
| 325 |
+
You can still include one if the pattern explicilty mentions it:
|
| 326 |
+
- to include a hidden file: "*/.hidden.txt" or "*/.*"
|
| 327 |
+
- to include a hidden directory: ".hidden/*" or ".*/*"
|
| 328 |
+
- to include a special directory: "__special__/*" or "__*/*"
|
| 329 |
+
|
| 330 |
+
Example::
|
| 331 |
+
|
| 332 |
+
>>> from datasets.data_files import resolve_pattern
|
| 333 |
+
>>> base_path = "."
|
| 334 |
+
>>> resolve_pattern("docs/**/*.py", base_path)
|
| 335 |
+
[/Users/mariosasko/Desktop/projects/datasets/docs/source/_config.py']
|
| 336 |
+
|
| 337 |
+
Args:
|
| 338 |
+
pattern (str): Unix pattern or paths or URLs of the data files to resolve.
|
| 339 |
+
The paths can be absolute or relative to base_path.
|
| 340 |
+
Remote filesystems using fsspec are supported, e.g. with the hf:// protocol.
|
| 341 |
+
base_path (str): Base path to use when resolving relative paths.
|
| 342 |
+
allowed_extensions (Optional[list], optional): White-list of file extensions to use. Defaults to None (all extensions).
|
| 343 |
+
For example: allowed_extensions=[".csv", ".json", ".txt", ".parquet"]
|
| 344 |
+
Returns:
|
| 345 |
+
List[str]: List of paths or URLs to the local or remote files that match the patterns.
|
| 346 |
+
"""
|
| 347 |
+
if is_relative_path(pattern):
|
| 348 |
+
pattern = xjoin(base_path, pattern)
|
| 349 |
+
elif is_local_path(pattern):
|
| 350 |
+
base_path = os.path.splitdrive(pattern)[0] + os.sep
|
| 351 |
+
else:
|
| 352 |
+
base_path = ""
|
| 353 |
+
pattern, storage_options = _prepare_path_and_storage_options(pattern, download_config=download_config)
|
| 354 |
+
fs, _, _ = get_fs_token_paths(pattern, storage_options=storage_options)
|
| 355 |
+
fs_base_path = base_path.split("::")[0].split("://")[-1] or fs.root_marker
|
| 356 |
+
fs_pattern = pattern.split("::")[0].split("://")[-1]
|
| 357 |
+
files_to_ignore = set(FILES_TO_IGNORE) - {xbasename(pattern)}
|
| 358 |
+
protocol = fs.protocol if isinstance(fs.protocol, str) else fs.protocol[0]
|
| 359 |
+
protocol_prefix = protocol + "://" if protocol != "file" else ""
|
| 360 |
+
glob_kwargs = {}
|
| 361 |
+
if protocol == "hf" and config.HF_HUB_VERSION >= version.parse("0.20.0"):
|
| 362 |
+
# 10 times faster glob with detail=True (ignores costly info like lastCommit)
|
| 363 |
+
glob_kwargs["expand_info"] = False
|
| 364 |
+
matched_paths = [
|
| 365 |
+
filepath if filepath.startswith(protocol_prefix) else protocol_prefix + filepath
|
| 366 |
+
for filepath, info in fs.glob(pattern, detail=True, **glob_kwargs).items()
|
| 367 |
+
if info["type"] == "file"
|
| 368 |
+
and (xbasename(filepath) not in files_to_ignore)
|
| 369 |
+
and not _is_inside_unrequested_special_dir(
|
| 370 |
+
os.path.relpath(filepath, fs_base_path), os.path.relpath(fs_pattern, fs_base_path)
|
| 371 |
+
)
|
| 372 |
+
and not _is_unrequested_hidden_file_or_is_inside_unrequested_hidden_dir(
|
| 373 |
+
os.path.relpath(filepath, fs_base_path), os.path.relpath(fs_pattern, fs_base_path)
|
| 374 |
+
)
|
| 375 |
+
] # ignore .ipynb and __pycache__, but keep /../
|
| 376 |
+
if allowed_extensions is not None:
|
| 377 |
+
out = [
|
| 378 |
+
filepath
|
| 379 |
+
for filepath in matched_paths
|
| 380 |
+
if any("." + suffix in allowed_extensions for suffix in xbasename(filepath).split(".")[1:])
|
| 381 |
+
]
|
| 382 |
+
if len(out) < len(matched_paths):
|
| 383 |
+
invalid_matched_files = list(set(matched_paths) - set(out))
|
| 384 |
+
logger.info(
|
| 385 |
+
f"Some files matched the pattern '{pattern}' but don't have valid data file extensions: {invalid_matched_files}"
|
| 386 |
+
)
|
| 387 |
+
else:
|
| 388 |
+
out = matched_paths
|
| 389 |
+
if not out:
|
| 390 |
+
error_msg = f"Unable to find '{pattern}'"
|
| 391 |
+
if allowed_extensions is not None:
|
| 392 |
+
error_msg += f" with any supported extension {list(allowed_extensions)}"
|
| 393 |
+
raise FileNotFoundError(error_msg)
|
| 394 |
+
return out
|
| 395 |
+
|
| 396 |
+
|
| 397 |
+
def get_data_patterns(base_path: str, download_config: Optional[DownloadConfig] = None) -> Dict[str, List[str]]:
|
| 398 |
+
"""
|
| 399 |
+
Get the default pattern from a directory testing all the supported patterns.
|
| 400 |
+
The first patterns to return a non-empty list of data files is returned.
|
| 401 |
+
|
| 402 |
+
Some examples of supported patterns:
|
| 403 |
+
|
| 404 |
+
Input:
|
| 405 |
+
|
| 406 |
+
my_dataset_repository/
|
| 407 |
+
βββ README.md
|
| 408 |
+
βββ dataset.csv
|
| 409 |
+
|
| 410 |
+
Output:
|
| 411 |
+
|
| 412 |
+
{"train": ["**"]}
|
| 413 |
+
|
| 414 |
+
Input:
|
| 415 |
+
|
| 416 |
+
my_dataset_repository/
|
| 417 |
+
βββ README.md
|
| 418 |
+
βββ train.csv
|
| 419 |
+
βββ test.csv
|
| 420 |
+
|
| 421 |
+
my_dataset_repository/
|
| 422 |
+
βββ README.md
|
| 423 |
+
βββ data/
|
| 424 |
+
βββ train.csv
|
| 425 |
+
βββ test.csv
|
| 426 |
+
|
| 427 |
+
my_dataset_repository/
|
| 428 |
+
βββ README.md
|
| 429 |
+
βββ train_0.csv
|
| 430 |
+
βββ train_1.csv
|
| 431 |
+
βββ train_2.csv
|
| 432 |
+
βββ train_3.csv
|
| 433 |
+
βββ test_0.csv
|
| 434 |
+
βββ test_1.csv
|
| 435 |
+
|
| 436 |
+
Output:
|
| 437 |
+
|
| 438 |
+
{'train': ['train[-._ 0-9/]**', '**/*[-._ 0-9/]train[-._ 0-9/]**', 'training[-._ 0-9/]**', '**/*[-._ 0-9/]training[-._ 0-9/]**'],
|
| 439 |
+
'test': ['test[-._ 0-9/]**', '**/*[-._ 0-9/]test[-._ 0-9/]**', 'testing[-._ 0-9/]**', '**/*[-._ 0-9/]testing[-._ 0-9/]**', ...]}
|
| 440 |
+
|
| 441 |
+
Input:
|
| 442 |
+
|
| 443 |
+
my_dataset_repository/
|
| 444 |
+
βββ README.md
|
| 445 |
+
βββ data/
|
| 446 |
+
βββ train/
|
| 447 |
+
β βββ shard_0.csv
|
| 448 |
+
β βββ shard_1.csv
|
| 449 |
+
β βββ shard_2.csv
|
| 450 |
+
β βββ shard_3.csv
|
| 451 |
+
βββ test/
|
| 452 |
+
βββ shard_0.csv
|
| 453 |
+
βββ shard_1.csv
|
| 454 |
+
|
| 455 |
+
Output:
|
| 456 |
+
|
| 457 |
+
{'train': ['train[-._ 0-9/]**', '**/*[-._ 0-9/]train[-._ 0-9/]**', 'training[-._ 0-9/]**', '**/*[-._ 0-9/]training[-._ 0-9/]**'],
|
| 458 |
+
'test': ['test[-._ 0-9/]**', '**/*[-._ 0-9/]test[-._ 0-9/]**', 'testing[-._ 0-9/]**', '**/*[-._ 0-9/]testing[-._ 0-9/]**', ...]}
|
| 459 |
+
|
| 460 |
+
Input:
|
| 461 |
+
|
| 462 |
+
my_dataset_repository/
|
| 463 |
+
βββ README.md
|
| 464 |
+
βββ data/
|
| 465 |
+
βββ train-00000-of-00003.csv
|
| 466 |
+
βββ train-00001-of-00003.csv
|
| 467 |
+
βββ train-00002-of-00003.csv
|
| 468 |
+
βββ test-00000-of-00001.csv
|
| 469 |
+
βββ random-00000-of-00003.csv
|
| 470 |
+
βββ random-00001-of-00003.csv
|
| 471 |
+
βββ random-00002-of-00003.csv
|
| 472 |
+
|
| 473 |
+
Output:
|
| 474 |
+
|
| 475 |
+
{'train': ['data/train-[0-9][0-9][0-9][0-9][0-9]-of-[0-9][0-9][0-9][0-9][0-9]*.*'],
|
| 476 |
+
'test': ['data/test-[0-9][0-9][0-9][0-9][0-9]-of-[0-9][0-9][0-9][0-9][0-9]*.*'],
|
| 477 |
+
'random': ['data/random-[0-9][0-9][0-9][0-9][0-9]-of-[0-9][0-9][0-9][0-9][0-9]*.*']}
|
| 478 |
+
|
| 479 |
+
In order, it first tests if SPLIT_PATTERN_SHARDED works, otherwise it tests the patterns in ALL_DEFAULT_PATTERNS.
|
| 480 |
+
"""
|
| 481 |
+
resolver = partial(resolve_pattern, base_path=base_path, download_config=download_config)
|
| 482 |
+
try:
|
| 483 |
+
return _get_data_files_patterns(resolver)
|
| 484 |
+
except FileNotFoundError:
|
| 485 |
+
raise EmptyDatasetError(f"The directory at {base_path} doesn't contain any data files") from None
|
| 486 |
+
|
| 487 |
+
|
| 488 |
+
def get_metadata_patterns(
|
| 489 |
+
base_path: str,
|
| 490 |
+
download_config: Optional[DownloadConfig] = None,
|
| 491 |
+
) -> List[str]:
|
| 492 |
+
"""
|
| 493 |
+
Get the supported metadata patterns from a local directory.
|
| 494 |
+
"""
|
| 495 |
+
resolver = partial(resolve_pattern, base_path=base_path, download_config=download_config)
|
| 496 |
+
try:
|
| 497 |
+
return _get_metadata_files_patterns(resolver)
|
| 498 |
+
except FileNotFoundError:
|
| 499 |
+
raise FileNotFoundError(f"The directory at {base_path} doesn't contain any metadata file") from None
|
| 500 |
+
|
| 501 |
+
|
| 502 |
+
def _get_single_origin_metadata(
|
| 503 |
+
data_file: str,
|
| 504 |
+
download_config: Optional[DownloadConfig] = None,
|
| 505 |
+
) -> Tuple[str]:
|
| 506 |
+
data_file, storage_options = _prepare_path_and_storage_options(data_file, download_config=download_config)
|
| 507 |
+
fs, _, _ = get_fs_token_paths(data_file, storage_options=storage_options)
|
| 508 |
+
if isinstance(fs, HfFileSystem):
|
| 509 |
+
resolved_path = fs.resolve_path(data_file)
|
| 510 |
+
return (resolved_path.repo_id, resolved_path.revision)
|
| 511 |
+
elif isinstance(fs, HTTPFileSystem) and data_file.startswith(config.HF_ENDPOINT):
|
| 512 |
+
hffs = HfFileSystem(endpoint=config.HF_ENDPOINT, token=download_config.token)
|
| 513 |
+
data_file = "hf://" + data_file[len(config.HF_ENDPOINT) + 1 :].replace("/resolve/", "@", 1)
|
| 514 |
+
resolved_path = hffs.resolve_path(data_file)
|
| 515 |
+
return (resolved_path.repo_id, resolved_path.revision)
|
| 516 |
+
info = fs.info(data_file)
|
| 517 |
+
# s3fs uses "ETag", gcsfs uses "etag", and for local we simply check mtime
|
| 518 |
+
for key in ["ETag", "etag", "mtime"]:
|
| 519 |
+
if key in info:
|
| 520 |
+
return (str(info[key]),)
|
| 521 |
+
return ()
|
| 522 |
+
|
| 523 |
+
|
| 524 |
+
def _get_origin_metadata(
|
| 525 |
+
data_files: List[str],
|
| 526 |
+
max_workers=64,
|
| 527 |
+
download_config: Optional[DownloadConfig] = None,
|
| 528 |
+
) -> Tuple[str]:
|
| 529 |
+
return thread_map(
|
| 530 |
+
partial(_get_single_origin_metadata, download_config=download_config),
|
| 531 |
+
data_files,
|
| 532 |
+
max_workers=max_workers,
|
| 533 |
+
tqdm_class=hf_tqdm,
|
| 534 |
+
desc="Resolving data files",
|
| 535 |
+
# set `disable=None` rather than `disable=False` by default to disable progress bar when no TTY attached
|
| 536 |
+
disable=len(data_files) <= 16 or None,
|
| 537 |
+
)
|
| 538 |
+
|
| 539 |
+
|
| 540 |
+
class DataFilesList(List[str]):
|
| 541 |
+
"""
|
| 542 |
+
List of data files (absolute local paths or URLs).
|
| 543 |
+
It has two construction methods given the user's data files patterns :
|
| 544 |
+
- ``from_hf_repo``: resolve patterns inside a dataset repository
|
| 545 |
+
- ``from_local_or_remote``: resolve patterns from a local path
|
| 546 |
+
|
| 547 |
+
Moreover DataFilesList has an additional attribute ``origin_metadata``.
|
| 548 |
+
It can store:
|
| 549 |
+
- the last modified time of local files
|
| 550 |
+
- ETag of remote files
|
| 551 |
+
- commit sha of a dataset repository
|
| 552 |
+
|
| 553 |
+
Thanks to this additional attribute, it is possible to hash the list
|
| 554 |
+
and get a different hash if and only if at least one file changed.
|
| 555 |
+
This is useful for caching Dataset objects that are obtained from a list of data files.
|
| 556 |
+
"""
|
| 557 |
+
|
| 558 |
+
def __init__(self, data_files: List[str], origin_metadata: List[Tuple[str]]):
|
| 559 |
+
super().__init__(data_files)
|
| 560 |
+
self.origin_metadata = origin_metadata
|
| 561 |
+
|
| 562 |
+
def __add__(self, other):
|
| 563 |
+
return DataFilesList([*self, *other], self.origin_metadata + other.origin_metadata)
|
| 564 |
+
|
| 565 |
+
@classmethod
|
| 566 |
+
def from_hf_repo(
|
| 567 |
+
cls,
|
| 568 |
+
patterns: List[str],
|
| 569 |
+
dataset_info: huggingface_hub.hf_api.DatasetInfo,
|
| 570 |
+
base_path: Optional[str] = None,
|
| 571 |
+
allowed_extensions: Optional[List[str]] = None,
|
| 572 |
+
download_config: Optional[DownloadConfig] = None,
|
| 573 |
+
) -> "DataFilesList":
|
| 574 |
+
base_path = f"hf://datasets/{dataset_info.id}@{dataset_info.sha}/{base_path or ''}".rstrip("/")
|
| 575 |
+
return cls.from_patterns(
|
| 576 |
+
patterns, base_path=base_path, allowed_extensions=allowed_extensions, download_config=download_config
|
| 577 |
+
)
|
| 578 |
+
|
| 579 |
+
@classmethod
|
| 580 |
+
def from_local_or_remote(
|
| 581 |
+
cls,
|
| 582 |
+
patterns: List[str],
|
| 583 |
+
base_path: Optional[str] = None,
|
| 584 |
+
allowed_extensions: Optional[List[str]] = None,
|
| 585 |
+
download_config: Optional[DownloadConfig] = None,
|
| 586 |
+
) -> "DataFilesList":
|
| 587 |
+
base_path = base_path if base_path is not None else Path().resolve().as_posix()
|
| 588 |
+
return cls.from_patterns(
|
| 589 |
+
patterns, base_path=base_path, allowed_extensions=allowed_extensions, download_config=download_config
|
| 590 |
+
)
|
| 591 |
+
|
| 592 |
+
@classmethod
|
| 593 |
+
def from_patterns(
|
| 594 |
+
cls,
|
| 595 |
+
patterns: List[str],
|
| 596 |
+
base_path: Optional[str] = None,
|
| 597 |
+
allowed_extensions: Optional[List[str]] = None,
|
| 598 |
+
download_config: Optional[DownloadConfig] = None,
|
| 599 |
+
) -> "DataFilesList":
|
| 600 |
+
base_path = base_path if base_path is not None else Path().resolve().as_posix()
|
| 601 |
+
data_files = []
|
| 602 |
+
for pattern in patterns:
|
| 603 |
+
try:
|
| 604 |
+
data_files.extend(
|
| 605 |
+
resolve_pattern(
|
| 606 |
+
pattern,
|
| 607 |
+
base_path=base_path,
|
| 608 |
+
allowed_extensions=allowed_extensions,
|
| 609 |
+
download_config=download_config,
|
| 610 |
+
)
|
| 611 |
+
)
|
| 612 |
+
except FileNotFoundError:
|
| 613 |
+
if not has_magic(pattern):
|
| 614 |
+
raise
|
| 615 |
+
origin_metadata = _get_origin_metadata(data_files, download_config=download_config)
|
| 616 |
+
return cls(data_files, origin_metadata)
|
| 617 |
+
|
| 618 |
+
def filter_extensions(self, extensions: List[str]) -> "DataFilesList":
|
| 619 |
+
pattern = "|".join("\\" + ext for ext in extensions)
|
| 620 |
+
pattern = re.compile(f".*({pattern})(\\..+)?$")
|
| 621 |
+
return DataFilesList(
|
| 622 |
+
[data_file for data_file in self if pattern.match(data_file)],
|
| 623 |
+
origin_metadata=self.origin_metadata,
|
| 624 |
+
)
|
| 625 |
+
|
| 626 |
+
|
| 627 |
+
class DataFilesDict(Dict[str, DataFilesList]):
|
| 628 |
+
"""
|
| 629 |
+
Dict of split_name -> list of data files (absolute local paths or URLs).
|
| 630 |
+
It has two construction methods given the user's data files patterns :
|
| 631 |
+
- ``from_hf_repo``: resolve patterns inside a dataset repository
|
| 632 |
+
- ``from_local_or_remote``: resolve patterns from a local path
|
| 633 |
+
|
| 634 |
+
Moreover each list is a DataFilesList. It is possible to hash the dictionary
|
| 635 |
+
and get a different hash if and only if at least one file changed.
|
| 636 |
+
For more info, see ``DataFilesList``.
|
| 637 |
+
|
| 638 |
+
This is useful for caching Dataset objects that are obtained from a list of data files.
|
| 639 |
+
|
| 640 |
+
Changing the order of the keys of this dictionary also doesn't change its hash.
|
| 641 |
+
"""
|
| 642 |
+
|
| 643 |
+
@classmethod
|
| 644 |
+
def from_local_or_remote(
|
| 645 |
+
cls,
|
| 646 |
+
patterns: Dict[str, Union[List[str], DataFilesList]],
|
| 647 |
+
base_path: Optional[str] = None,
|
| 648 |
+
allowed_extensions: Optional[List[str]] = None,
|
| 649 |
+
download_config: Optional[DownloadConfig] = None,
|
| 650 |
+
) -> "DataFilesDict":
|
| 651 |
+
out = cls()
|
| 652 |
+
for key, patterns_for_key in patterns.items():
|
| 653 |
+
out[key] = (
|
| 654 |
+
DataFilesList.from_local_or_remote(
|
| 655 |
+
patterns_for_key,
|
| 656 |
+
base_path=base_path,
|
| 657 |
+
allowed_extensions=allowed_extensions,
|
| 658 |
+
download_config=download_config,
|
| 659 |
+
)
|
| 660 |
+
if not isinstance(patterns_for_key, DataFilesList)
|
| 661 |
+
else patterns_for_key
|
| 662 |
+
)
|
| 663 |
+
return out
|
| 664 |
+
|
| 665 |
+
@classmethod
|
| 666 |
+
def from_hf_repo(
|
| 667 |
+
cls,
|
| 668 |
+
patterns: Dict[str, Union[List[str], DataFilesList]],
|
| 669 |
+
dataset_info: huggingface_hub.hf_api.DatasetInfo,
|
| 670 |
+
base_path: Optional[str] = None,
|
| 671 |
+
allowed_extensions: Optional[List[str]] = None,
|
| 672 |
+
download_config: Optional[DownloadConfig] = None,
|
| 673 |
+
) -> "DataFilesDict":
|
| 674 |
+
out = cls()
|
| 675 |
+
for key, patterns_for_key in patterns.items():
|
| 676 |
+
out[key] = (
|
| 677 |
+
DataFilesList.from_hf_repo(
|
| 678 |
+
patterns_for_key,
|
| 679 |
+
dataset_info=dataset_info,
|
| 680 |
+
base_path=base_path,
|
| 681 |
+
allowed_extensions=allowed_extensions,
|
| 682 |
+
download_config=download_config,
|
| 683 |
+
)
|
| 684 |
+
if not isinstance(patterns_for_key, DataFilesList)
|
| 685 |
+
else patterns_for_key
|
| 686 |
+
)
|
| 687 |
+
return out
|
| 688 |
+
|
| 689 |
+
@classmethod
|
| 690 |
+
def from_patterns(
|
| 691 |
+
cls,
|
| 692 |
+
patterns: Dict[str, Union[List[str], DataFilesList]],
|
| 693 |
+
base_path: Optional[str] = None,
|
| 694 |
+
allowed_extensions: Optional[List[str]] = None,
|
| 695 |
+
download_config: Optional[DownloadConfig] = None,
|
| 696 |
+
) -> "DataFilesDict":
|
| 697 |
+
out = cls()
|
| 698 |
+
for key, patterns_for_key in patterns.items():
|
| 699 |
+
out[key] = (
|
| 700 |
+
DataFilesList.from_patterns(
|
| 701 |
+
patterns_for_key,
|
| 702 |
+
base_path=base_path,
|
| 703 |
+
allowed_extensions=allowed_extensions,
|
| 704 |
+
download_config=download_config,
|
| 705 |
+
)
|
| 706 |
+
if not isinstance(patterns_for_key, DataFilesList)
|
| 707 |
+
else patterns_for_key
|
| 708 |
+
)
|
| 709 |
+
return out
|
| 710 |
+
|
| 711 |
+
def filter_extensions(self, extensions: List[str]) -> "DataFilesDict":
|
| 712 |
+
out = type(self)()
|
| 713 |
+
for key, data_files_list in self.items():
|
| 714 |
+
out[key] = data_files_list.filter_extensions(extensions)
|
| 715 |
+
return out
|
| 716 |
+
|
| 717 |
+
|
| 718 |
+
class DataFilesPatternsList(List[str]):
|
| 719 |
+
"""
|
| 720 |
+
List of data files patterns (absolute local paths or URLs).
|
| 721 |
+
For each pattern there should also be a list of allowed extensions
|
| 722 |
+
to keep, or a None ot keep all the files for the pattern.
|
| 723 |
+
"""
|
| 724 |
+
|
| 725 |
+
def __init__(
|
| 726 |
+
self,
|
| 727 |
+
patterns: List[str],
|
| 728 |
+
allowed_extensions: List[Optional[List[str]]],
|
| 729 |
+
):
|
| 730 |
+
super().__init__(patterns)
|
| 731 |
+
self.allowed_extensions = allowed_extensions
|
| 732 |
+
|
| 733 |
+
def __add__(self, other):
|
| 734 |
+
return DataFilesList([*self, *other], self.allowed_extensions + other.allowed_extensions)
|
| 735 |
+
|
| 736 |
+
@classmethod
|
| 737 |
+
def from_patterns(
|
| 738 |
+
cls, patterns: List[str], allowed_extensions: Optional[List[str]] = None
|
| 739 |
+
) -> "DataFilesPatternsDict":
|
| 740 |
+
return cls(patterns, [allowed_extensions] * len(patterns))
|
| 741 |
+
|
| 742 |
+
def resolve(
|
| 743 |
+
self,
|
| 744 |
+
base_path: str,
|
| 745 |
+
download_config: Optional[DownloadConfig] = None,
|
| 746 |
+
) -> "DataFilesList":
|
| 747 |
+
base_path = base_path if base_path is not None else Path().resolve().as_posix()
|
| 748 |
+
data_files = []
|
| 749 |
+
for pattern, allowed_extensions in zip(self, self.allowed_extensions):
|
| 750 |
+
try:
|
| 751 |
+
data_files.extend(
|
| 752 |
+
resolve_pattern(
|
| 753 |
+
pattern,
|
| 754 |
+
base_path=base_path,
|
| 755 |
+
allowed_extensions=allowed_extensions,
|
| 756 |
+
download_config=download_config,
|
| 757 |
+
)
|
| 758 |
+
)
|
| 759 |
+
except FileNotFoundError:
|
| 760 |
+
if not has_magic(pattern):
|
| 761 |
+
raise
|
| 762 |
+
origin_metadata = _get_origin_metadata(data_files, download_config=download_config)
|
| 763 |
+
return DataFilesList(data_files, origin_metadata)
|
| 764 |
+
|
| 765 |
+
def filter_extensions(self, extensions: List[str]) -> "DataFilesList":
|
| 766 |
+
return DataFilesPatternsList(
|
| 767 |
+
self, [allowed_extensions + extensions for allowed_extensions in self.allowed_extensions]
|
| 768 |
+
)
|
| 769 |
+
|
| 770 |
+
|
| 771 |
+
class DataFilesPatternsDict(Dict[str, DataFilesPatternsList]):
|
| 772 |
+
"""
|
| 773 |
+
Dict of split_name -> list of data files patterns (absolute local paths or URLs).
|
| 774 |
+
"""
|
| 775 |
+
|
| 776 |
+
@classmethod
|
| 777 |
+
def from_patterns(
|
| 778 |
+
cls, patterns: Dict[str, List[str]], allowed_extensions: Optional[List[str]] = None
|
| 779 |
+
) -> "DataFilesPatternsDict":
|
| 780 |
+
out = cls()
|
| 781 |
+
for key, patterns_for_key in patterns.items():
|
| 782 |
+
out[key] = (
|
| 783 |
+
DataFilesPatternsList.from_patterns(
|
| 784 |
+
patterns_for_key,
|
| 785 |
+
allowed_extensions=allowed_extensions,
|
| 786 |
+
)
|
| 787 |
+
if not isinstance(patterns_for_key, DataFilesPatternsList)
|
| 788 |
+
else patterns_for_key
|
| 789 |
+
)
|
| 790 |
+
return out
|
| 791 |
+
|
| 792 |
+
def resolve(
|
| 793 |
+
self,
|
| 794 |
+
base_path: str,
|
| 795 |
+
download_config: Optional[DownloadConfig] = None,
|
| 796 |
+
) -> "DataFilesDict":
|
| 797 |
+
out = DataFilesDict()
|
| 798 |
+
for key, data_files_patterns_list in self.items():
|
| 799 |
+
out[key] = data_files_patterns_list.resolve(base_path, download_config)
|
| 800 |
+
return out
|
| 801 |
+
|
| 802 |
+
def filter_extensions(self, extensions: List[str]) -> "DataFilesPatternsDict":
|
| 803 |
+
out = type(self)()
|
| 804 |
+
for key, data_files_patterns_list in self.items():
|
| 805 |
+
out[key] = data_files_patterns_list.filter_extensions(extensions)
|
| 806 |
+
return out
|
evalkit_tf437/lib/python3.10/site-packages/datasets/dataset_dict.py
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
evalkit_tf437/lib/python3.10/site-packages/datasets/distributed.py
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from typing import TypeVar
|
| 2 |
+
|
| 3 |
+
from .arrow_dataset import Dataset, _split_by_node_map_style_dataset
|
| 4 |
+
from .iterable_dataset import IterableDataset, _split_by_node_iterable_dataset
|
| 5 |
+
|
| 6 |
+
|
| 7 |
+
DatasetType = TypeVar("DatasetType", Dataset, IterableDataset)
|
| 8 |
+
|
| 9 |
+
|
| 10 |
+
def split_dataset_by_node(dataset: DatasetType, rank: int, world_size: int) -> DatasetType:
|
| 11 |
+
"""
|
| 12 |
+
Split a dataset for the node at rank `rank` in a pool of nodes of size `world_size`.
|
| 13 |
+
|
| 14 |
+
For map-style datasets:
|
| 15 |
+
|
| 16 |
+
Each node is assigned a chunk of data, e.g. rank 0 is given the first chunk of the dataset.
|
| 17 |
+
To maximize data loading throughput, chunks are made of contiguous data on disk if possible.
|
| 18 |
+
|
| 19 |
+
For iterable datasets:
|
| 20 |
+
|
| 21 |
+
If the dataset has a number of shards that is a factor of `world_size` (i.e. if `dataset.n_shards % world_size == 0`),
|
| 22 |
+
then the shards are evenly assigned across the nodes, which is the most optimized.
|
| 23 |
+
Otherwise, each node keeps 1 example out of `world_size`, skipping the other examples.
|
| 24 |
+
|
| 25 |
+
Args:
|
| 26 |
+
dataset ([`Dataset`] or [`IterableDataset`]):
|
| 27 |
+
The dataset to split by node.
|
| 28 |
+
rank (`int`):
|
| 29 |
+
Rank of the current node.
|
| 30 |
+
world_size (`int`):
|
| 31 |
+
Total number of nodes.
|
| 32 |
+
|
| 33 |
+
Returns:
|
| 34 |
+
[`Dataset`] or [`IterableDataset`]: The dataset to be used on the node at rank `rank`.
|
| 35 |
+
"""
|
| 36 |
+
if isinstance(dataset, Dataset):
|
| 37 |
+
return _split_by_node_map_style_dataset(dataset, rank=rank, world_size=world_size)
|
| 38 |
+
else:
|
| 39 |
+
return _split_by_node_iterable_dataset(dataset, rank=rank, world_size=world_size)
|
evalkit_tf437/lib/python3.10/site-packages/datasets/exceptions.py
ADDED
|
@@ -0,0 +1,85 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# SPDX-License-Identifier: Apache-2.0
|
| 2 |
+
# Copyright 2023 The HuggingFace Authors.
|
| 3 |
+
from typing import Any, Dict, List, Optional, Union
|
| 4 |
+
|
| 5 |
+
from huggingface_hub import HfFileSystem
|
| 6 |
+
|
| 7 |
+
from . import config
|
| 8 |
+
from .table import CastError
|
| 9 |
+
from .utils.track import TrackedIterable, tracked_list, tracked_str
|
| 10 |
+
|
| 11 |
+
|
| 12 |
+
class DatasetsError(Exception):
|
| 13 |
+
"""Base class for exceptions in this library."""
|
| 14 |
+
|
| 15 |
+
|
| 16 |
+
class DefunctDatasetError(DatasetsError):
|
| 17 |
+
"""The dataset has been defunct."""
|
| 18 |
+
|
| 19 |
+
|
| 20 |
+
class FileNotFoundDatasetsError(DatasetsError, FileNotFoundError):
|
| 21 |
+
"""FileNotFoundError raised by this library."""
|
| 22 |
+
|
| 23 |
+
|
| 24 |
+
class DataFilesNotFoundError(FileNotFoundDatasetsError):
|
| 25 |
+
"""No (supported) data files found."""
|
| 26 |
+
|
| 27 |
+
|
| 28 |
+
class DatasetNotFoundError(FileNotFoundDatasetsError):
|
| 29 |
+
"""Dataset not found.
|
| 30 |
+
|
| 31 |
+
Raised when trying to access:
|
| 32 |
+
- a missing dataset, or
|
| 33 |
+
- a private/gated dataset and the user is not authenticated.
|
| 34 |
+
"""
|
| 35 |
+
|
| 36 |
+
|
| 37 |
+
class DatasetBuildError(DatasetsError):
|
| 38 |
+
pass
|
| 39 |
+
|
| 40 |
+
|
| 41 |
+
class ManualDownloadError(DatasetBuildError):
|
| 42 |
+
pass
|
| 43 |
+
|
| 44 |
+
|
| 45 |
+
class FileFormatError(DatasetBuildError):
|
| 46 |
+
pass
|
| 47 |
+
|
| 48 |
+
|
| 49 |
+
class DatasetGenerationError(DatasetBuildError):
|
| 50 |
+
pass
|
| 51 |
+
|
| 52 |
+
|
| 53 |
+
class DatasetGenerationCastError(DatasetGenerationError):
|
| 54 |
+
@classmethod
|
| 55 |
+
def from_cast_error(
|
| 56 |
+
cls,
|
| 57 |
+
cast_error: CastError,
|
| 58 |
+
builder_name: str,
|
| 59 |
+
gen_kwargs: Dict[str, Any],
|
| 60 |
+
token: Optional[Union[bool, str]],
|
| 61 |
+
) -> "DatasetGenerationCastError":
|
| 62 |
+
explanation_message = (
|
| 63 |
+
f"\n\nAll the data files must have the same columns, but at some point {cast_error.details()}"
|
| 64 |
+
)
|
| 65 |
+
formatted_tracked_gen_kwargs: List[str] = []
|
| 66 |
+
for gen_kwarg in gen_kwargs.values():
|
| 67 |
+
if not isinstance(gen_kwarg, (tracked_str, tracked_list, TrackedIterable)):
|
| 68 |
+
continue
|
| 69 |
+
while isinstance(gen_kwarg, (tracked_list, TrackedIterable)) and gen_kwarg.last_item is not None:
|
| 70 |
+
gen_kwarg = gen_kwarg.last_item
|
| 71 |
+
if isinstance(gen_kwarg, tracked_str):
|
| 72 |
+
gen_kwarg = gen_kwarg.get_origin()
|
| 73 |
+
if isinstance(gen_kwarg, str) and gen_kwarg.startswith("hf://"):
|
| 74 |
+
resolved_path = HfFileSystem(endpoint=config.HF_ENDPOINT, token=token).resolve_path(gen_kwarg)
|
| 75 |
+
gen_kwarg = "hf://" + resolved_path.unresolve()
|
| 76 |
+
if "@" + resolved_path.revision in gen_kwarg:
|
| 77 |
+
gen_kwarg = (
|
| 78 |
+
gen_kwarg.replace("@" + resolved_path.revision, "", 1)
|
| 79 |
+
+ f" (at revision {resolved_path.revision})"
|
| 80 |
+
)
|
| 81 |
+
formatted_tracked_gen_kwargs.append(str(gen_kwarg))
|
| 82 |
+
if formatted_tracked_gen_kwargs:
|
| 83 |
+
explanation_message += f"\n\nThis happened while the {builder_name} dataset builder was generating data using\n\n{', '.join(formatted_tracked_gen_kwargs)}"
|
| 84 |
+
help_message = "\n\nPlease either edit the data files to have matching columns, or separate them into different configurations (see docs at https://hf.co/docs/hub/datasets-manual-configuration#multiple-configurations)"
|
| 85 |
+
return cls("An error occurred while generating the dataset" + explanation_message + help_message)
|
evalkit_tf437/lib/python3.10/site-packages/datasets/filesystems/__init__.py
ADDED
|
@@ -0,0 +1,86 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import importlib
|
| 2 |
+
import shutil
|
| 3 |
+
import threading
|
| 4 |
+
import warnings
|
| 5 |
+
from typing import List
|
| 6 |
+
|
| 7 |
+
import fsspec
|
| 8 |
+
import fsspec.asyn
|
| 9 |
+
from fsspec.implementations.local import LocalFileSystem
|
| 10 |
+
|
| 11 |
+
from ..utils.deprecation_utils import deprecated
|
| 12 |
+
from . import compression
|
| 13 |
+
|
| 14 |
+
|
| 15 |
+
_has_s3fs = importlib.util.find_spec("s3fs") is not None
|
| 16 |
+
|
| 17 |
+
if _has_s3fs:
|
| 18 |
+
from .s3filesystem import S3FileSystem # noqa: F401
|
| 19 |
+
|
| 20 |
+
COMPRESSION_FILESYSTEMS: List[compression.BaseCompressedFileFileSystem] = [
|
| 21 |
+
compression.Bz2FileSystem,
|
| 22 |
+
compression.GzipFileSystem,
|
| 23 |
+
compression.Lz4FileSystem,
|
| 24 |
+
compression.XzFileSystem,
|
| 25 |
+
compression.ZstdFileSystem,
|
| 26 |
+
]
|
| 27 |
+
|
| 28 |
+
# Register custom filesystems
|
| 29 |
+
for fs_class in COMPRESSION_FILESYSTEMS:
|
| 30 |
+
if fs_class.protocol in fsspec.registry and fsspec.registry[fs_class.protocol] is not fs_class:
|
| 31 |
+
warnings.warn(f"A filesystem protocol was already set for {fs_class.protocol} and will be overwritten.")
|
| 32 |
+
fsspec.register_implementation(fs_class.protocol, fs_class, clobber=True)
|
| 33 |
+
|
| 34 |
+
|
| 35 |
+
@deprecated(
|
| 36 |
+
"This function is deprecated and will be removed in a future version. Please use `fsspec.core.strip_protocol` instead."
|
| 37 |
+
)
|
| 38 |
+
def extract_path_from_uri(dataset_path: str) -> str:
|
| 39 |
+
"""
|
| 40 |
+
Preprocesses `dataset_path` and removes remote filesystem (e.g. removing `s3://`).
|
| 41 |
+
|
| 42 |
+
Args:
|
| 43 |
+
dataset_path (`str`):
|
| 44 |
+
Path (e.g. `dataset/train`) or remote uri (e.g. `s3://my-bucket/dataset/train`) of the dataset directory.
|
| 45 |
+
"""
|
| 46 |
+
if "://" in dataset_path:
|
| 47 |
+
dataset_path = dataset_path.split("://")[1]
|
| 48 |
+
return dataset_path
|
| 49 |
+
|
| 50 |
+
|
| 51 |
+
def is_remote_filesystem(fs: fsspec.AbstractFileSystem) -> bool:
|
| 52 |
+
"""
|
| 53 |
+
Checks if `fs` is a remote filesystem.
|
| 54 |
+
|
| 55 |
+
Args:
|
| 56 |
+
fs (`fsspec.spec.AbstractFileSystem`):
|
| 57 |
+
An abstract super-class for pythonic file-systems, e.g. `fsspec.filesystem(\'file\')` or [`datasets.filesystems.S3FileSystem`].
|
| 58 |
+
"""
|
| 59 |
+
return not isinstance(fs, LocalFileSystem)
|
| 60 |
+
|
| 61 |
+
|
| 62 |
+
def rename(fs: fsspec.AbstractFileSystem, src: str, dst: str):
|
| 63 |
+
"""
|
| 64 |
+
Renames the file `src` in `fs` to `dst`.
|
| 65 |
+
"""
|
| 66 |
+
if not is_remote_filesystem(fs):
|
| 67 |
+
# LocalFileSystem.mv does copy + rm, it is more efficient to simply move a local directory
|
| 68 |
+
shutil.move(fs._strip_protocol(src), fs._strip_protocol(dst))
|
| 69 |
+
else:
|
| 70 |
+
fs.mv(src, dst, recursive=True)
|
| 71 |
+
|
| 72 |
+
|
| 73 |
+
def _reset_fsspec_lock() -> None:
|
| 74 |
+
"""
|
| 75 |
+
Clear reference to the loop and thread.
|
| 76 |
+
This is necessary otherwise HTTPFileSystem hangs in the ML training loop.
|
| 77 |
+
Only required for fsspec >= 0.9.0
|
| 78 |
+
See https://github.com/fsspec/gcsfs/issues/379
|
| 79 |
+
"""
|
| 80 |
+
if hasattr(fsspec.asyn, "reset_lock"):
|
| 81 |
+
# for future fsspec>2022.05.0
|
| 82 |
+
fsspec.asyn.reset_lock()
|
| 83 |
+
else:
|
| 84 |
+
fsspec.asyn.iothread[0] = None
|
| 85 |
+
fsspec.asyn.loop[0] = None
|
| 86 |
+
fsspec.asyn.lock = threading.Lock()
|
evalkit_tf437/lib/python3.10/site-packages/datasets/filesystems/__pycache__/__init__.cpython-310.pyc
ADDED
|
Binary file (2.83 kB). View file
|
|
|
evalkit_tf437/lib/python3.10/site-packages/datasets/filesystems/__pycache__/s3filesystem.cpython-310.pyc
ADDED
|
Binary file (6.05 kB). View file
|
|
|
evalkit_tf437/lib/python3.10/site-packages/datasets/filesystems/compression.py
ADDED
|
@@ -0,0 +1,178 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import os
|
| 2 |
+
from typing import Optional
|
| 3 |
+
|
| 4 |
+
import fsspec
|
| 5 |
+
from fsspec.archive import AbstractArchiveFileSystem
|
| 6 |
+
from fsspec.utils import DEFAULT_BLOCK_SIZE
|
| 7 |
+
|
| 8 |
+
|
| 9 |
+
class BaseCompressedFileFileSystem(AbstractArchiveFileSystem):
|
| 10 |
+
"""Read contents of compressed file as a filesystem with one file inside."""
|
| 11 |
+
|
| 12 |
+
root_marker = ""
|
| 13 |
+
protocol: str = (
|
| 14 |
+
None # protocol passed in prefix to the url. ex: "gzip", for gzip://file.txt::http://foo.bar/file.txt.gz
|
| 15 |
+
)
|
| 16 |
+
compression: str = None # compression type in fsspec. ex: "gzip"
|
| 17 |
+
extension: str = None # extension of the filename to strip. ex: "".gz" to get file.txt from file.txt.gz
|
| 18 |
+
|
| 19 |
+
def __init__(
|
| 20 |
+
self, fo: str = "", target_protocol: Optional[str] = None, target_options: Optional[dict] = None, **kwargs
|
| 21 |
+
):
|
| 22 |
+
"""
|
| 23 |
+
The compressed file system can be instantiated from any compressed file.
|
| 24 |
+
It reads the contents of compressed file as a filesystem with one file inside, as if it was an archive.
|
| 25 |
+
|
| 26 |
+
The single file inside the filesystem is named after the compresssed file,
|
| 27 |
+
without the compression extension at the end of the filename.
|
| 28 |
+
|
| 29 |
+
Args:
|
| 30 |
+
fo (:obj:``str``): Path to compressed file. Will fetch file using ``fsspec.open()``
|
| 31 |
+
mode (:obj:``str``): Currently, only 'rb' accepted
|
| 32 |
+
target_protocol(:obj:``str``, optional): To override the FS protocol inferred from a URL.
|
| 33 |
+
target_options (:obj:``dict``, optional): Kwargs passed when instantiating the target FS.
|
| 34 |
+
"""
|
| 35 |
+
super().__init__(self, **kwargs)
|
| 36 |
+
# always open as "rb" since fsspec can then use the TextIOWrapper to make it work for "r" mode
|
| 37 |
+
self.file = fsspec.open(
|
| 38 |
+
fo,
|
| 39 |
+
mode="rb",
|
| 40 |
+
protocol=target_protocol,
|
| 41 |
+
compression=self.compression,
|
| 42 |
+
client_kwargs={
|
| 43 |
+
"requote_redirect_url": False, # see https://github.com/huggingface/datasets/pull/5459
|
| 44 |
+
"trust_env": True, # Enable reading proxy env variables.
|
| 45 |
+
**(target_options or {}).pop("client_kwargs", {}), # To avoid issues if it was already passed.
|
| 46 |
+
},
|
| 47 |
+
**(target_options or {}),
|
| 48 |
+
)
|
| 49 |
+
self.compressed_name = os.path.basename(self.file.path.split("::")[0])
|
| 50 |
+
self.uncompressed_name = (
|
| 51 |
+
self.compressed_name[: self.compressed_name.rindex(".")]
|
| 52 |
+
if "." in self.compressed_name
|
| 53 |
+
else self.compressed_name
|
| 54 |
+
)
|
| 55 |
+
self.dir_cache = None
|
| 56 |
+
|
| 57 |
+
@classmethod
|
| 58 |
+
def _strip_protocol(cls, path):
|
| 59 |
+
# compressed file paths are always relative to the archive root
|
| 60 |
+
return super()._strip_protocol(path).lstrip("/")
|
| 61 |
+
|
| 62 |
+
def _get_dirs(self):
|
| 63 |
+
if self.dir_cache is None:
|
| 64 |
+
f = {**self.file.fs.info(self.file.path), "name": self.uncompressed_name}
|
| 65 |
+
self.dir_cache = {f["name"]: f}
|
| 66 |
+
|
| 67 |
+
def cat(self, path: str):
|
| 68 |
+
return self.file.open().read()
|
| 69 |
+
|
| 70 |
+
def _open(
|
| 71 |
+
self,
|
| 72 |
+
path: str,
|
| 73 |
+
mode: str = "rb",
|
| 74 |
+
block_size=None,
|
| 75 |
+
autocommit=True,
|
| 76 |
+
cache_options=None,
|
| 77 |
+
**kwargs,
|
| 78 |
+
):
|
| 79 |
+
path = self._strip_protocol(path)
|
| 80 |
+
if mode != "rb":
|
| 81 |
+
raise ValueError(f"Tried to read with mode {mode} on file {self.file.path} opened with mode 'rb'")
|
| 82 |
+
return self.file.open()
|
| 83 |
+
|
| 84 |
+
|
| 85 |
+
class Bz2FileSystem(BaseCompressedFileFileSystem):
|
| 86 |
+
"""Read contents of BZ2 file as a filesystem with one file inside."""
|
| 87 |
+
|
| 88 |
+
protocol = "bz2"
|
| 89 |
+
compression = "bz2"
|
| 90 |
+
extension = ".bz2"
|
| 91 |
+
|
| 92 |
+
|
| 93 |
+
class GzipFileSystem(BaseCompressedFileFileSystem):
|
| 94 |
+
"""Read contents of GZIP file as a filesystem with one file inside."""
|
| 95 |
+
|
| 96 |
+
protocol = "gzip"
|
| 97 |
+
compression = "gzip"
|
| 98 |
+
extension = ".gz"
|
| 99 |
+
|
| 100 |
+
|
| 101 |
+
class Lz4FileSystem(BaseCompressedFileFileSystem):
|
| 102 |
+
"""Read contents of LZ4 file as a filesystem with one file inside."""
|
| 103 |
+
|
| 104 |
+
protocol = "lz4"
|
| 105 |
+
compression = "lz4"
|
| 106 |
+
extension = ".lz4"
|
| 107 |
+
|
| 108 |
+
|
| 109 |
+
class XzFileSystem(BaseCompressedFileFileSystem):
|
| 110 |
+
"""Read contents of .xz (LZMA) file as a filesystem with one file inside."""
|
| 111 |
+
|
| 112 |
+
protocol = "xz"
|
| 113 |
+
compression = "xz"
|
| 114 |
+
extension = ".xz"
|
| 115 |
+
|
| 116 |
+
|
| 117 |
+
class ZstdFileSystem(BaseCompressedFileFileSystem):
|
| 118 |
+
"""
|
| 119 |
+
Read contents of zstd file as a filesystem with one file inside.
|
| 120 |
+
|
| 121 |
+
Note that reading in binary mode with fsspec isn't supported yet:
|
| 122 |
+
https://github.com/indygreg/python-zstandard/issues/136
|
| 123 |
+
"""
|
| 124 |
+
|
| 125 |
+
protocol = "zstd"
|
| 126 |
+
compression = "zstd"
|
| 127 |
+
extension = ".zst"
|
| 128 |
+
|
| 129 |
+
def __init__(
|
| 130 |
+
self,
|
| 131 |
+
fo: str,
|
| 132 |
+
mode: str = "rb",
|
| 133 |
+
target_protocol: Optional[str] = None,
|
| 134 |
+
target_options: Optional[dict] = None,
|
| 135 |
+
block_size: int = DEFAULT_BLOCK_SIZE,
|
| 136 |
+
**kwargs,
|
| 137 |
+
):
|
| 138 |
+
super().__init__(
|
| 139 |
+
fo=fo,
|
| 140 |
+
mode=mode,
|
| 141 |
+
target_protocol=target_protocol,
|
| 142 |
+
target_options=target_options,
|
| 143 |
+
block_size=block_size,
|
| 144 |
+
**kwargs,
|
| 145 |
+
)
|
| 146 |
+
# We need to wrap the zstd decompressor to avoid this error in fsspec==2021.7.0 and zstandard==0.15.2:
|
| 147 |
+
#
|
| 148 |
+
# File "/Users/user/.virtualenvs/hf-datasets/lib/python3.7/site-packages/fsspec/core.py", line 145, in open
|
| 149 |
+
# out.close = close
|
| 150 |
+
# AttributeError: 'zstd.ZstdDecompressionReader' object attribute 'close' is read-only
|
| 151 |
+
#
|
| 152 |
+
# see https://github.com/intake/filesystem_spec/issues/725
|
| 153 |
+
_enter = self.file.__enter__
|
| 154 |
+
|
| 155 |
+
class WrappedFile:
|
| 156 |
+
def __init__(self, file_):
|
| 157 |
+
self._file = file_
|
| 158 |
+
|
| 159 |
+
def __enter__(self):
|
| 160 |
+
self._file.__enter__()
|
| 161 |
+
return self
|
| 162 |
+
|
| 163 |
+
def __exit__(self, *args, **kwargs):
|
| 164 |
+
self._file.__exit__(*args, **kwargs)
|
| 165 |
+
|
| 166 |
+
def __iter__(self):
|
| 167 |
+
return iter(self._file)
|
| 168 |
+
|
| 169 |
+
def __next__(self):
|
| 170 |
+
return next(self._file)
|
| 171 |
+
|
| 172 |
+
def __getattr__(self, attr):
|
| 173 |
+
return getattr(self._file, attr)
|
| 174 |
+
|
| 175 |
+
def fixed_enter(*args, **kwargs):
|
| 176 |
+
return WrappedFile(_enter(*args, **kwargs))
|
| 177 |
+
|
| 178 |
+
self.file.__enter__ = fixed_enter
|
evalkit_tf437/lib/python3.10/site-packages/datasets/filesystems/s3filesystem.py
ADDED
|
@@ -0,0 +1,116 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import s3fs
|
| 2 |
+
|
| 3 |
+
from ..utils.deprecation_utils import deprecated
|
| 4 |
+
|
| 5 |
+
|
| 6 |
+
@deprecated("Use s3fs.S3FileSystem instead.")
|
| 7 |
+
class S3FileSystem(s3fs.S3FileSystem):
|
| 8 |
+
"""
|
| 9 |
+
`datasets.filesystems.S3FileSystem` is a subclass of [`s3fs.S3FileSystem`](https://s3fs.readthedocs.io/en/latest/api.html).
|
| 10 |
+
|
| 11 |
+
Users can use this class to access S3 as if it were a file system. It exposes a filesystem-like API (ls, cp, open, etc.) on top of S3 storage. Provide credentials either explicitly (`key=`, `secret=`) or with boto's credential methods. See botocore documentation for more information. If no credentials are available, use `anon=True`.
|
| 12 |
+
|
| 13 |
+
Args:
|
| 14 |
+
anon (`bool`, default to `False`):
|
| 15 |
+
Whether to use anonymous connection (public buckets only). If `False`, uses the key/secret given,
|
| 16 |
+
or boto's credential resolver (client_kwargs, environment, variables, config files, EC2 IAM server, in that order).
|
| 17 |
+
key (`str`):
|
| 18 |
+
If not anonymous, use this access key ID, if specified.
|
| 19 |
+
secret (`str`):
|
| 20 |
+
If not anonymous, use this secret access key, if specified.
|
| 21 |
+
token (`str`):
|
| 22 |
+
If not anonymous, use this security token, if specified.
|
| 23 |
+
use_ssl (`bool`, defaults to `True`):
|
| 24 |
+
Whether to use SSL in connections to S3; may be faster without, but insecure. If `use_ssl` is
|
| 25 |
+
also set in `client_kwargs`, the value set in `client_kwargs` will take priority.
|
| 26 |
+
s3_additional_kwargs (`dict`):
|
| 27 |
+
Parameters that are used when calling S3 API methods. Typically used for things
|
| 28 |
+
like ServerSideEncryption.
|
| 29 |
+
client_kwargs (`dict`):
|
| 30 |
+
Parameters for the botocore client.
|
| 31 |
+
requester_pays (`bool`, defaults to `False`):
|
| 32 |
+
Whether `RequesterPays` buckets are supported.
|
| 33 |
+
default_block_size (`int`):
|
| 34 |
+
If given, the default block size value used for `open()`, if no specific value is given at all time.
|
| 35 |
+
The built-in default is 5MB.
|
| 36 |
+
default_fill_cache (`bool`, defaults to `True`):
|
| 37 |
+
Whether to use cache filling with open by default. Refer to `S3File.open`.
|
| 38 |
+
default_cache_type (`str`, defaults to `bytes`):
|
| 39 |
+
If given, the default `cache_type` value used for `open()`. Set to `none` if no
|
| 40 |
+
caching is desired. See fsspec's documentation for other available `cache_type` values.
|
| 41 |
+
version_aware (`bool`, defaults to `False`):
|
| 42 |
+
Whether to support bucket versioning. If enable this will require the user to have
|
| 43 |
+
the necessary IAM permissions for dealing with versioned objects.
|
| 44 |
+
cache_regions (`bool`, defaults to `False`):
|
| 45 |
+
Whether to cache bucket regions. Whenever a new bucket is used, it will
|
| 46 |
+
first find out which region it belongs to and then use the client for that region.
|
| 47 |
+
asynchronous (`bool`, defaults to `False`):
|
| 48 |
+
Whether this instance is to be used from inside coroutines.
|
| 49 |
+
config_kwargs (`dict`):
|
| 50 |
+
Parameters passed to `botocore.client.Config`.
|
| 51 |
+
**kwargs:
|
| 52 |
+
Other parameters for core session.
|
| 53 |
+
session (`aiobotocore.session.AioSession`):
|
| 54 |
+
Session to be used for all connections. This session will be used inplace of creating
|
| 55 |
+
a new session inside S3FileSystem. For example: `aiobotocore.session.AioSession(profile='test_user')`.
|
| 56 |
+
skip_instance_cache (`bool`):
|
| 57 |
+
Control reuse of instances. Passed on to `fsspec`.
|
| 58 |
+
use_listings_cache (`bool`):
|
| 59 |
+
Control reuse of directory listings. Passed on to `fsspec`.
|
| 60 |
+
listings_expiry_time (`int` or `float`):
|
| 61 |
+
Control reuse of directory listings. Passed on to `fsspec`.
|
| 62 |
+
max_paths (`int`): Control reuse of directory listings. Passed on to `fsspec`.
|
| 63 |
+
|
| 64 |
+
Examples:
|
| 65 |
+
|
| 66 |
+
Listing files from public S3 bucket.
|
| 67 |
+
|
| 68 |
+
```py
|
| 69 |
+
>>> import datasets
|
| 70 |
+
>>> s3 = datasets.filesystems.S3FileSystem(anon=True) # doctest: +SKIP
|
| 71 |
+
>>> s3.ls('public-datasets/imdb/train') # doctest: +SKIP
|
| 72 |
+
['dataset_info.json.json','dataset.arrow','state.json']
|
| 73 |
+
```
|
| 74 |
+
|
| 75 |
+
Listing files from private S3 bucket using `aws_access_key_id` and `aws_secret_access_key`.
|
| 76 |
+
|
| 77 |
+
```py
|
| 78 |
+
>>> import datasets
|
| 79 |
+
>>> s3 = datasets.filesystems.S3FileSystem(key=aws_access_key_id, secret=aws_secret_access_key) # doctest: +SKIP
|
| 80 |
+
>>> s3.ls('my-private-datasets/imdb/train') # doctest: +SKIP
|
| 81 |
+
['dataset_info.json.json','dataset.arrow','state.json']
|
| 82 |
+
```
|
| 83 |
+
|
| 84 |
+
Using `S3Filesystem` with `botocore.session.Session` and custom `aws_profile`.
|
| 85 |
+
|
| 86 |
+
```py
|
| 87 |
+
>>> import botocore
|
| 88 |
+
>>> from datasets.filesystems import S3Filesystem
|
| 89 |
+
|
| 90 |
+
>>> s3_session = botocore.session.Session(profile_name='my_profile_name')
|
| 91 |
+
>>> s3 = S3FileSystem(session=s3_session) # doctest: +SKIP
|
| 92 |
+
```
|
| 93 |
+
|
| 94 |
+
Loading dataset from S3 using `S3Filesystem` and [`load_from_disk`].
|
| 95 |
+
|
| 96 |
+
```py
|
| 97 |
+
>>> from datasets import load_from_disk
|
| 98 |
+
>>> from datasets.filesystems import S3Filesystem
|
| 99 |
+
|
| 100 |
+
>>> s3 = S3FileSystem(key=aws_access_key_id, secret=aws_secret_access_key) # doctest: +SKIP
|
| 101 |
+
>>> dataset = load_from_disk('s3://my-private-datasets/imdb/train', storage_options=s3.storage_options) # doctest: +SKIP
|
| 102 |
+
>>> print(len(dataset))
|
| 103 |
+
25000
|
| 104 |
+
```
|
| 105 |
+
|
| 106 |
+
Saving dataset to S3 using `S3Filesystem` and [`Dataset.save_to_disk`].
|
| 107 |
+
|
| 108 |
+
```py
|
| 109 |
+
>>> from datasets import load_dataset
|
| 110 |
+
>>> from datasets.filesystems import S3Filesystem
|
| 111 |
+
|
| 112 |
+
>>> dataset = load_dataset("imdb")
|
| 113 |
+
>>> s3 = S3FileSystem(key=aws_access_key_id, secret=aws_secret_access_key) # doctest: +SKIP
|
| 114 |
+
>>> dataset.save_to_disk('s3://my-private-datasets/imdb/train', storage_options=s3.storage_options) # doctest: +SKIP
|
| 115 |
+
```
|
| 116 |
+
"""
|
evalkit_tf437/lib/python3.10/site-packages/datasets/info.py
ADDED
|
@@ -0,0 +1,592 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors.
|
| 2 |
+
#
|
| 3 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
| 4 |
+
# you may not use this file except in compliance with the License.
|
| 5 |
+
# You may obtain a copy of the License at
|
| 6 |
+
#
|
| 7 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
| 8 |
+
#
|
| 9 |
+
# Unless required by applicable law or agreed to in writing, software
|
| 10 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
| 11 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| 12 |
+
# See the License for the specific language governing permissions and
|
| 13 |
+
# limitations under the License.
|
| 14 |
+
|
| 15 |
+
# Lint as: python3
|
| 16 |
+
"""DatasetInfo and MetricInfo record information we know about a dataset and a metric.
|
| 17 |
+
|
| 18 |
+
This includes things that we know about the dataset statically, i.e.:
|
| 19 |
+
- description
|
| 20 |
+
- canonical location
|
| 21 |
+
- does it have validation and tests splits
|
| 22 |
+
- size
|
| 23 |
+
- etc.
|
| 24 |
+
|
| 25 |
+
This also includes the things that can and should be computed once we've
|
| 26 |
+
processed the dataset as well:
|
| 27 |
+
- number of examples (in each split)
|
| 28 |
+
- etc.
|
| 29 |
+
"""
|
| 30 |
+
|
| 31 |
+
import copy
|
| 32 |
+
import dataclasses
|
| 33 |
+
import json
|
| 34 |
+
import os
|
| 35 |
+
import posixpath
|
| 36 |
+
import warnings
|
| 37 |
+
from dataclasses import dataclass
|
| 38 |
+
from pathlib import Path
|
| 39 |
+
from typing import ClassVar, Dict, List, Optional, Union
|
| 40 |
+
|
| 41 |
+
import fsspec
|
| 42 |
+
from huggingface_hub import DatasetCard, DatasetCardData
|
| 43 |
+
|
| 44 |
+
from . import config
|
| 45 |
+
from .features import Features, Value
|
| 46 |
+
from .splits import SplitDict
|
| 47 |
+
from .tasks import TaskTemplate, task_template_from_dict
|
| 48 |
+
from .utils import Version
|
| 49 |
+
from .utils.logging import get_logger
|
| 50 |
+
from .utils.py_utils import asdict, unique_values
|
| 51 |
+
|
| 52 |
+
|
| 53 |
+
logger = get_logger(__name__)
|
| 54 |
+
|
| 55 |
+
|
| 56 |
+
@dataclass
|
| 57 |
+
class SupervisedKeysData:
|
| 58 |
+
input: str = ""
|
| 59 |
+
output: str = ""
|
| 60 |
+
|
| 61 |
+
|
| 62 |
+
@dataclass
|
| 63 |
+
class DownloadChecksumsEntryData:
|
| 64 |
+
key: str = ""
|
| 65 |
+
value: str = ""
|
| 66 |
+
|
| 67 |
+
|
| 68 |
+
class MissingCachedSizesConfigError(Exception):
|
| 69 |
+
"""The expected cached sizes of the download file are missing."""
|
| 70 |
+
|
| 71 |
+
|
| 72 |
+
class NonMatchingCachedSizesError(Exception):
|
| 73 |
+
"""The prepared split doesn't have expected sizes."""
|
| 74 |
+
|
| 75 |
+
|
| 76 |
+
@dataclass
|
| 77 |
+
class PostProcessedInfo:
|
| 78 |
+
features: Optional[Features] = None
|
| 79 |
+
resources_checksums: Optional[dict] = None
|
| 80 |
+
|
| 81 |
+
def __post_init__(self):
|
| 82 |
+
# Convert back to the correct classes when we reload from dict
|
| 83 |
+
if self.features is not None and not isinstance(self.features, Features):
|
| 84 |
+
self.features = Features.from_dict(self.features)
|
| 85 |
+
|
| 86 |
+
@classmethod
|
| 87 |
+
def from_dict(cls, post_processed_info_dict: dict) -> "PostProcessedInfo":
|
| 88 |
+
field_names = {f.name for f in dataclasses.fields(cls)}
|
| 89 |
+
return cls(**{k: v for k, v in post_processed_info_dict.items() if k in field_names})
|
| 90 |
+
|
| 91 |
+
|
| 92 |
+
@dataclass
|
| 93 |
+
class DatasetInfo:
|
| 94 |
+
"""Information about a dataset.
|
| 95 |
+
|
| 96 |
+
`DatasetInfo` documents datasets, including its name, version, and features.
|
| 97 |
+
See the constructor arguments and properties for a full list.
|
| 98 |
+
|
| 99 |
+
Not all fields are known on construction and may be updated later.
|
| 100 |
+
|
| 101 |
+
Attributes:
|
| 102 |
+
description (`str`):
|
| 103 |
+
A description of the dataset.
|
| 104 |
+
citation (`str`):
|
| 105 |
+
A BibTeX citation of the dataset.
|
| 106 |
+
homepage (`str`):
|
| 107 |
+
A URL to the official homepage for the dataset.
|
| 108 |
+
license (`str`):
|
| 109 |
+
The dataset's license. It can be the name of the license or a paragraph containing the terms of the license.
|
| 110 |
+
features ([`Features`], *optional*):
|
| 111 |
+
The features used to specify the dataset's column types.
|
| 112 |
+
post_processed (`PostProcessedInfo`, *optional*):
|
| 113 |
+
Information regarding the resources of a possible post-processing of a dataset. For example, it can contain the information of an index.
|
| 114 |
+
supervised_keys (`SupervisedKeysData`, *optional*):
|
| 115 |
+
Specifies the input feature and the label for supervised learning if applicable for the dataset (legacy from TFDS).
|
| 116 |
+
builder_name (`str`, *optional*):
|
| 117 |
+
The name of the `GeneratorBasedBuilder` subclass used to create the dataset. Usually matched to the corresponding script name. It is also the snake_case version of the dataset builder class name.
|
| 118 |
+
config_name (`str`, *optional*):
|
| 119 |
+
The name of the configuration derived from [`BuilderConfig`].
|
| 120 |
+
version (`str` or [`Version`], *optional*):
|
| 121 |
+
The version of the dataset.
|
| 122 |
+
splits (`dict`, *optional*):
|
| 123 |
+
The mapping between split name and metadata.
|
| 124 |
+
download_checksums (`dict`, *optional*):
|
| 125 |
+
The mapping between the URL to download the dataset's checksums and corresponding metadata.
|
| 126 |
+
download_size (`int`, *optional*):
|
| 127 |
+
The size of the files to download to generate the dataset, in bytes.
|
| 128 |
+
post_processing_size (`int`, *optional*):
|
| 129 |
+
Size of the dataset in bytes after post-processing, if any.
|
| 130 |
+
dataset_size (`int`, *optional*):
|
| 131 |
+
The combined size in bytes of the Arrow tables for all splits.
|
| 132 |
+
size_in_bytes (`int`, *optional*):
|
| 133 |
+
The combined size in bytes of all files associated with the dataset (downloaded files + Arrow files).
|
| 134 |
+
task_templates (`List[TaskTemplate]`, *optional*):
|
| 135 |
+
The task templates to prepare the dataset for during training and evaluation. Each template casts the dataset's [`Features`] to standardized column names and types as detailed in `datasets.tasks`.
|
| 136 |
+
**config_kwargs (additional keyword arguments):
|
| 137 |
+
Keyword arguments to be passed to the [`BuilderConfig`] and used in the [`DatasetBuilder`].
|
| 138 |
+
"""
|
| 139 |
+
|
| 140 |
+
# Set in the dataset scripts
|
| 141 |
+
description: str = dataclasses.field(default_factory=str)
|
| 142 |
+
citation: str = dataclasses.field(default_factory=str)
|
| 143 |
+
homepage: str = dataclasses.field(default_factory=str)
|
| 144 |
+
license: str = dataclasses.field(default_factory=str)
|
| 145 |
+
features: Optional[Features] = None
|
| 146 |
+
post_processed: Optional[PostProcessedInfo] = None
|
| 147 |
+
supervised_keys: Optional[SupervisedKeysData] = None
|
| 148 |
+
task_templates: Optional[List[TaskTemplate]] = None
|
| 149 |
+
|
| 150 |
+
# Set later by the builder
|
| 151 |
+
builder_name: Optional[str] = None
|
| 152 |
+
dataset_name: Optional[str] = None # for packaged builders, to be different from builder_name
|
| 153 |
+
config_name: Optional[str] = None
|
| 154 |
+
version: Optional[Union[str, Version]] = None
|
| 155 |
+
# Set later by `download_and_prepare`
|
| 156 |
+
splits: Optional[dict] = None
|
| 157 |
+
download_checksums: Optional[dict] = None
|
| 158 |
+
download_size: Optional[int] = None
|
| 159 |
+
post_processing_size: Optional[int] = None
|
| 160 |
+
dataset_size: Optional[int] = None
|
| 161 |
+
size_in_bytes: Optional[int] = None
|
| 162 |
+
|
| 163 |
+
_INCLUDED_INFO_IN_YAML: ClassVar[List[str]] = [
|
| 164 |
+
"config_name",
|
| 165 |
+
"download_size",
|
| 166 |
+
"dataset_size",
|
| 167 |
+
"features",
|
| 168 |
+
"splits",
|
| 169 |
+
]
|
| 170 |
+
|
| 171 |
+
def __post_init__(self):
|
| 172 |
+
# Convert back to the correct classes when we reload from dict
|
| 173 |
+
if self.features is not None and not isinstance(self.features, Features):
|
| 174 |
+
self.features = Features.from_dict(self.features)
|
| 175 |
+
if self.post_processed is not None and not isinstance(self.post_processed, PostProcessedInfo):
|
| 176 |
+
self.post_processed = PostProcessedInfo.from_dict(self.post_processed)
|
| 177 |
+
if self.version is not None and not isinstance(self.version, Version):
|
| 178 |
+
if isinstance(self.version, str):
|
| 179 |
+
self.version = Version(self.version)
|
| 180 |
+
else:
|
| 181 |
+
self.version = Version.from_dict(self.version)
|
| 182 |
+
if self.splits is not None and not isinstance(self.splits, SplitDict):
|
| 183 |
+
self.splits = SplitDict.from_split_dict(self.splits)
|
| 184 |
+
if self.supervised_keys is not None and not isinstance(self.supervised_keys, SupervisedKeysData):
|
| 185 |
+
if isinstance(self.supervised_keys, (tuple, list)):
|
| 186 |
+
self.supervised_keys = SupervisedKeysData(*self.supervised_keys)
|
| 187 |
+
else:
|
| 188 |
+
self.supervised_keys = SupervisedKeysData(**self.supervised_keys)
|
| 189 |
+
|
| 190 |
+
# Parse and make a list of templates
|
| 191 |
+
if self.task_templates is not None:
|
| 192 |
+
if isinstance(self.task_templates, (list, tuple)):
|
| 193 |
+
templates = [
|
| 194 |
+
template if isinstance(template, TaskTemplate) else task_template_from_dict(template)
|
| 195 |
+
for template in self.task_templates
|
| 196 |
+
]
|
| 197 |
+
self.task_templates = [template for template in templates if template is not None]
|
| 198 |
+
elif isinstance(self.task_templates, TaskTemplate):
|
| 199 |
+
self.task_templates = [self.task_templates]
|
| 200 |
+
else:
|
| 201 |
+
template = task_template_from_dict(self.task_templates)
|
| 202 |
+
self.task_templates = [template] if template is not None else []
|
| 203 |
+
|
| 204 |
+
# Align task templates with features
|
| 205 |
+
if self.task_templates is not None:
|
| 206 |
+
self.task_templates = list(self.task_templates)
|
| 207 |
+
if self.features is not None:
|
| 208 |
+
self.task_templates = [
|
| 209 |
+
template.align_with_features(self.features) for template in (self.task_templates)
|
| 210 |
+
]
|
| 211 |
+
|
| 212 |
+
def write_to_directory(
|
| 213 |
+
self, dataset_info_dir, pretty_print=False, fs="deprecated", storage_options: Optional[dict] = None
|
| 214 |
+
):
|
| 215 |
+
"""Write `DatasetInfo` and license (if present) as JSON files to `dataset_info_dir`.
|
| 216 |
+
|
| 217 |
+
Args:
|
| 218 |
+
dataset_info_dir (`str`):
|
| 219 |
+
Destination directory.
|
| 220 |
+
pretty_print (`bool`, defaults to `False`):
|
| 221 |
+
If `True`, the JSON will be pretty-printed with the indent level of 4.
|
| 222 |
+
fs (`fsspec.spec.AbstractFileSystem`, *optional*):
|
| 223 |
+
Instance of the remote filesystem used to download the files from.
|
| 224 |
+
|
| 225 |
+
<Deprecated version="2.9.0">
|
| 226 |
+
|
| 227 |
+
`fs` was deprecated in version 2.9.0 and will be removed in 3.0.0.
|
| 228 |
+
Please use `storage_options` instead, e.g. `storage_options=fs.storage_options`.
|
| 229 |
+
|
| 230 |
+
</Deprecated>
|
| 231 |
+
|
| 232 |
+
storage_options (`dict`, *optional*):
|
| 233 |
+
Key/value pairs to be passed on to the file-system backend, if any.
|
| 234 |
+
|
| 235 |
+
<Added version="2.9.0"/>
|
| 236 |
+
|
| 237 |
+
Example:
|
| 238 |
+
|
| 239 |
+
```py
|
| 240 |
+
>>> from datasets import load_dataset
|
| 241 |
+
>>> ds = load_dataset("rotten_tomatoes", split="validation")
|
| 242 |
+
>>> ds.info.write_to_directory("/path/to/directory/")
|
| 243 |
+
```
|
| 244 |
+
"""
|
| 245 |
+
if fs != "deprecated":
|
| 246 |
+
warnings.warn(
|
| 247 |
+
"'fs' was deprecated in favor of 'storage_options' in version 2.9.0 and will be removed in 3.0.0.\n"
|
| 248 |
+
"You can remove this warning by passing 'storage_options=fs.storage_options' instead.",
|
| 249 |
+
FutureWarning,
|
| 250 |
+
)
|
| 251 |
+
storage_options = fs.storage_options
|
| 252 |
+
|
| 253 |
+
fs: fsspec.AbstractFileSystem
|
| 254 |
+
fs, _, _ = fsspec.get_fs_token_paths(dataset_info_dir, storage_options=storage_options)
|
| 255 |
+
with fs.open(posixpath.join(dataset_info_dir, config.DATASET_INFO_FILENAME), "wb") as f:
|
| 256 |
+
self._dump_info(f, pretty_print=pretty_print)
|
| 257 |
+
if self.license:
|
| 258 |
+
with fs.open(posixpath.join(dataset_info_dir, config.LICENSE_FILENAME), "wb") as f:
|
| 259 |
+
self._dump_license(f)
|
| 260 |
+
|
| 261 |
+
def _dump_info(self, file, pretty_print=False):
|
| 262 |
+
"""Dump info in `file` file-like object open in bytes mode (to support remote files)"""
|
| 263 |
+
file.write(json.dumps(asdict(self), indent=4 if pretty_print else None).encode("utf-8"))
|
| 264 |
+
|
| 265 |
+
def _dump_license(self, file):
|
| 266 |
+
"""Dump license in `file` file-like object open in bytes mode (to support remote files)"""
|
| 267 |
+
file.write(self.license.encode("utf-8"))
|
| 268 |
+
|
| 269 |
+
@classmethod
|
| 270 |
+
def from_merge(cls, dataset_infos: List["DatasetInfo"]):
|
| 271 |
+
dataset_infos = [dset_info.copy() for dset_info in dataset_infos if dset_info is not None]
|
| 272 |
+
|
| 273 |
+
if len(dataset_infos) > 0 and all(dataset_infos[0] == dset_info for dset_info in dataset_infos):
|
| 274 |
+
# if all dataset_infos are equal we don't need to merge. Just return the first.
|
| 275 |
+
return dataset_infos[0]
|
| 276 |
+
|
| 277 |
+
description = "\n\n".join(unique_values(info.description for info in dataset_infos)).strip()
|
| 278 |
+
citation = "\n\n".join(unique_values(info.citation for info in dataset_infos)).strip()
|
| 279 |
+
homepage = "\n\n".join(unique_values(info.homepage for info in dataset_infos)).strip()
|
| 280 |
+
license = "\n\n".join(unique_values(info.license for info in dataset_infos)).strip()
|
| 281 |
+
features = None
|
| 282 |
+
supervised_keys = None
|
| 283 |
+
task_templates = None
|
| 284 |
+
|
| 285 |
+
# Find common task templates across all dataset infos
|
| 286 |
+
all_task_templates = [info.task_templates for info in dataset_infos if info.task_templates is not None]
|
| 287 |
+
if len(all_task_templates) > 1:
|
| 288 |
+
task_templates = list(set(all_task_templates[0]).intersection(*all_task_templates[1:]))
|
| 289 |
+
elif len(all_task_templates):
|
| 290 |
+
task_templates = list(set(all_task_templates[0]))
|
| 291 |
+
# If no common task templates found, replace empty list with None
|
| 292 |
+
task_templates = task_templates if task_templates else None
|
| 293 |
+
|
| 294 |
+
return cls(
|
| 295 |
+
description=description,
|
| 296 |
+
citation=citation,
|
| 297 |
+
homepage=homepage,
|
| 298 |
+
license=license,
|
| 299 |
+
features=features,
|
| 300 |
+
supervised_keys=supervised_keys,
|
| 301 |
+
task_templates=task_templates,
|
| 302 |
+
)
|
| 303 |
+
|
| 304 |
+
@classmethod
|
| 305 |
+
def from_directory(
|
| 306 |
+
cls, dataset_info_dir: str, fs="deprecated", storage_options: Optional[dict] = None
|
| 307 |
+
) -> "DatasetInfo":
|
| 308 |
+
"""Create [`DatasetInfo`] from the JSON file in `dataset_info_dir`.
|
| 309 |
+
|
| 310 |
+
This function updates all the dynamically generated fields (num_examples,
|
| 311 |
+
hash, time of creation,...) of the [`DatasetInfo`].
|
| 312 |
+
|
| 313 |
+
This will overwrite all previous metadata.
|
| 314 |
+
|
| 315 |
+
Args:
|
| 316 |
+
dataset_info_dir (`str`):
|
| 317 |
+
The directory containing the metadata file. This
|
| 318 |
+
should be the root directory of a specific dataset version.
|
| 319 |
+
fs (`fsspec.spec.AbstractFileSystem`, *optional*):
|
| 320 |
+
Instance of the remote filesystem used to download the files from.
|
| 321 |
+
|
| 322 |
+
<Deprecated version="2.9.0">
|
| 323 |
+
|
| 324 |
+
`fs` was deprecated in version 2.9.0 and will be removed in 3.0.0.
|
| 325 |
+
Please use `storage_options` instead, e.g. `storage_options=fs.storage_options`.
|
| 326 |
+
|
| 327 |
+
</Deprecated>
|
| 328 |
+
|
| 329 |
+
storage_options (`dict`, *optional*):
|
| 330 |
+
Key/value pairs to be passed on to the file-system backend, if any.
|
| 331 |
+
|
| 332 |
+
<Added version="2.9.0"/>
|
| 333 |
+
|
| 334 |
+
Example:
|
| 335 |
+
|
| 336 |
+
```py
|
| 337 |
+
>>> from datasets import DatasetInfo
|
| 338 |
+
>>> ds_info = DatasetInfo.from_directory("/path/to/directory/")
|
| 339 |
+
```
|
| 340 |
+
"""
|
| 341 |
+
if fs != "deprecated":
|
| 342 |
+
warnings.warn(
|
| 343 |
+
"'fs' was deprecated in favor of 'storage_options' in version 2.9.0 and will be removed in 3.0.0.\n"
|
| 344 |
+
"You can remove this warning by passing 'storage_options=fs.storage_options' instead.",
|
| 345 |
+
FutureWarning,
|
| 346 |
+
)
|
| 347 |
+
storage_options = fs.storage_options
|
| 348 |
+
|
| 349 |
+
fs: fsspec.AbstractFileSystem
|
| 350 |
+
fs, _, _ = fsspec.get_fs_token_paths(dataset_info_dir, storage_options=storage_options)
|
| 351 |
+
logger.info(f"Loading Dataset info from {dataset_info_dir}")
|
| 352 |
+
if not dataset_info_dir:
|
| 353 |
+
raise ValueError("Calling DatasetInfo.from_directory() with undefined dataset_info_dir.")
|
| 354 |
+
with fs.open(posixpath.join(dataset_info_dir, config.DATASET_INFO_FILENAME), "r", encoding="utf-8") as f:
|
| 355 |
+
dataset_info_dict = json.load(f)
|
| 356 |
+
return cls.from_dict(dataset_info_dict)
|
| 357 |
+
|
| 358 |
+
@classmethod
|
| 359 |
+
def from_dict(cls, dataset_info_dict: dict) -> "DatasetInfo":
|
| 360 |
+
field_names = {f.name for f in dataclasses.fields(cls)}
|
| 361 |
+
return cls(**{k: v for k, v in dataset_info_dict.items() if k in field_names})
|
| 362 |
+
|
| 363 |
+
def update(self, other_dataset_info: "DatasetInfo", ignore_none=True):
|
| 364 |
+
self_dict = self.__dict__
|
| 365 |
+
self_dict.update(
|
| 366 |
+
**{
|
| 367 |
+
k: copy.deepcopy(v)
|
| 368 |
+
for k, v in other_dataset_info.__dict__.items()
|
| 369 |
+
if (v is not None or not ignore_none)
|
| 370 |
+
}
|
| 371 |
+
)
|
| 372 |
+
|
| 373 |
+
def copy(self) -> "DatasetInfo":
|
| 374 |
+
return self.__class__(**{k: copy.deepcopy(v) for k, v in self.__dict__.items()})
|
| 375 |
+
|
| 376 |
+
def _to_yaml_dict(self) -> dict:
|
| 377 |
+
yaml_dict = {}
|
| 378 |
+
dataset_info_dict = asdict(self)
|
| 379 |
+
for key in dataset_info_dict:
|
| 380 |
+
if key in self._INCLUDED_INFO_IN_YAML:
|
| 381 |
+
value = getattr(self, key)
|
| 382 |
+
if hasattr(value, "_to_yaml_list"): # Features, SplitDict
|
| 383 |
+
yaml_dict[key] = value._to_yaml_list()
|
| 384 |
+
elif hasattr(value, "_to_yaml_string"): # Version
|
| 385 |
+
yaml_dict[key] = value._to_yaml_string()
|
| 386 |
+
else:
|
| 387 |
+
yaml_dict[key] = value
|
| 388 |
+
return yaml_dict
|
| 389 |
+
|
| 390 |
+
@classmethod
|
| 391 |
+
def _from_yaml_dict(cls, yaml_data: dict) -> "DatasetInfo":
|
| 392 |
+
yaml_data = copy.deepcopy(yaml_data)
|
| 393 |
+
if yaml_data.get("features") is not None:
|
| 394 |
+
yaml_data["features"] = Features._from_yaml_list(yaml_data["features"])
|
| 395 |
+
if yaml_data.get("splits") is not None:
|
| 396 |
+
yaml_data["splits"] = SplitDict._from_yaml_list(yaml_data["splits"])
|
| 397 |
+
field_names = {f.name for f in dataclasses.fields(cls)}
|
| 398 |
+
return cls(**{k: v for k, v in yaml_data.items() if k in field_names})
|
| 399 |
+
|
| 400 |
+
|
| 401 |
+
class DatasetInfosDict(Dict[str, DatasetInfo]):
|
| 402 |
+
def write_to_directory(self, dataset_infos_dir, overwrite=False, pretty_print=False) -> None:
|
| 403 |
+
total_dataset_infos = {}
|
| 404 |
+
dataset_infos_path = os.path.join(dataset_infos_dir, config.DATASETDICT_INFOS_FILENAME)
|
| 405 |
+
dataset_readme_path = os.path.join(dataset_infos_dir, config.REPOCARD_FILENAME)
|
| 406 |
+
if not overwrite:
|
| 407 |
+
total_dataset_infos = self.from_directory(dataset_infos_dir)
|
| 408 |
+
total_dataset_infos.update(self)
|
| 409 |
+
if os.path.exists(dataset_infos_path):
|
| 410 |
+
# for backward compatibility, let's update the JSON file if it exists
|
| 411 |
+
with open(dataset_infos_path, "w", encoding="utf-8") as f:
|
| 412 |
+
dataset_infos_dict = {
|
| 413 |
+
config_name: asdict(dset_info) for config_name, dset_info in total_dataset_infos.items()
|
| 414 |
+
}
|
| 415 |
+
json.dump(dataset_infos_dict, f, indent=4 if pretty_print else None)
|
| 416 |
+
# Dump the infos in the YAML part of the README.md file
|
| 417 |
+
if os.path.exists(dataset_readme_path):
|
| 418 |
+
dataset_card = DatasetCard.load(dataset_readme_path)
|
| 419 |
+
dataset_card_data = dataset_card.data
|
| 420 |
+
else:
|
| 421 |
+
dataset_card = None
|
| 422 |
+
dataset_card_data = DatasetCardData()
|
| 423 |
+
if total_dataset_infos:
|
| 424 |
+
total_dataset_infos.to_dataset_card_data(dataset_card_data)
|
| 425 |
+
dataset_card = (
|
| 426 |
+
DatasetCard("---\n" + str(dataset_card_data) + "\n---\n") if dataset_card is None else dataset_card
|
| 427 |
+
)
|
| 428 |
+
dataset_card.save(Path(dataset_readme_path))
|
| 429 |
+
|
| 430 |
+
@classmethod
|
| 431 |
+
def from_directory(cls, dataset_infos_dir) -> "DatasetInfosDict":
|
| 432 |
+
logger.info(f"Loading Dataset Infos from {dataset_infos_dir}")
|
| 433 |
+
# Load the info from the YAML part of README.md
|
| 434 |
+
if os.path.exists(os.path.join(dataset_infos_dir, config.REPOCARD_FILENAME)):
|
| 435 |
+
dataset_card_data = DatasetCard.load(Path(dataset_infos_dir) / config.REPOCARD_FILENAME).data
|
| 436 |
+
if "dataset_info" in dataset_card_data:
|
| 437 |
+
return cls.from_dataset_card_data(dataset_card_data)
|
| 438 |
+
if os.path.exists(os.path.join(dataset_infos_dir, config.DATASETDICT_INFOS_FILENAME)):
|
| 439 |
+
# this is just to have backward compatibility with dataset_infos.json files
|
| 440 |
+
with open(os.path.join(dataset_infos_dir, config.DATASETDICT_INFOS_FILENAME), encoding="utf-8") as f:
|
| 441 |
+
return cls(
|
| 442 |
+
{
|
| 443 |
+
config_name: DatasetInfo.from_dict(dataset_info_dict)
|
| 444 |
+
for config_name, dataset_info_dict in json.load(f).items()
|
| 445 |
+
}
|
| 446 |
+
)
|
| 447 |
+
else:
|
| 448 |
+
return cls()
|
| 449 |
+
|
| 450 |
+
@classmethod
|
| 451 |
+
def from_dataset_card_data(cls, dataset_card_data: DatasetCardData) -> "DatasetInfosDict":
|
| 452 |
+
if isinstance(dataset_card_data.get("dataset_info"), (list, dict)):
|
| 453 |
+
if isinstance(dataset_card_data["dataset_info"], list):
|
| 454 |
+
return cls(
|
| 455 |
+
{
|
| 456 |
+
dataset_info_yaml_dict.get("config_name", "default"): DatasetInfo._from_yaml_dict(
|
| 457 |
+
dataset_info_yaml_dict
|
| 458 |
+
)
|
| 459 |
+
for dataset_info_yaml_dict in dataset_card_data["dataset_info"]
|
| 460 |
+
}
|
| 461 |
+
)
|
| 462 |
+
else:
|
| 463 |
+
dataset_info = DatasetInfo._from_yaml_dict(dataset_card_data["dataset_info"])
|
| 464 |
+
dataset_info.config_name = dataset_card_data["dataset_info"].get("config_name", "default")
|
| 465 |
+
return cls({dataset_info.config_name: dataset_info})
|
| 466 |
+
else:
|
| 467 |
+
return cls()
|
| 468 |
+
|
| 469 |
+
def to_dataset_card_data(self, dataset_card_data: DatasetCardData) -> None:
|
| 470 |
+
if self:
|
| 471 |
+
# first get existing metadata info
|
| 472 |
+
if "dataset_info" in dataset_card_data and isinstance(dataset_card_data["dataset_info"], dict):
|
| 473 |
+
dataset_metadata_infos = {
|
| 474 |
+
dataset_card_data["dataset_info"].get("config_name", "default"): dataset_card_data["dataset_info"]
|
| 475 |
+
}
|
| 476 |
+
elif "dataset_info" in dataset_card_data and isinstance(dataset_card_data["dataset_info"], list):
|
| 477 |
+
dataset_metadata_infos = {
|
| 478 |
+
config_metadata["config_name"]: config_metadata
|
| 479 |
+
for config_metadata in dataset_card_data["dataset_info"]
|
| 480 |
+
}
|
| 481 |
+
else:
|
| 482 |
+
dataset_metadata_infos = {}
|
| 483 |
+
# update/rewrite existing metadata info with the one to dump
|
| 484 |
+
total_dataset_infos = {
|
| 485 |
+
**dataset_metadata_infos,
|
| 486 |
+
**{config_name: dset_info._to_yaml_dict() for config_name, dset_info in self.items()},
|
| 487 |
+
}
|
| 488 |
+
# the config_name from the dataset_infos_dict takes over the config_name of the DatasetInfo
|
| 489 |
+
for config_name, dset_info_yaml_dict in total_dataset_infos.items():
|
| 490 |
+
dset_info_yaml_dict["config_name"] = config_name
|
| 491 |
+
if len(total_dataset_infos) == 1:
|
| 492 |
+
# use a struct instead of a list of configurations, since there's only one
|
| 493 |
+
dataset_card_data["dataset_info"] = next(iter(total_dataset_infos.values()))
|
| 494 |
+
config_name = dataset_card_data["dataset_info"].pop("config_name", None)
|
| 495 |
+
if config_name != "default":
|
| 496 |
+
# if config_name is not "default" preserve it and put at the first position
|
| 497 |
+
dataset_card_data["dataset_info"] = {
|
| 498 |
+
"config_name": config_name,
|
| 499 |
+
**dataset_card_data["dataset_info"],
|
| 500 |
+
}
|
| 501 |
+
else:
|
| 502 |
+
dataset_card_data["dataset_info"] = []
|
| 503 |
+
for config_name, dataset_info_yaml_dict in sorted(total_dataset_infos.items()):
|
| 504 |
+
# add the config_name field in first position
|
| 505 |
+
dataset_info_yaml_dict.pop("config_name", None)
|
| 506 |
+
dataset_info_yaml_dict = {"config_name": config_name, **dataset_info_yaml_dict}
|
| 507 |
+
dataset_card_data["dataset_info"].append(dataset_info_yaml_dict)
|
| 508 |
+
|
| 509 |
+
|
| 510 |
+
@dataclass
|
| 511 |
+
class MetricInfo:
|
| 512 |
+
"""Information about a metric.
|
| 513 |
+
|
| 514 |
+
`MetricInfo` documents a metric, including its name, version, and features.
|
| 515 |
+
See the constructor arguments and properties for a full list.
|
| 516 |
+
|
| 517 |
+
Note: Not all fields are known on construction and may be updated later.
|
| 518 |
+
"""
|
| 519 |
+
|
| 520 |
+
# Set in the dataset scripts
|
| 521 |
+
description: str
|
| 522 |
+
citation: str
|
| 523 |
+
features: Features
|
| 524 |
+
inputs_description: str = dataclasses.field(default_factory=str)
|
| 525 |
+
homepage: str = dataclasses.field(default_factory=str)
|
| 526 |
+
license: str = dataclasses.field(default_factory=str)
|
| 527 |
+
codebase_urls: List[str] = dataclasses.field(default_factory=list)
|
| 528 |
+
reference_urls: List[str] = dataclasses.field(default_factory=list)
|
| 529 |
+
streamable: bool = False
|
| 530 |
+
format: Optional[str] = None
|
| 531 |
+
|
| 532 |
+
# Set later by the builder
|
| 533 |
+
metric_name: Optional[str] = None
|
| 534 |
+
config_name: Optional[str] = None
|
| 535 |
+
experiment_id: Optional[str] = None
|
| 536 |
+
|
| 537 |
+
def __post_init__(self):
|
| 538 |
+
if self.format is not None:
|
| 539 |
+
for key, value in self.features.items():
|
| 540 |
+
if not isinstance(value, Value):
|
| 541 |
+
raise ValueError(
|
| 542 |
+
f"When using 'numpy' format, all features should be a `datasets.Value` feature. "
|
| 543 |
+
f"Here {key} is an instance of {value.__class__.__name__}"
|
| 544 |
+
)
|
| 545 |
+
|
| 546 |
+
def write_to_directory(self, metric_info_dir, pretty_print=False):
|
| 547 |
+
"""Write `MetricInfo` as JSON to `metric_info_dir`.
|
| 548 |
+
Also save the license separately in LICENCE.
|
| 549 |
+
If `pretty_print` is True, the JSON will be pretty-printed with the indent level of 4.
|
| 550 |
+
|
| 551 |
+
Example:
|
| 552 |
+
|
| 553 |
+
```py
|
| 554 |
+
>>> from datasets import load_metric
|
| 555 |
+
>>> metric = load_metric("accuracy")
|
| 556 |
+
>>> metric.info.write_to_directory("/path/to/directory/")
|
| 557 |
+
```
|
| 558 |
+
"""
|
| 559 |
+
with open(os.path.join(metric_info_dir, config.METRIC_INFO_FILENAME), "w", encoding="utf-8") as f:
|
| 560 |
+
json.dump(asdict(self), f, indent=4 if pretty_print else None)
|
| 561 |
+
|
| 562 |
+
if self.license:
|
| 563 |
+
with open(os.path.join(metric_info_dir, config.LICENSE_FILENAME), "w", encoding="utf-8") as f:
|
| 564 |
+
f.write(self.license)
|
| 565 |
+
|
| 566 |
+
@classmethod
|
| 567 |
+
def from_directory(cls, metric_info_dir) -> "MetricInfo":
|
| 568 |
+
"""Create MetricInfo from the JSON file in `metric_info_dir`.
|
| 569 |
+
|
| 570 |
+
Args:
|
| 571 |
+
metric_info_dir: `str` The directory containing the metadata file. This
|
| 572 |
+
should be the root directory of a specific dataset version.
|
| 573 |
+
|
| 574 |
+
Example:
|
| 575 |
+
|
| 576 |
+
```py
|
| 577 |
+
>>> from datasets import MetricInfo
|
| 578 |
+
>>> metric_info = MetricInfo.from_directory("/path/to/directory/")
|
| 579 |
+
```
|
| 580 |
+
"""
|
| 581 |
+
logger.info(f"Loading Metric info from {metric_info_dir}")
|
| 582 |
+
if not metric_info_dir:
|
| 583 |
+
raise ValueError("Calling MetricInfo.from_directory() with undefined metric_info_dir.")
|
| 584 |
+
|
| 585 |
+
with open(os.path.join(metric_info_dir, config.METRIC_INFO_FILENAME), encoding="utf-8") as f:
|
| 586 |
+
metric_info_dict = json.load(f)
|
| 587 |
+
return cls.from_dict(metric_info_dict)
|
| 588 |
+
|
| 589 |
+
@classmethod
|
| 590 |
+
def from_dict(cls, metric_info_dict: dict) -> "MetricInfo":
|
| 591 |
+
field_names = {f.name for f in dataclasses.fields(cls)}
|
| 592 |
+
return cls(**{k: v for k, v in metric_info_dict.items() if k in field_names})
|
evalkit_tf437/lib/python3.10/site-packages/datasets/inspect.py
ADDED
|
@@ -0,0 +1,581 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright 2020 The HuggingFace Datasets Authors.
|
| 2 |
+
#
|
| 3 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
| 4 |
+
# you may not use this file except in compliance with the License.
|
| 5 |
+
# You may obtain a copy of the License at
|
| 6 |
+
#
|
| 7 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
| 8 |
+
#
|
| 9 |
+
# Unless required by applicable law or agreed to in writing, software
|
| 10 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
| 11 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| 12 |
+
# See the License for the specific language governing permissions and
|
| 13 |
+
# limitations under the License.
|
| 14 |
+
|
| 15 |
+
# Lint as: python3
|
| 16 |
+
"""List and inspect datasets."""
|
| 17 |
+
|
| 18 |
+
import inspect
|
| 19 |
+
import os
|
| 20 |
+
import shutil
|
| 21 |
+
import warnings
|
| 22 |
+
from pathlib import Path, PurePath
|
| 23 |
+
from typing import Dict, List, Mapping, Optional, Sequence, Union
|
| 24 |
+
|
| 25 |
+
import huggingface_hub
|
| 26 |
+
|
| 27 |
+
from . import config
|
| 28 |
+
from .download.download_config import DownloadConfig
|
| 29 |
+
from .download.download_manager import DownloadMode
|
| 30 |
+
from .download.streaming_download_manager import StreamingDownloadManager
|
| 31 |
+
from .info import DatasetInfo
|
| 32 |
+
from .load import (
|
| 33 |
+
dataset_module_factory,
|
| 34 |
+
get_dataset_builder_class,
|
| 35 |
+
import_main_class,
|
| 36 |
+
load_dataset_builder,
|
| 37 |
+
metric_module_factory,
|
| 38 |
+
)
|
| 39 |
+
from .utils.deprecation_utils import deprecated
|
| 40 |
+
from .utils.file_utils import relative_to_absolute_path
|
| 41 |
+
from .utils.logging import get_logger
|
| 42 |
+
from .utils.version import Version
|
| 43 |
+
|
| 44 |
+
|
| 45 |
+
logger = get_logger(__name__)
|
| 46 |
+
|
| 47 |
+
|
| 48 |
+
class SplitsNotFoundError(ValueError):
|
| 49 |
+
pass
|
| 50 |
+
|
| 51 |
+
|
| 52 |
+
@deprecated("Use 'huggingface_hub.list_datasets' instead.")
|
| 53 |
+
def list_datasets(with_community_datasets=True, with_details=False):
|
| 54 |
+
"""List all the datasets scripts available on the Hugging Face Hub.
|
| 55 |
+
|
| 56 |
+
Args:
|
| 57 |
+
with_community_datasets (`bool`, *optional*, defaults to `True`):
|
| 58 |
+
Include the community provided datasets.
|
| 59 |
+
with_details (`bool`, *optional*, defaults to `False`):
|
| 60 |
+
Return the full details on the datasets instead of only the short name.
|
| 61 |
+
|
| 62 |
+
Example:
|
| 63 |
+
|
| 64 |
+
```py
|
| 65 |
+
>>> from datasets import list_datasets
|
| 66 |
+
>>> list_datasets()
|
| 67 |
+
['acronym_identification',
|
| 68 |
+
'ade_corpus_v2',
|
| 69 |
+
'adversarial_qa',
|
| 70 |
+
'aeslc',
|
| 71 |
+
'afrikaans_ner_corpus',
|
| 72 |
+
'ag_news',
|
| 73 |
+
...
|
| 74 |
+
]
|
| 75 |
+
```
|
| 76 |
+
"""
|
| 77 |
+
datasets = huggingface_hub.list_datasets(full=with_details)
|
| 78 |
+
if not with_community_datasets:
|
| 79 |
+
datasets = [dataset for dataset in datasets if "/" not in dataset.id]
|
| 80 |
+
if not with_details:
|
| 81 |
+
datasets = [dataset.id for dataset in datasets]
|
| 82 |
+
return list(datasets)
|
| 83 |
+
|
| 84 |
+
|
| 85 |
+
@deprecated(
|
| 86 |
+
"Use 'evaluate.list_evaluation_modules' instead, from the new library π€ Evaluate: https://huggingface.co/docs/evaluate"
|
| 87 |
+
)
|
| 88 |
+
def list_metrics(with_community_metrics=True, with_details=False):
|
| 89 |
+
"""List all the metrics script available on the Hugging Face Hub.
|
| 90 |
+
|
| 91 |
+
<Deprecated version="2.5.0">
|
| 92 |
+
|
| 93 |
+
Use `evaluate.list_evaluation_modules` instead, from the new library π€ Evaluate: https://huggingface.co/docs/evaluate
|
| 94 |
+
|
| 95 |
+
</Deprecated>
|
| 96 |
+
|
| 97 |
+
Args:
|
| 98 |
+
with_community_metrics (:obj:`bool`, optional, default ``True``): Include the community provided metrics.
|
| 99 |
+
with_details (:obj:`bool`, optional, default ``False``): Return the full details on the metrics instead of only the short name.
|
| 100 |
+
|
| 101 |
+
Example:
|
| 102 |
+
|
| 103 |
+
```py
|
| 104 |
+
>>> from datasets import list_metrics
|
| 105 |
+
>>> list_metrics()
|
| 106 |
+
['accuracy',
|
| 107 |
+
'bertscore',
|
| 108 |
+
'bleu',
|
| 109 |
+
'bleurt',
|
| 110 |
+
'cer',
|
| 111 |
+
'chrf',
|
| 112 |
+
...
|
| 113 |
+
]
|
| 114 |
+
```
|
| 115 |
+
"""
|
| 116 |
+
metrics = huggingface_hub.list_metrics()
|
| 117 |
+
if not with_community_metrics:
|
| 118 |
+
metrics = [metric for metric in metrics if "/" not in metric.id]
|
| 119 |
+
if not with_details:
|
| 120 |
+
metrics = [metric.id for metric in metrics]
|
| 121 |
+
return metrics
|
| 122 |
+
|
| 123 |
+
|
| 124 |
+
@deprecated("Clone the dataset repository from the Hugging Face Hub instead.")
|
| 125 |
+
def inspect_dataset(path: str, local_path: str, download_config: Optional[DownloadConfig] = None, **download_kwargs):
|
| 126 |
+
"""
|
| 127 |
+
Allow inspection/modification of a dataset script by copying on local drive at local_path.
|
| 128 |
+
|
| 129 |
+
Args:
|
| 130 |
+
path (`str`): Path to the dataset processing script with the dataset builder. Can be either:
|
| 131 |
+
|
| 132 |
+
- a local path to processing script or the directory containing the script (if the script has the same name
|
| 133 |
+
as the directory),
|
| 134 |
+
e.g. `'./dataset/squad'` or `'./dataset/squad/squad.py'`.
|
| 135 |
+
- a dataset identifier on the Hugging Face Hub (list all available datasets and ids with [`list_datasets`])
|
| 136 |
+
e.g. `'squad'`, `'glue'` or `'openai/webtext'`.
|
| 137 |
+
local_path (`str`):
|
| 138 |
+
Path to the local folder to copy the dataset script to.
|
| 139 |
+
download_config ([`DownloadConfig`], *optional*):
|
| 140 |
+
Specific download configuration parameters.
|
| 141 |
+
**download_kwargs (additional keyword arguments):
|
| 142 |
+
Optional arguments for [`DownloadConfig`] which will override
|
| 143 |
+
the attributes of `download_config` if supplied.
|
| 144 |
+
"""
|
| 145 |
+
if download_config is None:
|
| 146 |
+
download_config = DownloadConfig(**download_kwargs)
|
| 147 |
+
if os.path.isfile(path):
|
| 148 |
+
path = str(Path(path).parent)
|
| 149 |
+
if os.path.isdir(path):
|
| 150 |
+
shutil.copytree(path, local_path, dirs_exist_ok=True)
|
| 151 |
+
else:
|
| 152 |
+
huggingface_hub.HfApi(endpoint=config.HF_ENDPOINT, token=download_config.token).snapshot_download(
|
| 153 |
+
repo_id=path, repo_type="dataset", local_dir=local_path, force_download=download_config.force_download
|
| 154 |
+
)
|
| 155 |
+
print(
|
| 156 |
+
f"The dataset {path} can be inspected at {local_path}. "
|
| 157 |
+
f'You can modify this loading script if it has one and use it with `datasets.load_dataset("{PurePath(local_path).as_posix()}")`.'
|
| 158 |
+
)
|
| 159 |
+
|
| 160 |
+
|
| 161 |
+
@deprecated(
|
| 162 |
+
"Use 'evaluate.inspect_evaluation_module' instead, from the new library π€ Evaluate: https://huggingface.co/docs/evaluate"
|
| 163 |
+
)
|
| 164 |
+
def inspect_metric(path: str, local_path: str, download_config: Optional[DownloadConfig] = None, **download_kwargs):
|
| 165 |
+
r"""
|
| 166 |
+
Allow inspection/modification of a metric script by copying it on local drive at local_path.
|
| 167 |
+
|
| 168 |
+
<Deprecated version="2.5.0">
|
| 169 |
+
|
| 170 |
+
Use `evaluate.inspect_evaluation_module` instead, from the new library π€ Evaluate instead: https://huggingface.co/docs/evaluate
|
| 171 |
+
|
| 172 |
+
</Deprecated>
|
| 173 |
+
|
| 174 |
+
Args:
|
| 175 |
+
path (``str``): path to the dataset processing script with the dataset builder. Can be either:
|
| 176 |
+
|
| 177 |
+
- a local path to processing script or the directory containing the script (if the script has the same name as the directory),
|
| 178 |
+
e.g. ``'./dataset/squad'`` or ``'./dataset/squad/squad.py'``
|
| 179 |
+
- a dataset identifier on the Hugging Face Hub (list all available datasets and ids with ``datasets.list_datasets()``)
|
| 180 |
+
e.g. ``'squad'``, ``'glue'`` or ``'openai/webtext'``
|
| 181 |
+
local_path (``str``): path to the local folder to copy the datset script to.
|
| 182 |
+
download_config (Optional ``datasets.DownloadConfig``): specific download configuration parameters.
|
| 183 |
+
**download_kwargs (additional keyword arguments): optional attributes for DownloadConfig() which will override the attributes in download_config if supplied.
|
| 184 |
+
"""
|
| 185 |
+
metric_module = metric_module_factory(path, download_config=download_config, **download_kwargs)
|
| 186 |
+
metric_cls = import_main_class(metric_module.module_path, dataset=False)
|
| 187 |
+
module_source_path = inspect.getsourcefile(metric_cls)
|
| 188 |
+
module_source_dirpath = os.path.dirname(module_source_path)
|
| 189 |
+
for dirpath, dirnames, filenames in os.walk(module_source_dirpath):
|
| 190 |
+
dst_dirpath = os.path.join(local_path, os.path.relpath(dirpath, module_source_dirpath))
|
| 191 |
+
os.makedirs(dst_dirpath, exist_ok=True)
|
| 192 |
+
# skipping hidden directories; prune the search
|
| 193 |
+
dirnames[:] = [dirname for dirname in dirnames if not dirname.startswith((".", "__"))]
|
| 194 |
+
for filename in filenames:
|
| 195 |
+
shutil.copy2(os.path.join(dirpath, filename), os.path.join(dst_dirpath, filename))
|
| 196 |
+
shutil.copystat(dirpath, dst_dirpath)
|
| 197 |
+
local_path = relative_to_absolute_path(local_path)
|
| 198 |
+
print(
|
| 199 |
+
f"The processing scripts for metric {path} can be inspected at {local_path}. "
|
| 200 |
+
f"The main class is in {module_source_dirpath}. "
|
| 201 |
+
f'You can modify this processing scripts and use it with `datasets.load_metric("{PurePath(local_path).as_posix()}")`.'
|
| 202 |
+
)
|
| 203 |
+
|
| 204 |
+
|
| 205 |
+
def get_dataset_infos(
|
| 206 |
+
path: str,
|
| 207 |
+
data_files: Optional[Union[Dict, List, str]] = None,
|
| 208 |
+
download_config: Optional[DownloadConfig] = None,
|
| 209 |
+
download_mode: Optional[Union[DownloadMode, str]] = None,
|
| 210 |
+
revision: Optional[Union[str, Version]] = None,
|
| 211 |
+
token: Optional[Union[bool, str]] = None,
|
| 212 |
+
use_auth_token="deprecated",
|
| 213 |
+
**config_kwargs,
|
| 214 |
+
):
|
| 215 |
+
"""Get the meta information about a dataset, returned as a dict mapping config name to DatasetInfoDict.
|
| 216 |
+
|
| 217 |
+
Args:
|
| 218 |
+
path (`str`): path to the dataset processing script with the dataset builder. Can be either:
|
| 219 |
+
|
| 220 |
+
- a local path to processing script or the directory containing the script (if the script has the same name as the directory),
|
| 221 |
+
e.g. `'./dataset/squad'` or `'./dataset/squad/squad.py'`
|
| 222 |
+
- a dataset identifier on the Hugging Face Hub (list all available datasets and ids with [`datasets.list_datasets`])
|
| 223 |
+
e.g. `'squad'`, `'glue'` or``'openai/webtext'`
|
| 224 |
+
revision (`Union[str, datasets.Version]`, *optional*):
|
| 225 |
+
If specified, the dataset module will be loaded from the datasets repository at this version.
|
| 226 |
+
By default:
|
| 227 |
+
- it is set to the local version of the lib.
|
| 228 |
+
- it will also try to load it from the main branch if it's not available at the local version of the lib.
|
| 229 |
+
Specifying a version that is different from your local version of the lib might cause compatibility issues.
|
| 230 |
+
download_config ([`DownloadConfig`], *optional*):
|
| 231 |
+
Specific download configuration parameters.
|
| 232 |
+
download_mode ([`DownloadMode`] or `str`, defaults to `REUSE_DATASET_IF_EXISTS`):
|
| 233 |
+
Download/generate mode.
|
| 234 |
+
data_files (`Union[Dict, List, str]`, *optional*):
|
| 235 |
+
Defining the data_files of the dataset configuration.
|
| 236 |
+
token (`str` or `bool`, *optional*):
|
| 237 |
+
Optional string or boolean to use as Bearer token for remote files on the Datasets Hub.
|
| 238 |
+
If `True`, or not specified, will get token from `"~/.huggingface"`.
|
| 239 |
+
use_auth_token (`str` or `bool`, *optional*):
|
| 240 |
+
Optional string or boolean to use as Bearer token for remote files on the Datasets Hub.
|
| 241 |
+
If `True`, or not specified, will get token from `"~/.huggingface"`.
|
| 242 |
+
|
| 243 |
+
<Deprecated version="2.14.0">
|
| 244 |
+
|
| 245 |
+
`use_auth_token` was deprecated in favor of `token` in version 2.14.0 and will be removed in 3.0.0.
|
| 246 |
+
|
| 247 |
+
</Deprecated>
|
| 248 |
+
|
| 249 |
+
**config_kwargs (additional keyword arguments):
|
| 250 |
+
Optional attributes for builder class which will override the attributes if supplied.
|
| 251 |
+
|
| 252 |
+
Example:
|
| 253 |
+
|
| 254 |
+
```py
|
| 255 |
+
>>> from datasets import get_dataset_infos
|
| 256 |
+
>>> get_dataset_infos('rotten_tomatoes')
|
| 257 |
+
{'default': DatasetInfo(description="Movie Review Dataset.\nThis is a dataset of containing 5,331 positive and 5,331 negative processed\nsentences from Rotten Tomatoes movie reviews...), ...}
|
| 258 |
+
```
|
| 259 |
+
"""
|
| 260 |
+
if use_auth_token != "deprecated":
|
| 261 |
+
warnings.warn(
|
| 262 |
+
"'use_auth_token' was deprecated in favor of 'token' in version 2.14.0 and will be removed in 3.0.0.\n"
|
| 263 |
+
"You can remove this warning by passing 'token=<use_auth_token>' instead.",
|
| 264 |
+
FutureWarning,
|
| 265 |
+
)
|
| 266 |
+
token = use_auth_token
|
| 267 |
+
|
| 268 |
+
config_names = get_dataset_config_names(
|
| 269 |
+
path=path,
|
| 270 |
+
revision=revision,
|
| 271 |
+
download_config=download_config,
|
| 272 |
+
download_mode=download_mode,
|
| 273 |
+
data_files=data_files,
|
| 274 |
+
token=token,
|
| 275 |
+
)
|
| 276 |
+
return {
|
| 277 |
+
config_name: get_dataset_config_info(
|
| 278 |
+
path=path,
|
| 279 |
+
config_name=config_name,
|
| 280 |
+
data_files=data_files,
|
| 281 |
+
download_config=download_config,
|
| 282 |
+
download_mode=download_mode,
|
| 283 |
+
revision=revision,
|
| 284 |
+
token=token,
|
| 285 |
+
**config_kwargs,
|
| 286 |
+
)
|
| 287 |
+
for config_name in config_names
|
| 288 |
+
}
|
| 289 |
+
|
| 290 |
+
|
| 291 |
+
def get_dataset_config_names(
|
| 292 |
+
path: str,
|
| 293 |
+
revision: Optional[Union[str, Version]] = None,
|
| 294 |
+
download_config: Optional[DownloadConfig] = None,
|
| 295 |
+
download_mode: Optional[Union[DownloadMode, str]] = None,
|
| 296 |
+
dynamic_modules_path: Optional[str] = None,
|
| 297 |
+
data_files: Optional[Union[Dict, List, str]] = None,
|
| 298 |
+
**download_kwargs,
|
| 299 |
+
):
|
| 300 |
+
"""Get the list of available config names for a particular dataset.
|
| 301 |
+
|
| 302 |
+
Args:
|
| 303 |
+
path (`str`): path to the dataset processing script with the dataset builder. Can be either:
|
| 304 |
+
|
| 305 |
+
- a local path to processing script or the directory containing the script (if the script has the same name as the directory),
|
| 306 |
+
e.g. `'./dataset/squad'` or `'./dataset/squad/squad.py'`
|
| 307 |
+
- a dataset identifier on the Hugging Face Hub (list all available datasets and ids with [`datasets.list_datasets`])
|
| 308 |
+
e.g. `'squad'`, `'glue'` or `'openai/webtext'`
|
| 309 |
+
revision (`Union[str, datasets.Version]`, *optional*):
|
| 310 |
+
If specified, the dataset module will be loaded from the datasets repository at this version.
|
| 311 |
+
By default:
|
| 312 |
+
- it is set to the local version of the lib.
|
| 313 |
+
- it will also try to load it from the main branch if it's not available at the local version of the lib.
|
| 314 |
+
Specifying a version that is different from your local version of the lib might cause compatibility issues.
|
| 315 |
+
download_config ([`DownloadConfig`], *optional*):
|
| 316 |
+
Specific download configuration parameters.
|
| 317 |
+
download_mode ([`DownloadMode`] or `str`, defaults to `REUSE_DATASET_IF_EXISTS`):
|
| 318 |
+
Download/generate mode.
|
| 319 |
+
dynamic_modules_path (`str`, defaults to `~/.cache/huggingface/modules/datasets_modules`):
|
| 320 |
+
Optional path to the directory in which the dynamic modules are saved. It must have been initialized with `init_dynamic_modules`.
|
| 321 |
+
By default the datasets and metrics are stored inside the `datasets_modules` module.
|
| 322 |
+
data_files (`Union[Dict, List, str]`, *optional*):
|
| 323 |
+
Defining the data_files of the dataset configuration.
|
| 324 |
+
**download_kwargs (additional keyword arguments):
|
| 325 |
+
Optional attributes for [`DownloadConfig`] which will override the attributes in `download_config` if supplied,
|
| 326 |
+
for example `token`.
|
| 327 |
+
|
| 328 |
+
Example:
|
| 329 |
+
|
| 330 |
+
```py
|
| 331 |
+
>>> from datasets import get_dataset_config_names
|
| 332 |
+
>>> get_dataset_config_names("glue")
|
| 333 |
+
['cola',
|
| 334 |
+
'sst2',
|
| 335 |
+
'mrpc',
|
| 336 |
+
'qqp',
|
| 337 |
+
'stsb',
|
| 338 |
+
'mnli',
|
| 339 |
+
'mnli_mismatched',
|
| 340 |
+
'mnli_matched',
|
| 341 |
+
'qnli',
|
| 342 |
+
'rte',
|
| 343 |
+
'wnli',
|
| 344 |
+
'ax']
|
| 345 |
+
```
|
| 346 |
+
"""
|
| 347 |
+
dataset_module = dataset_module_factory(
|
| 348 |
+
path,
|
| 349 |
+
revision=revision,
|
| 350 |
+
download_config=download_config,
|
| 351 |
+
download_mode=download_mode,
|
| 352 |
+
dynamic_modules_path=dynamic_modules_path,
|
| 353 |
+
data_files=data_files,
|
| 354 |
+
**download_kwargs,
|
| 355 |
+
)
|
| 356 |
+
builder_cls = get_dataset_builder_class(dataset_module, dataset_name=os.path.basename(path))
|
| 357 |
+
return list(builder_cls.builder_configs.keys()) or [
|
| 358 |
+
dataset_module.builder_kwargs.get("config_name", builder_cls.DEFAULT_CONFIG_NAME or "default")
|
| 359 |
+
]
|
| 360 |
+
|
| 361 |
+
|
| 362 |
+
def get_dataset_default_config_name(
|
| 363 |
+
path: str,
|
| 364 |
+
revision: Optional[Union[str, Version]] = None,
|
| 365 |
+
download_config: Optional[DownloadConfig] = None,
|
| 366 |
+
download_mode: Optional[Union[DownloadMode, str]] = None,
|
| 367 |
+
dynamic_modules_path: Optional[str] = None,
|
| 368 |
+
data_files: Optional[Union[Dict, List, str]] = None,
|
| 369 |
+
**download_kwargs,
|
| 370 |
+
) -> Optional[str]:
|
| 371 |
+
"""Get the default config name for a particular dataset.
|
| 372 |
+
|
| 373 |
+
Args:
|
| 374 |
+
path (`str`): path to the dataset processing script with the dataset builder. Can be either:
|
| 375 |
+
|
| 376 |
+
- a local path to processing script or the directory containing the script (if the script has the same name as the directory),
|
| 377 |
+
e.g. `'./dataset/squad'` or `'./dataset/squad/squad.py'`
|
| 378 |
+
- a dataset identifier on the Hugging Face Hub (list all available datasets and ids with [`datasets.list_datasets`])
|
| 379 |
+
e.g. `'squad'`, `'glue'` or `'openai/webtext'`
|
| 380 |
+
revision (`Union[str, datasets.Version]`, *optional*):
|
| 381 |
+
If specified, the dataset module will be loaded from the datasets repository at this version.
|
| 382 |
+
By default:
|
| 383 |
+
- it is set to the local version of the lib.
|
| 384 |
+
- it will also try to load it from the main branch if it's not available at the local version of the lib.
|
| 385 |
+
Specifying a version that is different from your local version of the lib might cause compatibility issues.
|
| 386 |
+
download_config ([`DownloadConfig`], *optional*):
|
| 387 |
+
Specific download configuration parameters.
|
| 388 |
+
download_mode ([`DownloadMode`] or `str`, defaults to `REUSE_DATASET_IF_EXISTS`):
|
| 389 |
+
Download/generate mode.
|
| 390 |
+
dynamic_modules_path (`str`, defaults to `~/.cache/huggingface/modules/datasets_modules`):
|
| 391 |
+
Optional path to the directory in which the dynamic modules are saved. It must have been initialized with `init_dynamic_modules`.
|
| 392 |
+
By default the datasets and metrics are stored inside the `datasets_modules` module.
|
| 393 |
+
data_files (`Union[Dict, List, str]`, *optional*):
|
| 394 |
+
Defining the data_files of the dataset configuration.
|
| 395 |
+
**download_kwargs (additional keyword arguments):
|
| 396 |
+
Optional attributes for [`DownloadConfig`] which will override the attributes in `download_config` if supplied,
|
| 397 |
+
for example `token`.
|
| 398 |
+
|
| 399 |
+
Returns:
|
| 400 |
+
Optional[str]
|
| 401 |
+
|
| 402 |
+
Example:
|
| 403 |
+
|
| 404 |
+
```py
|
| 405 |
+
>>> from datasets import get_dataset_default_config_name
|
| 406 |
+
>>> get_dataset_default_config_name("openbookqa")
|
| 407 |
+
'main'
|
| 408 |
+
```
|
| 409 |
+
"""
|
| 410 |
+
dataset_module = dataset_module_factory(
|
| 411 |
+
path,
|
| 412 |
+
revision=revision,
|
| 413 |
+
download_config=download_config,
|
| 414 |
+
download_mode=download_mode,
|
| 415 |
+
dynamic_modules_path=dynamic_modules_path,
|
| 416 |
+
data_files=data_files,
|
| 417 |
+
**download_kwargs,
|
| 418 |
+
)
|
| 419 |
+
builder_cls = get_dataset_builder_class(dataset_module, dataset_name=os.path.basename(path))
|
| 420 |
+
builder_configs = list(builder_cls.builder_configs.keys())
|
| 421 |
+
if builder_configs:
|
| 422 |
+
default_config_name = builder_configs[0] if len(builder_configs) == 1 else None
|
| 423 |
+
else:
|
| 424 |
+
default_config_name = "default"
|
| 425 |
+
return builder_cls.DEFAULT_CONFIG_NAME or default_config_name
|
| 426 |
+
|
| 427 |
+
|
| 428 |
+
def get_dataset_config_info(
|
| 429 |
+
path: str,
|
| 430 |
+
config_name: Optional[str] = None,
|
| 431 |
+
data_files: Optional[Union[str, Sequence[str], Mapping[str, Union[str, Sequence[str]]]]] = None,
|
| 432 |
+
download_config: Optional[DownloadConfig] = None,
|
| 433 |
+
download_mode: Optional[Union[DownloadMode, str]] = None,
|
| 434 |
+
revision: Optional[Union[str, Version]] = None,
|
| 435 |
+
token: Optional[Union[bool, str]] = None,
|
| 436 |
+
use_auth_token="deprecated",
|
| 437 |
+
**config_kwargs,
|
| 438 |
+
) -> DatasetInfo:
|
| 439 |
+
"""Get the meta information (DatasetInfo) about a dataset for a particular config
|
| 440 |
+
|
| 441 |
+
Args:
|
| 442 |
+
path (``str``): path to the dataset processing script with the dataset builder. Can be either:
|
| 443 |
+
|
| 444 |
+
- a local path to processing script or the directory containing the script (if the script has the same name as the directory),
|
| 445 |
+
e.g. ``'./dataset/squad'`` or ``'./dataset/squad/squad.py'``
|
| 446 |
+
- a dataset identifier on the Hugging Face Hub (list all available datasets and ids with ``datasets.list_datasets()``)
|
| 447 |
+
e.g. ``'squad'``, ``'glue'`` or ``'openai/webtext'``
|
| 448 |
+
config_name (:obj:`str`, optional): Defining the name of the dataset configuration.
|
| 449 |
+
data_files (:obj:`str` or :obj:`Sequence` or :obj:`Mapping`, optional): Path(s) to source data file(s).
|
| 450 |
+
download_config (:class:`~download.DownloadConfig`, optional): Specific download configuration parameters.
|
| 451 |
+
download_mode (:class:`DownloadMode` or :obj:`str`, default ``REUSE_DATASET_IF_EXISTS``): Download/generate mode.
|
| 452 |
+
revision (:class:`~utils.Version` or :obj:`str`, optional): Version of the dataset script to load.
|
| 453 |
+
As datasets have their own git repository on the Datasets Hub, the default version "main" corresponds to their "main" branch.
|
| 454 |
+
You can specify a different version than the default "main" by using a commit SHA or a git tag of the dataset repository.
|
| 455 |
+
token (``str`` or :obj:`bool`, optional): Optional string or boolean to use as Bearer token for remote files on the Datasets Hub.
|
| 456 |
+
If True, or not specified, will get token from `"~/.huggingface"`.
|
| 457 |
+
use_auth_token (``str`` or :obj:`bool`, optional): Optional string or boolean to use as Bearer token for remote files on the Datasets Hub.
|
| 458 |
+
If True, or not specified, will get token from `"~/.huggingface"`.
|
| 459 |
+
|
| 460 |
+
<Deprecated version="2.14.0">
|
| 461 |
+
|
| 462 |
+
`use_auth_token` was deprecated in favor of `token` in version 2.14.0 and will be removed in 3.0.0.
|
| 463 |
+
|
| 464 |
+
</Deprecated>
|
| 465 |
+
|
| 466 |
+
**config_kwargs (additional keyword arguments): optional attributes for builder class which will override the attributes if supplied.
|
| 467 |
+
|
| 468 |
+
"""
|
| 469 |
+
if use_auth_token != "deprecated":
|
| 470 |
+
warnings.warn(
|
| 471 |
+
"'use_auth_token' was deprecated in favor of 'token' in version 2.14.0 and will be removed in 3.0.0.\n"
|
| 472 |
+
"You can remove this warning by passing 'token=<use_auth_token>' instead.",
|
| 473 |
+
FutureWarning,
|
| 474 |
+
)
|
| 475 |
+
token = use_auth_token
|
| 476 |
+
|
| 477 |
+
builder = load_dataset_builder(
|
| 478 |
+
path,
|
| 479 |
+
name=config_name,
|
| 480 |
+
data_files=data_files,
|
| 481 |
+
download_config=download_config,
|
| 482 |
+
download_mode=download_mode,
|
| 483 |
+
revision=revision,
|
| 484 |
+
token=token,
|
| 485 |
+
**config_kwargs,
|
| 486 |
+
)
|
| 487 |
+
info = builder.info
|
| 488 |
+
if info.splits is None:
|
| 489 |
+
download_config = download_config.copy() if download_config else DownloadConfig()
|
| 490 |
+
if token is not None:
|
| 491 |
+
download_config.token = token
|
| 492 |
+
builder._check_manual_download(
|
| 493 |
+
StreamingDownloadManager(base_path=builder.base_path, download_config=download_config)
|
| 494 |
+
)
|
| 495 |
+
try:
|
| 496 |
+
info.splits = {
|
| 497 |
+
split_generator.name: {"name": split_generator.name, "dataset_name": path}
|
| 498 |
+
for split_generator in builder._split_generators(
|
| 499 |
+
StreamingDownloadManager(base_path=builder.base_path, download_config=download_config)
|
| 500 |
+
)
|
| 501 |
+
}
|
| 502 |
+
except Exception as err:
|
| 503 |
+
raise SplitsNotFoundError("The split names could not be parsed from the dataset config.") from err
|
| 504 |
+
return info
|
| 505 |
+
|
| 506 |
+
|
| 507 |
+
def get_dataset_split_names(
|
| 508 |
+
path: str,
|
| 509 |
+
config_name: Optional[str] = None,
|
| 510 |
+
data_files: Optional[Union[str, Sequence[str], Mapping[str, Union[str, Sequence[str]]]]] = None,
|
| 511 |
+
download_config: Optional[DownloadConfig] = None,
|
| 512 |
+
download_mode: Optional[Union[DownloadMode, str]] = None,
|
| 513 |
+
revision: Optional[Union[str, Version]] = None,
|
| 514 |
+
token: Optional[Union[bool, str]] = None,
|
| 515 |
+
use_auth_token="deprecated",
|
| 516 |
+
**config_kwargs,
|
| 517 |
+
):
|
| 518 |
+
"""Get the list of available splits for a particular config and dataset.
|
| 519 |
+
|
| 520 |
+
Args:
|
| 521 |
+
path (`str`): path to the dataset processing script with the dataset builder. Can be either:
|
| 522 |
+
|
| 523 |
+
- a local path to processing script or the directory containing the script (if the script has the same name as the directory),
|
| 524 |
+
e.g. `'./dataset/squad'` or `'./dataset/squad/squad.py'`
|
| 525 |
+
- a dataset identifier on the Hugging Face Hub (list all available datasets and ids with [`datasets.list_datasets`])
|
| 526 |
+
e.g. `'squad'`, `'glue'` or `'openai/webtext'`
|
| 527 |
+
config_name (`str`, *optional*):
|
| 528 |
+
Defining the name of the dataset configuration.
|
| 529 |
+
data_files (`str` or `Sequence` or `Mapping`, *optional*):
|
| 530 |
+
Path(s) to source data file(s).
|
| 531 |
+
download_config ([`DownloadConfig`], *optional*):
|
| 532 |
+
Specific download configuration parameters.
|
| 533 |
+
download_mode ([`DownloadMode`] or `str`, defaults to `REUSE_DATASET_IF_EXISTS`):
|
| 534 |
+
Download/generate mode.
|
| 535 |
+
revision ([`Version`] or `str`, *optional*):
|
| 536 |
+
Version of the dataset script to load.
|
| 537 |
+
As datasets have their own git repository on the Datasets Hub, the default version "main" corresponds to their "main" branch.
|
| 538 |
+
You can specify a different version than the default "main" by using a commit SHA or a git tag of the dataset repository.
|
| 539 |
+
token (`str` or `bool`, *optional*):
|
| 540 |
+
Optional string or boolean to use as Bearer token for remote files on the Datasets Hub.
|
| 541 |
+
If `True`, or not specified, will get token from `"~/.huggingface"`.
|
| 542 |
+
use_auth_token (`str` or `bool`, *optional*):
|
| 543 |
+
Optional string or boolean to use as Bearer token for remote files on the Datasets Hub.
|
| 544 |
+
If `True`, or not specified, will get token from `"~/.huggingface"`.
|
| 545 |
+
|
| 546 |
+
<Deprecated version="2.14.0">
|
| 547 |
+
|
| 548 |
+
`use_auth_token` was deprecated in favor of `token` in version 2.14.0 and will be removed in 3.0.0.
|
| 549 |
+
|
| 550 |
+
</Deprecated>
|
| 551 |
+
|
| 552 |
+
**config_kwargs (additional keyword arguments):
|
| 553 |
+
Optional attributes for builder class which will override the attributes if supplied.
|
| 554 |
+
|
| 555 |
+
Example:
|
| 556 |
+
|
| 557 |
+
```py
|
| 558 |
+
>>> from datasets import get_dataset_split_names
|
| 559 |
+
>>> get_dataset_split_names('rotten_tomatoes')
|
| 560 |
+
['train', 'validation', 'test']
|
| 561 |
+
```
|
| 562 |
+
"""
|
| 563 |
+
if use_auth_token != "deprecated":
|
| 564 |
+
warnings.warn(
|
| 565 |
+
"'use_auth_token' was deprecated in favor of 'token' in version 2.14.0 and will be removed in 3.0.0.\n"
|
| 566 |
+
"You can remove this warning by passing 'token=<use_auth_token>' instead.",
|
| 567 |
+
FutureWarning,
|
| 568 |
+
)
|
| 569 |
+
token = use_auth_token
|
| 570 |
+
|
| 571 |
+
info = get_dataset_config_info(
|
| 572 |
+
path,
|
| 573 |
+
config_name=config_name,
|
| 574 |
+
data_files=data_files,
|
| 575 |
+
download_config=download_config,
|
| 576 |
+
download_mode=download_mode,
|
| 577 |
+
revision=revision,
|
| 578 |
+
token=token,
|
| 579 |
+
**config_kwargs,
|
| 580 |
+
)
|
| 581 |
+
return list(info.splits.keys())
|
evalkit_tf437/lib/python3.10/site-packages/datasets/io/__pycache__/csv.cpython-310.pyc
ADDED
|
Binary file (4.37 kB). View file
|
|
|
evalkit_tf437/lib/python3.10/site-packages/datasets/io/__pycache__/generator.cpython-310.pyc
ADDED
|
Binary file (1.65 kB). View file
|
|
|
evalkit_tf437/lib/python3.10/site-packages/datasets/io/__pycache__/json.cpython-310.pyc
ADDED
|
Binary file (4.95 kB). View file
|
|
|
evalkit_tf437/lib/python3.10/site-packages/datasets/io/__pycache__/parquet.cpython-310.pyc
ADDED
|
Binary file (5.24 kB). View file
|
|
|
evalkit_tf437/lib/python3.10/site-packages/datasets/io/__pycache__/sql.cpython-310.pyc
ADDED
|
Binary file (3.94 kB). View file
|
|
|
evalkit_tf437/lib/python3.10/site-packages/datasets/io/__pycache__/text.cpython-310.pyc
ADDED
|
Binary file (1.73 kB). View file
|
|
|
evalkit_tf437/lib/python3.10/site-packages/datasets/io/abc.py
ADDED
|
@@ -0,0 +1,53 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from abc import ABC, abstractmethod
|
| 2 |
+
from typing import Optional, Union
|
| 3 |
+
|
| 4 |
+
from .. import Dataset, DatasetDict, Features, IterableDataset, IterableDatasetDict, NamedSplit
|
| 5 |
+
from ..utils.typing import NestedDataStructureLike, PathLike
|
| 6 |
+
|
| 7 |
+
|
| 8 |
+
class AbstractDatasetReader(ABC):
|
| 9 |
+
def __init__(
|
| 10 |
+
self,
|
| 11 |
+
path_or_paths: Optional[NestedDataStructureLike[PathLike]] = None,
|
| 12 |
+
split: Optional[NamedSplit] = None,
|
| 13 |
+
features: Optional[Features] = None,
|
| 14 |
+
cache_dir: str = None,
|
| 15 |
+
keep_in_memory: bool = False,
|
| 16 |
+
streaming: bool = False,
|
| 17 |
+
num_proc: Optional[int] = None,
|
| 18 |
+
**kwargs,
|
| 19 |
+
):
|
| 20 |
+
self.path_or_paths = path_or_paths
|
| 21 |
+
self.split = split if split or isinstance(path_or_paths, dict) else "train"
|
| 22 |
+
self.features = features
|
| 23 |
+
self.cache_dir = cache_dir
|
| 24 |
+
self.keep_in_memory = keep_in_memory
|
| 25 |
+
self.streaming = streaming
|
| 26 |
+
self.num_proc = num_proc
|
| 27 |
+
self.kwargs = kwargs
|
| 28 |
+
|
| 29 |
+
@abstractmethod
|
| 30 |
+
def read(self) -> Union[Dataset, DatasetDict, IterableDataset, IterableDatasetDict]:
|
| 31 |
+
pass
|
| 32 |
+
|
| 33 |
+
|
| 34 |
+
class AbstractDatasetInputStream(ABC):
|
| 35 |
+
def __init__(
|
| 36 |
+
self,
|
| 37 |
+
features: Optional[Features] = None,
|
| 38 |
+
cache_dir: str = None,
|
| 39 |
+
keep_in_memory: bool = False,
|
| 40 |
+
streaming: bool = False,
|
| 41 |
+
num_proc: Optional[int] = None,
|
| 42 |
+
**kwargs,
|
| 43 |
+
):
|
| 44 |
+
self.features = features
|
| 45 |
+
self.cache_dir = cache_dir
|
| 46 |
+
self.keep_in_memory = keep_in_memory
|
| 47 |
+
self.streaming = streaming
|
| 48 |
+
self.num_proc = num_proc
|
| 49 |
+
self.kwargs = kwargs
|
| 50 |
+
|
| 51 |
+
@abstractmethod
|
| 52 |
+
def read(self) -> Union[Dataset, IterableDataset]:
|
| 53 |
+
pass
|
evalkit_tf437/lib/python3.10/site-packages/datasets/io/generator.py
ADDED
|
@@ -0,0 +1,58 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from typing import Callable, Optional
|
| 2 |
+
|
| 3 |
+
from .. import Features
|
| 4 |
+
from ..packaged_modules.generator.generator import Generator
|
| 5 |
+
from .abc import AbstractDatasetInputStream
|
| 6 |
+
|
| 7 |
+
|
| 8 |
+
class GeneratorDatasetInputStream(AbstractDatasetInputStream):
|
| 9 |
+
def __init__(
|
| 10 |
+
self,
|
| 11 |
+
generator: Callable,
|
| 12 |
+
features: Optional[Features] = None,
|
| 13 |
+
cache_dir: str = None,
|
| 14 |
+
keep_in_memory: bool = False,
|
| 15 |
+
streaming: bool = False,
|
| 16 |
+
gen_kwargs: Optional[dict] = None,
|
| 17 |
+
num_proc: Optional[int] = None,
|
| 18 |
+
**kwargs,
|
| 19 |
+
):
|
| 20 |
+
super().__init__(
|
| 21 |
+
features=features,
|
| 22 |
+
cache_dir=cache_dir,
|
| 23 |
+
keep_in_memory=keep_in_memory,
|
| 24 |
+
streaming=streaming,
|
| 25 |
+
num_proc=num_proc,
|
| 26 |
+
**kwargs,
|
| 27 |
+
)
|
| 28 |
+
self.builder = Generator(
|
| 29 |
+
cache_dir=cache_dir,
|
| 30 |
+
features=features,
|
| 31 |
+
generator=generator,
|
| 32 |
+
gen_kwargs=gen_kwargs,
|
| 33 |
+
**kwargs,
|
| 34 |
+
)
|
| 35 |
+
|
| 36 |
+
def read(self):
|
| 37 |
+
# Build iterable dataset
|
| 38 |
+
if self.streaming:
|
| 39 |
+
dataset = self.builder.as_streaming_dataset(split="train")
|
| 40 |
+
# Build regular (map-style) dataset
|
| 41 |
+
else:
|
| 42 |
+
download_config = None
|
| 43 |
+
download_mode = None
|
| 44 |
+
verification_mode = None
|
| 45 |
+
base_path = None
|
| 46 |
+
|
| 47 |
+
self.builder.download_and_prepare(
|
| 48 |
+
download_config=download_config,
|
| 49 |
+
download_mode=download_mode,
|
| 50 |
+
verification_mode=verification_mode,
|
| 51 |
+
try_from_hf_gcs=False,
|
| 52 |
+
base_path=base_path,
|
| 53 |
+
num_proc=self.num_proc,
|
| 54 |
+
)
|
| 55 |
+
dataset = self.builder.as_dataset(
|
| 56 |
+
split="train", verification_mode=verification_mode, in_memory=self.keep_in_memory
|
| 57 |
+
)
|
| 58 |
+
return dataset
|
evalkit_tf437/lib/python3.10/site-packages/datasets/io/json.py
ADDED
|
@@ -0,0 +1,166 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import multiprocessing
|
| 2 |
+
import os
|
| 3 |
+
from typing import BinaryIO, Optional, Union
|
| 4 |
+
|
| 5 |
+
import fsspec
|
| 6 |
+
|
| 7 |
+
from .. import Dataset, Features, NamedSplit, config
|
| 8 |
+
from ..formatting import query_table
|
| 9 |
+
from ..packaged_modules.json.json import Json
|
| 10 |
+
from ..utils import tqdm as hf_tqdm
|
| 11 |
+
from ..utils.typing import NestedDataStructureLike, PathLike
|
| 12 |
+
from .abc import AbstractDatasetReader
|
| 13 |
+
|
| 14 |
+
|
| 15 |
+
class JsonDatasetReader(AbstractDatasetReader):
|
| 16 |
+
def __init__(
|
| 17 |
+
self,
|
| 18 |
+
path_or_paths: NestedDataStructureLike[PathLike],
|
| 19 |
+
split: Optional[NamedSplit] = None,
|
| 20 |
+
features: Optional[Features] = None,
|
| 21 |
+
cache_dir: str = None,
|
| 22 |
+
keep_in_memory: bool = False,
|
| 23 |
+
streaming: bool = False,
|
| 24 |
+
field: Optional[str] = None,
|
| 25 |
+
num_proc: Optional[int] = None,
|
| 26 |
+
**kwargs,
|
| 27 |
+
):
|
| 28 |
+
super().__init__(
|
| 29 |
+
path_or_paths,
|
| 30 |
+
split=split,
|
| 31 |
+
features=features,
|
| 32 |
+
cache_dir=cache_dir,
|
| 33 |
+
keep_in_memory=keep_in_memory,
|
| 34 |
+
streaming=streaming,
|
| 35 |
+
num_proc=num_proc,
|
| 36 |
+
**kwargs,
|
| 37 |
+
)
|
| 38 |
+
self.field = field
|
| 39 |
+
path_or_paths = path_or_paths if isinstance(path_or_paths, dict) else {self.split: path_or_paths}
|
| 40 |
+
self.builder = Json(
|
| 41 |
+
cache_dir=cache_dir,
|
| 42 |
+
data_files=path_or_paths,
|
| 43 |
+
features=features,
|
| 44 |
+
field=field,
|
| 45 |
+
**kwargs,
|
| 46 |
+
)
|
| 47 |
+
|
| 48 |
+
def read(self):
|
| 49 |
+
# Build iterable dataset
|
| 50 |
+
if self.streaming:
|
| 51 |
+
dataset = self.builder.as_streaming_dataset(split=self.split)
|
| 52 |
+
# Build regular (map-style) dataset
|
| 53 |
+
else:
|
| 54 |
+
download_config = None
|
| 55 |
+
download_mode = None
|
| 56 |
+
verification_mode = None
|
| 57 |
+
base_path = None
|
| 58 |
+
|
| 59 |
+
self.builder.download_and_prepare(
|
| 60 |
+
download_config=download_config,
|
| 61 |
+
download_mode=download_mode,
|
| 62 |
+
verification_mode=verification_mode,
|
| 63 |
+
# try_from_hf_gcs=try_from_hf_gcs,
|
| 64 |
+
base_path=base_path,
|
| 65 |
+
num_proc=self.num_proc,
|
| 66 |
+
)
|
| 67 |
+
dataset = self.builder.as_dataset(
|
| 68 |
+
split=self.split, verification_mode=verification_mode, in_memory=self.keep_in_memory
|
| 69 |
+
)
|
| 70 |
+
return dataset
|
| 71 |
+
|
| 72 |
+
|
| 73 |
+
class JsonDatasetWriter:
|
| 74 |
+
def __init__(
|
| 75 |
+
self,
|
| 76 |
+
dataset: Dataset,
|
| 77 |
+
path_or_buf: Union[PathLike, BinaryIO],
|
| 78 |
+
batch_size: Optional[int] = None,
|
| 79 |
+
num_proc: Optional[int] = None,
|
| 80 |
+
**to_json_kwargs,
|
| 81 |
+
):
|
| 82 |
+
if num_proc is not None and num_proc <= 0:
|
| 83 |
+
raise ValueError(f"num_proc {num_proc} must be an integer > 0.")
|
| 84 |
+
|
| 85 |
+
self.dataset = dataset
|
| 86 |
+
self.path_or_buf = path_or_buf
|
| 87 |
+
self.batch_size = batch_size if batch_size else config.DEFAULT_MAX_BATCH_SIZE
|
| 88 |
+
self.num_proc = num_proc
|
| 89 |
+
self.encoding = "utf-8"
|
| 90 |
+
self.to_json_kwargs = to_json_kwargs
|
| 91 |
+
|
| 92 |
+
def write(self) -> int:
|
| 93 |
+
_ = self.to_json_kwargs.pop("path_or_buf", None)
|
| 94 |
+
orient = self.to_json_kwargs.pop("orient", "records")
|
| 95 |
+
lines = self.to_json_kwargs.pop("lines", True if orient == "records" else False)
|
| 96 |
+
if "index" not in self.to_json_kwargs and orient in ["split", "table"]:
|
| 97 |
+
self.to_json_kwargs["index"] = False
|
| 98 |
+
|
| 99 |
+
# Determine the default compression value based on self.path_or_buf type
|
| 100 |
+
default_compression = "infer" if isinstance(self.path_or_buf, (str, bytes, os.PathLike)) else None
|
| 101 |
+
compression = self.to_json_kwargs.pop("compression", default_compression)
|
| 102 |
+
|
| 103 |
+
if compression not in [None, "infer", "gzip", "bz2", "xz"]:
|
| 104 |
+
raise NotImplementedError(f"`datasets` currently does not support {compression} compression")
|
| 105 |
+
|
| 106 |
+
if isinstance(self.path_or_buf, (str, bytes, os.PathLike)):
|
| 107 |
+
with fsspec.open(self.path_or_buf, "wb", compression=compression) as buffer:
|
| 108 |
+
written = self._write(file_obj=buffer, orient=orient, lines=lines, **self.to_json_kwargs)
|
| 109 |
+
else:
|
| 110 |
+
if compression:
|
| 111 |
+
raise NotImplementedError(
|
| 112 |
+
f"The compression parameter is not supported when writing to a buffer, but compression={compression}"
|
| 113 |
+
" was passed. Please provide a local path instead."
|
| 114 |
+
)
|
| 115 |
+
written = self._write(file_obj=self.path_or_buf, orient=orient, lines=lines, **self.to_json_kwargs)
|
| 116 |
+
return written
|
| 117 |
+
|
| 118 |
+
def _batch_json(self, args):
|
| 119 |
+
offset, orient, lines, to_json_kwargs = args
|
| 120 |
+
|
| 121 |
+
batch = query_table(
|
| 122 |
+
table=self.dataset.data,
|
| 123 |
+
key=slice(offset, offset + self.batch_size),
|
| 124 |
+
indices=self.dataset._indices,
|
| 125 |
+
)
|
| 126 |
+
json_str = batch.to_pandas().to_json(path_or_buf=None, orient=orient, lines=lines, **to_json_kwargs)
|
| 127 |
+
if not json_str.endswith("\n"):
|
| 128 |
+
json_str += "\n"
|
| 129 |
+
return json_str.encode(self.encoding)
|
| 130 |
+
|
| 131 |
+
def _write(
|
| 132 |
+
self,
|
| 133 |
+
file_obj: BinaryIO,
|
| 134 |
+
orient,
|
| 135 |
+
lines,
|
| 136 |
+
**to_json_kwargs,
|
| 137 |
+
) -> int:
|
| 138 |
+
"""Writes the pyarrow table as JSON lines to a binary file handle.
|
| 139 |
+
|
| 140 |
+
Caller is responsible for opening and closing the handle.
|
| 141 |
+
"""
|
| 142 |
+
written = 0
|
| 143 |
+
|
| 144 |
+
if self.num_proc is None or self.num_proc == 1:
|
| 145 |
+
for offset in hf_tqdm(
|
| 146 |
+
range(0, len(self.dataset), self.batch_size),
|
| 147 |
+
unit="ba",
|
| 148 |
+
desc="Creating json from Arrow format",
|
| 149 |
+
):
|
| 150 |
+
json_str = self._batch_json((offset, orient, lines, to_json_kwargs))
|
| 151 |
+
written += file_obj.write(json_str)
|
| 152 |
+
else:
|
| 153 |
+
num_rows, batch_size = len(self.dataset), self.batch_size
|
| 154 |
+
with multiprocessing.Pool(self.num_proc) as pool:
|
| 155 |
+
for json_str in hf_tqdm(
|
| 156 |
+
pool.imap(
|
| 157 |
+
self._batch_json,
|
| 158 |
+
[(offset, orient, lines, to_json_kwargs) for offset in range(0, num_rows, batch_size)],
|
| 159 |
+
),
|
| 160 |
+
total=(num_rows // batch_size) + 1 if num_rows % batch_size else num_rows // batch_size,
|
| 161 |
+
unit="ba",
|
| 162 |
+
desc="Creating json from Arrow format",
|
| 163 |
+
):
|
| 164 |
+
written += file_obj.write(json_str)
|
| 165 |
+
|
| 166 |
+
return written
|
evalkit_tf437/lib/python3.10/site-packages/datasets/io/parquet.py
ADDED
|
@@ -0,0 +1,156 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import os
|
| 2 |
+
from typing import BinaryIO, Optional, Union
|
| 3 |
+
|
| 4 |
+
import numpy as np
|
| 5 |
+
import pyarrow.parquet as pq
|
| 6 |
+
|
| 7 |
+
from .. import Audio, Dataset, Features, Image, NamedSplit, Value, config
|
| 8 |
+
from ..features.features import FeatureType, _visit
|
| 9 |
+
from ..formatting import query_table
|
| 10 |
+
from ..packaged_modules import _PACKAGED_DATASETS_MODULES
|
| 11 |
+
from ..packaged_modules.parquet.parquet import Parquet
|
| 12 |
+
from ..utils import tqdm as hf_tqdm
|
| 13 |
+
from ..utils.typing import NestedDataStructureLike, PathLike
|
| 14 |
+
from .abc import AbstractDatasetReader
|
| 15 |
+
|
| 16 |
+
|
| 17 |
+
def get_writer_batch_size(features: Features) -> Optional[int]:
|
| 18 |
+
"""
|
| 19 |
+
Get the writer_batch_size that defines the maximum row group size in the parquet files.
|
| 20 |
+
The default in `datasets` is 1,000 but we lower it to 100 for image datasets.
|
| 21 |
+
This allows to optimize random access to parquet file, since accessing 1 row requires
|
| 22 |
+
to read its entire row group.
|
| 23 |
+
|
| 24 |
+
This can be improved to get optimized size for querying/iterating
|
| 25 |
+
but at least it matches the dataset viewer expectations on HF.
|
| 26 |
+
|
| 27 |
+
Args:
|
| 28 |
+
ds_config_info (`datasets.info.DatasetInfo`):
|
| 29 |
+
Dataset info from `datasets`.
|
| 30 |
+
Returns:
|
| 31 |
+
writer_batch_size (`Optional[int]`):
|
| 32 |
+
Writer batch size to pass to a dataset builder.
|
| 33 |
+
If `None`, then it will use the `datasets` default.
|
| 34 |
+
"""
|
| 35 |
+
|
| 36 |
+
batch_size = np.inf
|
| 37 |
+
|
| 38 |
+
def set_batch_size(feature: FeatureType) -> None:
|
| 39 |
+
nonlocal batch_size
|
| 40 |
+
if isinstance(feature, Image):
|
| 41 |
+
batch_size = min(batch_size, config.PARQUET_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS)
|
| 42 |
+
elif isinstance(feature, Audio):
|
| 43 |
+
batch_size = min(batch_size, config.PARQUET_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS)
|
| 44 |
+
elif isinstance(feature, Value) and feature.dtype == "binary":
|
| 45 |
+
batch_size = min(batch_size, config.PARQUET_ROW_GROUP_SIZE_FOR_BINARY_DATASETS)
|
| 46 |
+
|
| 47 |
+
_visit(features, set_batch_size)
|
| 48 |
+
|
| 49 |
+
return None if batch_size is np.inf else batch_size
|
| 50 |
+
|
| 51 |
+
|
| 52 |
+
class ParquetDatasetReader(AbstractDatasetReader):
|
| 53 |
+
def __init__(
|
| 54 |
+
self,
|
| 55 |
+
path_or_paths: NestedDataStructureLike[PathLike],
|
| 56 |
+
split: Optional[NamedSplit] = None,
|
| 57 |
+
features: Optional[Features] = None,
|
| 58 |
+
cache_dir: str = None,
|
| 59 |
+
keep_in_memory: bool = False,
|
| 60 |
+
streaming: bool = False,
|
| 61 |
+
num_proc: Optional[int] = None,
|
| 62 |
+
**kwargs,
|
| 63 |
+
):
|
| 64 |
+
super().__init__(
|
| 65 |
+
path_or_paths,
|
| 66 |
+
split=split,
|
| 67 |
+
features=features,
|
| 68 |
+
cache_dir=cache_dir,
|
| 69 |
+
keep_in_memory=keep_in_memory,
|
| 70 |
+
streaming=streaming,
|
| 71 |
+
num_proc=num_proc,
|
| 72 |
+
**kwargs,
|
| 73 |
+
)
|
| 74 |
+
path_or_paths = path_or_paths if isinstance(path_or_paths, dict) else {self.split: path_or_paths}
|
| 75 |
+
hash = _PACKAGED_DATASETS_MODULES["parquet"][1]
|
| 76 |
+
self.builder = Parquet(
|
| 77 |
+
cache_dir=cache_dir,
|
| 78 |
+
data_files=path_or_paths,
|
| 79 |
+
features=features,
|
| 80 |
+
hash=hash,
|
| 81 |
+
**kwargs,
|
| 82 |
+
)
|
| 83 |
+
|
| 84 |
+
def read(self):
|
| 85 |
+
# Build iterable dataset
|
| 86 |
+
if self.streaming:
|
| 87 |
+
dataset = self.builder.as_streaming_dataset(split=self.split)
|
| 88 |
+
# Build regular (map-style) dataset
|
| 89 |
+
else:
|
| 90 |
+
download_config = None
|
| 91 |
+
download_mode = None
|
| 92 |
+
verification_mode = None
|
| 93 |
+
base_path = None
|
| 94 |
+
|
| 95 |
+
self.builder.download_and_prepare(
|
| 96 |
+
download_config=download_config,
|
| 97 |
+
download_mode=download_mode,
|
| 98 |
+
verification_mode=verification_mode,
|
| 99 |
+
# try_from_hf_gcs=try_from_hf_gcs,
|
| 100 |
+
base_path=base_path,
|
| 101 |
+
num_proc=self.num_proc,
|
| 102 |
+
)
|
| 103 |
+
dataset = self.builder.as_dataset(
|
| 104 |
+
split=self.split, verification_mode=verification_mode, in_memory=self.keep_in_memory
|
| 105 |
+
)
|
| 106 |
+
return dataset
|
| 107 |
+
|
| 108 |
+
|
| 109 |
+
class ParquetDatasetWriter:
|
| 110 |
+
def __init__(
|
| 111 |
+
self,
|
| 112 |
+
dataset: Dataset,
|
| 113 |
+
path_or_buf: Union[PathLike, BinaryIO],
|
| 114 |
+
batch_size: Optional[int] = None,
|
| 115 |
+
**parquet_writer_kwargs,
|
| 116 |
+
):
|
| 117 |
+
self.dataset = dataset
|
| 118 |
+
self.path_or_buf = path_or_buf
|
| 119 |
+
self.batch_size = batch_size or get_writer_batch_size(dataset.features)
|
| 120 |
+
self.parquet_writer_kwargs = parquet_writer_kwargs
|
| 121 |
+
|
| 122 |
+
def write(self) -> int:
|
| 123 |
+
batch_size = self.batch_size if self.batch_size else config.DEFAULT_MAX_BATCH_SIZE
|
| 124 |
+
|
| 125 |
+
if isinstance(self.path_or_buf, (str, bytes, os.PathLike)):
|
| 126 |
+
with open(self.path_or_buf, "wb+") as buffer:
|
| 127 |
+
written = self._write(file_obj=buffer, batch_size=batch_size, **self.parquet_writer_kwargs)
|
| 128 |
+
else:
|
| 129 |
+
written = self._write(file_obj=self.path_or_buf, batch_size=batch_size, **self.parquet_writer_kwargs)
|
| 130 |
+
return written
|
| 131 |
+
|
| 132 |
+
def _write(self, file_obj: BinaryIO, batch_size: int, **parquet_writer_kwargs) -> int:
|
| 133 |
+
"""Writes the pyarrow table as Parquet to a binary file handle.
|
| 134 |
+
|
| 135 |
+
Caller is responsible for opening and closing the handle.
|
| 136 |
+
"""
|
| 137 |
+
written = 0
|
| 138 |
+
_ = parquet_writer_kwargs.pop("path_or_buf", None)
|
| 139 |
+
schema = self.dataset.features.arrow_schema
|
| 140 |
+
|
| 141 |
+
writer = pq.ParquetWriter(file_obj, schema=schema, **parquet_writer_kwargs)
|
| 142 |
+
|
| 143 |
+
for offset in hf_tqdm(
|
| 144 |
+
range(0, len(self.dataset), batch_size),
|
| 145 |
+
unit="ba",
|
| 146 |
+
desc="Creating parquet from Arrow format",
|
| 147 |
+
):
|
| 148 |
+
batch = query_table(
|
| 149 |
+
table=self.dataset._data,
|
| 150 |
+
key=slice(offset, offset + batch_size),
|
| 151 |
+
indices=self.dataset._indices,
|
| 152 |
+
)
|
| 153 |
+
writer.write_table(batch)
|
| 154 |
+
written += batch.nbytes
|
| 155 |
+
writer.close()
|
| 156 |
+
return written
|
evalkit_tf437/lib/python3.10/site-packages/datasets/io/spark.py
ADDED
|
@@ -0,0 +1,57 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from typing import Optional
|
| 2 |
+
|
| 3 |
+
import pyspark
|
| 4 |
+
|
| 5 |
+
from .. import Features, NamedSplit
|
| 6 |
+
from ..download import DownloadMode
|
| 7 |
+
from ..packaged_modules.spark.spark import Spark
|
| 8 |
+
from .abc import AbstractDatasetReader
|
| 9 |
+
|
| 10 |
+
|
| 11 |
+
class SparkDatasetReader(AbstractDatasetReader):
|
| 12 |
+
"""A dataset reader that reads from a Spark DataFrame.
|
| 13 |
+
|
| 14 |
+
When caching, cache materialization is parallelized over Spark; an NFS that is accessible to the driver must be
|
| 15 |
+
provided. Streaming is not currently supported.
|
| 16 |
+
"""
|
| 17 |
+
|
| 18 |
+
def __init__(
|
| 19 |
+
self,
|
| 20 |
+
df: pyspark.sql.DataFrame,
|
| 21 |
+
split: Optional[NamedSplit] = None,
|
| 22 |
+
features: Optional[Features] = None,
|
| 23 |
+
streaming: bool = True,
|
| 24 |
+
cache_dir: str = None,
|
| 25 |
+
keep_in_memory: bool = False,
|
| 26 |
+
working_dir: str = None,
|
| 27 |
+
load_from_cache_file: bool = True,
|
| 28 |
+
file_format: str = "arrow",
|
| 29 |
+
**kwargs,
|
| 30 |
+
):
|
| 31 |
+
super().__init__(
|
| 32 |
+
split=split,
|
| 33 |
+
features=features,
|
| 34 |
+
cache_dir=cache_dir,
|
| 35 |
+
keep_in_memory=keep_in_memory,
|
| 36 |
+
streaming=streaming,
|
| 37 |
+
**kwargs,
|
| 38 |
+
)
|
| 39 |
+
self._load_from_cache_file = load_from_cache_file
|
| 40 |
+
self._file_format = file_format
|
| 41 |
+
self.builder = Spark(
|
| 42 |
+
df=df,
|
| 43 |
+
features=features,
|
| 44 |
+
cache_dir=cache_dir,
|
| 45 |
+
working_dir=working_dir,
|
| 46 |
+
**kwargs,
|
| 47 |
+
)
|
| 48 |
+
|
| 49 |
+
def read(self):
|
| 50 |
+
if self.streaming:
|
| 51 |
+
return self.builder.as_streaming_dataset(split=self.split)
|
| 52 |
+
download_mode = None if self._load_from_cache_file else DownloadMode.FORCE_REDOWNLOAD
|
| 53 |
+
self.builder.download_and_prepare(
|
| 54 |
+
download_mode=download_mode,
|
| 55 |
+
file_format=self._file_format,
|
| 56 |
+
)
|
| 57 |
+
return self.builder.as_dataset(split=self.split)
|
evalkit_tf437/lib/python3.10/site-packages/datasets/io/sql.py
ADDED
|
@@ -0,0 +1,125 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import multiprocessing
|
| 2 |
+
from typing import TYPE_CHECKING, Optional, Union
|
| 3 |
+
|
| 4 |
+
from .. import Dataset, Features, config
|
| 5 |
+
from ..formatting import query_table
|
| 6 |
+
from ..packaged_modules.sql.sql import Sql
|
| 7 |
+
from ..utils import tqdm as hf_tqdm
|
| 8 |
+
from .abc import AbstractDatasetInputStream
|
| 9 |
+
|
| 10 |
+
|
| 11 |
+
if TYPE_CHECKING:
|
| 12 |
+
import sqlite3
|
| 13 |
+
|
| 14 |
+
import sqlalchemy
|
| 15 |
+
|
| 16 |
+
|
| 17 |
+
class SqlDatasetReader(AbstractDatasetInputStream):
|
| 18 |
+
def __init__(
|
| 19 |
+
self,
|
| 20 |
+
sql: Union[str, "sqlalchemy.sql.Selectable"],
|
| 21 |
+
con: Union[str, "sqlalchemy.engine.Connection", "sqlalchemy.engine.Engine", "sqlite3.Connection"],
|
| 22 |
+
features: Optional[Features] = None,
|
| 23 |
+
cache_dir: str = None,
|
| 24 |
+
keep_in_memory: bool = False,
|
| 25 |
+
**kwargs,
|
| 26 |
+
):
|
| 27 |
+
super().__init__(features=features, cache_dir=cache_dir, keep_in_memory=keep_in_memory, **kwargs)
|
| 28 |
+
self.builder = Sql(
|
| 29 |
+
cache_dir=cache_dir,
|
| 30 |
+
features=features,
|
| 31 |
+
sql=sql,
|
| 32 |
+
con=con,
|
| 33 |
+
**kwargs,
|
| 34 |
+
)
|
| 35 |
+
|
| 36 |
+
def read(self):
|
| 37 |
+
download_config = None
|
| 38 |
+
download_mode = None
|
| 39 |
+
verification_mode = None
|
| 40 |
+
base_path = None
|
| 41 |
+
|
| 42 |
+
self.builder.download_and_prepare(
|
| 43 |
+
download_config=download_config,
|
| 44 |
+
download_mode=download_mode,
|
| 45 |
+
verification_mode=verification_mode,
|
| 46 |
+
# try_from_hf_gcs=try_from_hf_gcs,
|
| 47 |
+
base_path=base_path,
|
| 48 |
+
)
|
| 49 |
+
|
| 50 |
+
# Build dataset for splits
|
| 51 |
+
dataset = self.builder.as_dataset(
|
| 52 |
+
split="train", verification_mode=verification_mode, in_memory=self.keep_in_memory
|
| 53 |
+
)
|
| 54 |
+
return dataset
|
| 55 |
+
|
| 56 |
+
|
| 57 |
+
class SqlDatasetWriter:
|
| 58 |
+
def __init__(
|
| 59 |
+
self,
|
| 60 |
+
dataset: Dataset,
|
| 61 |
+
name: str,
|
| 62 |
+
con: Union[str, "sqlalchemy.engine.Connection", "sqlalchemy.engine.Engine", "sqlite3.Connection"],
|
| 63 |
+
batch_size: Optional[int] = None,
|
| 64 |
+
num_proc: Optional[int] = None,
|
| 65 |
+
**to_sql_kwargs,
|
| 66 |
+
):
|
| 67 |
+
if num_proc is not None and num_proc <= 0:
|
| 68 |
+
raise ValueError(f"num_proc {num_proc} must be an integer > 0.")
|
| 69 |
+
|
| 70 |
+
self.dataset = dataset
|
| 71 |
+
self.name = name
|
| 72 |
+
self.con = con
|
| 73 |
+
self.batch_size = batch_size if batch_size else config.DEFAULT_MAX_BATCH_SIZE
|
| 74 |
+
self.num_proc = num_proc
|
| 75 |
+
self.to_sql_kwargs = to_sql_kwargs
|
| 76 |
+
|
| 77 |
+
def write(self) -> int:
|
| 78 |
+
_ = self.to_sql_kwargs.pop("sql", None)
|
| 79 |
+
_ = self.to_sql_kwargs.pop("con", None)
|
| 80 |
+
index = self.to_sql_kwargs.pop("index", False)
|
| 81 |
+
|
| 82 |
+
written = self._write(index=index, **self.to_sql_kwargs)
|
| 83 |
+
return written
|
| 84 |
+
|
| 85 |
+
def _batch_sql(self, args):
|
| 86 |
+
offset, index, to_sql_kwargs = args
|
| 87 |
+
to_sql_kwargs = {**to_sql_kwargs, "if_exists": "append"} if offset > 0 else to_sql_kwargs
|
| 88 |
+
batch = query_table(
|
| 89 |
+
table=self.dataset.data,
|
| 90 |
+
key=slice(offset, offset + self.batch_size),
|
| 91 |
+
indices=self.dataset._indices,
|
| 92 |
+
)
|
| 93 |
+
df = batch.to_pandas()
|
| 94 |
+
num_rows = df.to_sql(self.name, self.con, index=index, **to_sql_kwargs)
|
| 95 |
+
return num_rows or len(df)
|
| 96 |
+
|
| 97 |
+
def _write(self, index, **to_sql_kwargs) -> int:
|
| 98 |
+
"""Writes the pyarrow table as SQL to a database.
|
| 99 |
+
|
| 100 |
+
Caller is responsible for opening and closing the SQL connection.
|
| 101 |
+
"""
|
| 102 |
+
written = 0
|
| 103 |
+
|
| 104 |
+
if self.num_proc is None or self.num_proc == 1:
|
| 105 |
+
for offset in hf_tqdm(
|
| 106 |
+
range(0, len(self.dataset), self.batch_size),
|
| 107 |
+
unit="ba",
|
| 108 |
+
desc="Creating SQL from Arrow format",
|
| 109 |
+
):
|
| 110 |
+
written += self._batch_sql((offset, index, to_sql_kwargs))
|
| 111 |
+
else:
|
| 112 |
+
num_rows, batch_size = len(self.dataset), self.batch_size
|
| 113 |
+
with multiprocessing.Pool(self.num_proc) as pool:
|
| 114 |
+
for num_rows in hf_tqdm(
|
| 115 |
+
pool.imap(
|
| 116 |
+
self._batch_sql,
|
| 117 |
+
[(offset, index, to_sql_kwargs) for offset in range(0, num_rows, batch_size)],
|
| 118 |
+
),
|
| 119 |
+
total=(num_rows // batch_size) + 1 if num_rows % batch_size else num_rows // batch_size,
|
| 120 |
+
unit="ba",
|
| 121 |
+
desc="Creating SQL from Arrow format",
|
| 122 |
+
):
|
| 123 |
+
written += num_rows
|
| 124 |
+
|
| 125 |
+
return written
|
evalkit_tf437/lib/python3.10/site-packages/datasets/io/text.py
ADDED
|
@@ -0,0 +1,61 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from typing import Optional
|
| 2 |
+
|
| 3 |
+
from .. import Features, NamedSplit
|
| 4 |
+
from ..packaged_modules.text.text import Text
|
| 5 |
+
from ..utils.typing import NestedDataStructureLike, PathLike
|
| 6 |
+
from .abc import AbstractDatasetReader
|
| 7 |
+
|
| 8 |
+
|
| 9 |
+
class TextDatasetReader(AbstractDatasetReader):
|
| 10 |
+
def __init__(
|
| 11 |
+
self,
|
| 12 |
+
path_or_paths: NestedDataStructureLike[PathLike],
|
| 13 |
+
split: Optional[NamedSplit] = None,
|
| 14 |
+
features: Optional[Features] = None,
|
| 15 |
+
cache_dir: str = None,
|
| 16 |
+
keep_in_memory: bool = False,
|
| 17 |
+
streaming: bool = False,
|
| 18 |
+
num_proc: Optional[int] = None,
|
| 19 |
+
**kwargs,
|
| 20 |
+
):
|
| 21 |
+
super().__init__(
|
| 22 |
+
path_or_paths,
|
| 23 |
+
split=split,
|
| 24 |
+
features=features,
|
| 25 |
+
cache_dir=cache_dir,
|
| 26 |
+
keep_in_memory=keep_in_memory,
|
| 27 |
+
streaming=streaming,
|
| 28 |
+
num_proc=num_proc,
|
| 29 |
+
**kwargs,
|
| 30 |
+
)
|
| 31 |
+
path_or_paths = path_or_paths if isinstance(path_or_paths, dict) else {self.split: path_or_paths}
|
| 32 |
+
self.builder = Text(
|
| 33 |
+
cache_dir=cache_dir,
|
| 34 |
+
data_files=path_or_paths,
|
| 35 |
+
features=features,
|
| 36 |
+
**kwargs,
|
| 37 |
+
)
|
| 38 |
+
|
| 39 |
+
def read(self):
|
| 40 |
+
# Build iterable dataset
|
| 41 |
+
if self.streaming:
|
| 42 |
+
dataset = self.builder.as_streaming_dataset(split=self.split)
|
| 43 |
+
# Build regular (map-style) dataset
|
| 44 |
+
else:
|
| 45 |
+
download_config = None
|
| 46 |
+
download_mode = None
|
| 47 |
+
verification_mode = None
|
| 48 |
+
base_path = None
|
| 49 |
+
|
| 50 |
+
self.builder.download_and_prepare(
|
| 51 |
+
download_config=download_config,
|
| 52 |
+
download_mode=download_mode,
|
| 53 |
+
verification_mode=verification_mode,
|
| 54 |
+
# try_from_hf_gcs=try_from_hf_gcs,
|
| 55 |
+
base_path=base_path,
|
| 56 |
+
num_proc=self.num_proc,
|
| 57 |
+
)
|
| 58 |
+
dataset = self.builder.as_dataset(
|
| 59 |
+
split=self.split, verification_mode=verification_mode, in_memory=self.keep_in_memory
|
| 60 |
+
)
|
| 61 |
+
return dataset
|
evalkit_tf437/lib/python3.10/site-packages/datasets/iterable_dataset.py
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
evalkit_tf437/lib/python3.10/site-packages/datasets/keyhash.py
ADDED
|
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors.
|
| 2 |
+
#
|
| 3 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
| 4 |
+
# you may not use this file except in compliance with the License.
|
| 5 |
+
# You may obtain a copy of the License at
|
| 6 |
+
#
|
| 7 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
| 8 |
+
#
|
| 9 |
+
# Unless required by applicable law or agreed to in writing, software
|
| 10 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
| 11 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| 12 |
+
# See the License for the specific language governing permissions and
|
| 13 |
+
# limitations under the License.
|
| 14 |
+
|
| 15 |
+
# Lint as: python3
|
| 16 |
+
|
| 17 |
+
"""
|
| 18 |
+
Hashing function for dataset keys using `hashlib.md5`
|
| 19 |
+
|
| 20 |
+
Requirements for the hash function:
|
| 21 |
+
|
| 22 |
+
- Provides a uniformly distributed hash from random space
|
| 23 |
+
- Adequately fast speed
|
| 24 |
+
- Working with multiple input types (in this case, `str`, `int` or `bytes`)
|
| 25 |
+
- Should be platform independent (generates same hash on different OS and systems)
|
| 26 |
+
|
| 27 |
+
The hashing function provides a unique 128-bit integer hash of the key provided.
|
| 28 |
+
|
| 29 |
+
The split name is being used here as the hash salt to avoid having same hashes
|
| 30 |
+
in different splits due to same keys
|
| 31 |
+
"""
|
| 32 |
+
|
| 33 |
+
from typing import Union
|
| 34 |
+
|
| 35 |
+
from huggingface_hub.utils import insecure_hashlib
|
| 36 |
+
|
| 37 |
+
|
| 38 |
+
def _as_bytes(hash_data: Union[str, int, bytes]) -> bytes:
|
| 39 |
+
"""
|
| 40 |
+
Returns the input hash_data in its bytes form
|
| 41 |
+
|
| 42 |
+
Args:
|
| 43 |
+
hash_data: the hash salt/key to be converted to bytes
|
| 44 |
+
"""
|
| 45 |
+
if isinstance(hash_data, bytes):
|
| 46 |
+
# Data already in bytes, returns as it as
|
| 47 |
+
return hash_data
|
| 48 |
+
elif isinstance(hash_data, str):
|
| 49 |
+
# We keep the data as it as for it ot be later encoded to UTF-8
|
| 50 |
+
# However replace `\\` with `/` for Windows compatibility
|
| 51 |
+
hash_data = hash_data.replace("\\", "/")
|
| 52 |
+
elif isinstance(hash_data, int):
|
| 53 |
+
hash_data = str(hash_data)
|
| 54 |
+
else:
|
| 55 |
+
# If data is not of the required type, raise error
|
| 56 |
+
raise InvalidKeyError(hash_data)
|
| 57 |
+
|
| 58 |
+
return hash_data.encode("utf-8")
|
| 59 |
+
|
| 60 |
+
|
| 61 |
+
class InvalidKeyError(Exception):
|
| 62 |
+
"""Raises an error when given key is of invalid datatype."""
|
| 63 |
+
|
| 64 |
+
def __init__(self, hash_data):
|
| 65 |
+
self.prefix = "\nFAILURE TO GENERATE DATASET: Invalid key type detected"
|
| 66 |
+
self.err_msg = f"\nFound Key {hash_data} of type {type(hash_data)}"
|
| 67 |
+
self.suffix = "\nKeys should be either str, int or bytes type"
|
| 68 |
+
super().__init__(f"{self.prefix}{self.err_msg}{self.suffix}")
|
| 69 |
+
|
| 70 |
+
|
| 71 |
+
class DuplicatedKeysError(Exception):
|
| 72 |
+
"""Raise an error when duplicate key found."""
|
| 73 |
+
|
| 74 |
+
def __init__(self, key, duplicate_key_indices, fix_msg=""):
|
| 75 |
+
self.key = key
|
| 76 |
+
self.duplicate_key_indices = duplicate_key_indices
|
| 77 |
+
self.fix_msg = fix_msg
|
| 78 |
+
self.prefix = "Found multiple examples generated with the same key"
|
| 79 |
+
if len(duplicate_key_indices) <= 20:
|
| 80 |
+
self.err_msg = f"\nThe examples at index {', '.join(duplicate_key_indices)} have the key {key}"
|
| 81 |
+
else:
|
| 82 |
+
self.err_msg = f"\nThe examples at index {', '.join(duplicate_key_indices[:20])}... ({len(duplicate_key_indices) - 20} more) have the key {key}"
|
| 83 |
+
self.suffix = "\n" + fix_msg if fix_msg else ""
|
| 84 |
+
super().__init__(f"{self.prefix}{self.err_msg}{self.suffix}")
|
| 85 |
+
|
| 86 |
+
|
| 87 |
+
class KeyHasher:
|
| 88 |
+
"""KeyHasher class for providing hash using md5"""
|
| 89 |
+
|
| 90 |
+
def __init__(self, hash_salt: str):
|
| 91 |
+
self._split_md5 = insecure_hashlib.md5(_as_bytes(hash_salt))
|
| 92 |
+
|
| 93 |
+
def hash(self, key: Union[str, int, bytes]) -> int:
|
| 94 |
+
"""Returns 128-bits unique hash of input key
|
| 95 |
+
|
| 96 |
+
Args:
|
| 97 |
+
key: the input key to be hashed (should be str, int or bytes)
|
| 98 |
+
|
| 99 |
+
Returns: 128-bit int hash key"""
|
| 100 |
+
md5 = self._split_md5.copy()
|
| 101 |
+
byte_key = _as_bytes(key)
|
| 102 |
+
md5.update(byte_key)
|
| 103 |
+
# Convert to integer with hexadecimal conversion
|
| 104 |
+
return int(md5.hexdigest(), 16)
|
evalkit_tf437/lib/python3.10/site-packages/datasets/metric.py
ADDED
|
@@ -0,0 +1,652 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright 2020 The HuggingFace Datasets Authors
|
| 2 |
+
#
|
| 3 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
| 4 |
+
# you may not use this file except in compliance with the License.
|
| 5 |
+
# You may obtain a copy of the License at
|
| 6 |
+
#
|
| 7 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
| 8 |
+
#
|
| 9 |
+
# Unless required by applicable law or agreed to in writing, software
|
| 10 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
| 11 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| 12 |
+
# See the License for the specific language governing permissions and
|
| 13 |
+
# limitations under the License.
|
| 14 |
+
|
| 15 |
+
# Lint as: python3
|
| 16 |
+
"""Metrics base class."""
|
| 17 |
+
|
| 18 |
+
import os
|
| 19 |
+
import types
|
| 20 |
+
import uuid
|
| 21 |
+
from typing import Any, Dict, List, Optional, Tuple, Union
|
| 22 |
+
|
| 23 |
+
import numpy as np
|
| 24 |
+
import pyarrow as pa
|
| 25 |
+
from filelock import BaseFileLock, Timeout
|
| 26 |
+
|
| 27 |
+
from . import config
|
| 28 |
+
from .arrow_dataset import Dataset
|
| 29 |
+
from .arrow_reader import ArrowReader
|
| 30 |
+
from .arrow_writer import ArrowWriter
|
| 31 |
+
from .download.download_config import DownloadConfig
|
| 32 |
+
from .download.download_manager import DownloadManager
|
| 33 |
+
from .features import Features
|
| 34 |
+
from .info import DatasetInfo, MetricInfo
|
| 35 |
+
from .naming import camelcase_to_snakecase
|
| 36 |
+
from .utils._filelock import FileLock
|
| 37 |
+
from .utils.deprecation_utils import deprecated
|
| 38 |
+
from .utils.logging import get_logger
|
| 39 |
+
from .utils.py_utils import copyfunc, temp_seed
|
| 40 |
+
|
| 41 |
+
|
| 42 |
+
logger = get_logger(__name__)
|
| 43 |
+
|
| 44 |
+
|
| 45 |
+
class FileFreeLock(BaseFileLock):
|
| 46 |
+
"""Thread lock until a file **cannot** be locked"""
|
| 47 |
+
|
| 48 |
+
def __init__(self, lock_file, *args, **kwargs):
|
| 49 |
+
self.filelock = FileLock(lock_file)
|
| 50 |
+
super().__init__(self.filelock.lock_file, *args, **kwargs)
|
| 51 |
+
|
| 52 |
+
def _acquire(self):
|
| 53 |
+
try:
|
| 54 |
+
self.filelock.acquire(timeout=0.01, poll_intervall=0.02) # Try to lock once
|
| 55 |
+
except Timeout:
|
| 56 |
+
# We couldn't acquire the lock, the file is locked!
|
| 57 |
+
self._context.lock_file_fd = self.filelock.lock_file
|
| 58 |
+
else:
|
| 59 |
+
# We were able to acquire the lock, the file is not yet locked!
|
| 60 |
+
self.filelock.release()
|
| 61 |
+
self._context.lock_file_fd = None
|
| 62 |
+
|
| 63 |
+
def _release(self):
|
| 64 |
+
self._context.lock_file_fd = None
|
| 65 |
+
|
| 66 |
+
|
| 67 |
+
# lists - summarize long lists similarly to NumPy
|
| 68 |
+
# arrays/tensors - let the frameworks control formatting
|
| 69 |
+
def summarize_if_long_list(obj):
|
| 70 |
+
if not type(obj) == list or len(obj) <= 6: # noqa: E721
|
| 71 |
+
return f"{obj}"
|
| 72 |
+
|
| 73 |
+
def format_chunk(chunk):
|
| 74 |
+
return ", ".join(repr(x) for x in chunk)
|
| 75 |
+
|
| 76 |
+
return f"[{format_chunk(obj[:3])}, ..., {format_chunk(obj[-3:])}]"
|
| 77 |
+
|
| 78 |
+
|
| 79 |
+
class MetricInfoMixin:
|
| 80 |
+
"""This base class exposes some attributes of MetricInfo
|
| 81 |
+
at the base level of the Metric for easy access.
|
| 82 |
+
|
| 83 |
+
<Deprecated version="2.5.0">
|
| 84 |
+
|
| 85 |
+
Use the new library π€ Evaluate instead: https://huggingface.co/docs/evaluate
|
| 86 |
+
|
| 87 |
+
</Deprecated>
|
| 88 |
+
|
| 89 |
+
"""
|
| 90 |
+
|
| 91 |
+
def __init__(self, info: MetricInfo):
|
| 92 |
+
self._metric_info = info
|
| 93 |
+
|
| 94 |
+
@property
|
| 95 |
+
def info(self):
|
| 96 |
+
""":class:`datasets.MetricInfo` object containing all the metadata in the metric."""
|
| 97 |
+
return self._metric_info
|
| 98 |
+
|
| 99 |
+
@property
|
| 100 |
+
def name(self) -> str:
|
| 101 |
+
return self._metric_info.metric_name
|
| 102 |
+
|
| 103 |
+
@property
|
| 104 |
+
def experiment_id(self) -> Optional[str]:
|
| 105 |
+
return self._metric_info.experiment_id
|
| 106 |
+
|
| 107 |
+
@property
|
| 108 |
+
def description(self) -> str:
|
| 109 |
+
return self._metric_info.description
|
| 110 |
+
|
| 111 |
+
@property
|
| 112 |
+
def citation(self) -> str:
|
| 113 |
+
return self._metric_info.citation
|
| 114 |
+
|
| 115 |
+
@property
|
| 116 |
+
def features(self) -> Features:
|
| 117 |
+
return self._metric_info.features
|
| 118 |
+
|
| 119 |
+
@property
|
| 120 |
+
def inputs_description(self) -> str:
|
| 121 |
+
return self._metric_info.inputs_description
|
| 122 |
+
|
| 123 |
+
@property
|
| 124 |
+
def homepage(self) -> Optional[str]:
|
| 125 |
+
return self._metric_info.homepage
|
| 126 |
+
|
| 127 |
+
@property
|
| 128 |
+
def license(self) -> str:
|
| 129 |
+
return self._metric_info.license
|
| 130 |
+
|
| 131 |
+
@property
|
| 132 |
+
def codebase_urls(self) -> Optional[List[str]]:
|
| 133 |
+
return self._metric_info.codebase_urls
|
| 134 |
+
|
| 135 |
+
@property
|
| 136 |
+
def reference_urls(self) -> Optional[List[str]]:
|
| 137 |
+
return self._metric_info.reference_urls
|
| 138 |
+
|
| 139 |
+
@property
|
| 140 |
+
def streamable(self) -> bool:
|
| 141 |
+
return self._metric_info.streamable
|
| 142 |
+
|
| 143 |
+
@property
|
| 144 |
+
def format(self) -> Optional[str]:
|
| 145 |
+
return self._metric_info.format
|
| 146 |
+
|
| 147 |
+
|
| 148 |
+
class Metric(MetricInfoMixin):
|
| 149 |
+
"""A Metric is the base class and common API for all metrics.
|
| 150 |
+
|
| 151 |
+
<Deprecated version="2.5.0">
|
| 152 |
+
|
| 153 |
+
Use the new library π€ Evaluate instead: https://huggingface.co/docs/evaluate
|
| 154 |
+
|
| 155 |
+
</Deprecated>
|
| 156 |
+
|
| 157 |
+
Args:
|
| 158 |
+
config_name (``str``): This is used to define a hash specific to a metrics computation script and prevents the metric's data
|
| 159 |
+
to be overridden when the metric loading script is modified.
|
| 160 |
+
keep_in_memory (:obj:`bool`): keep all predictions and references in memory. Not possible in distributed settings.
|
| 161 |
+
cache_dir (``str``): Path to a directory in which temporary prediction/references data will be stored.
|
| 162 |
+
The data directory should be located on a shared file-system in distributed setups.
|
| 163 |
+
num_process (``int``): specify the total number of nodes in a distributed settings.
|
| 164 |
+
This is useful to compute metrics in distributed setups (in particular non-additive metrics like F1).
|
| 165 |
+
process_id (``int``): specify the id of the current process in a distributed setup (between 0 and num_process-1)
|
| 166 |
+
This is useful to compute metrics in distributed setups (in particular non-additive metrics like F1).
|
| 167 |
+
seed (:obj:`int`, optional): If specified, this will temporarily set numpy's random seed when :func:`datasets.Metric.compute` is run.
|
| 168 |
+
experiment_id (``str``): A specific experiment id. This is used if several distributed evaluations share the same file system.
|
| 169 |
+
This is useful to compute metrics in distributed setups (in particular non-additive metrics like F1).
|
| 170 |
+
max_concurrent_cache_files (``int``): Max number of concurrent metrics cache files (default 10000).
|
| 171 |
+
timeout (``Union[int, float]``): Timeout in second for distributed setting synchronization.
|
| 172 |
+
"""
|
| 173 |
+
|
| 174 |
+
@deprecated("Use the new library π€ Evaluate instead: https://huggingface.co/docs/evaluate")
|
| 175 |
+
def __init__(
|
| 176 |
+
self,
|
| 177 |
+
config_name: Optional[str] = None,
|
| 178 |
+
keep_in_memory: bool = False,
|
| 179 |
+
cache_dir: Optional[str] = None,
|
| 180 |
+
num_process: int = 1,
|
| 181 |
+
process_id: int = 0,
|
| 182 |
+
seed: Optional[int] = None,
|
| 183 |
+
experiment_id: Optional[str] = None,
|
| 184 |
+
max_concurrent_cache_files: int = 10000,
|
| 185 |
+
timeout: Union[int, float] = 100,
|
| 186 |
+
**kwargs,
|
| 187 |
+
):
|
| 188 |
+
# prepare info
|
| 189 |
+
self.config_name = config_name or "default"
|
| 190 |
+
info = self._info()
|
| 191 |
+
info.metric_name = camelcase_to_snakecase(self.__class__.__name__)
|
| 192 |
+
info.config_name = self.config_name
|
| 193 |
+
info.experiment_id = experiment_id or "default_experiment"
|
| 194 |
+
MetricInfoMixin.__init__(self, info) # For easy access on low level
|
| 195 |
+
|
| 196 |
+
# Safety checks on num_process and process_id
|
| 197 |
+
if not isinstance(process_id, int) or process_id < 0:
|
| 198 |
+
raise ValueError("'process_id' should be a number greater than 0")
|
| 199 |
+
if not isinstance(num_process, int) or num_process <= process_id:
|
| 200 |
+
raise ValueError("'num_process' should be a number greater than process_id")
|
| 201 |
+
if keep_in_memory and num_process != 1:
|
| 202 |
+
raise ValueError("Using 'keep_in_memory' is not possible in distributed setting (num_process > 1).")
|
| 203 |
+
|
| 204 |
+
self.num_process = num_process
|
| 205 |
+
self.process_id = process_id
|
| 206 |
+
self.max_concurrent_cache_files = max_concurrent_cache_files
|
| 207 |
+
|
| 208 |
+
self.keep_in_memory = keep_in_memory
|
| 209 |
+
self._data_dir_root = os.path.expanduser(cache_dir or config.HF_METRICS_CACHE)
|
| 210 |
+
self.data_dir = self._build_data_dir()
|
| 211 |
+
if seed is None:
|
| 212 |
+
_, seed, pos, *_ = np.random.get_state()
|
| 213 |
+
self.seed: int = seed[pos] if pos < 624 else seed[0]
|
| 214 |
+
else:
|
| 215 |
+
self.seed: int = seed
|
| 216 |
+
self.timeout: Union[int, float] = timeout
|
| 217 |
+
|
| 218 |
+
# Update 'compute' and 'add' docstring
|
| 219 |
+
# methods need to be copied otherwise it changes the docstrings of every instance
|
| 220 |
+
self.compute = types.MethodType(copyfunc(self.compute), self)
|
| 221 |
+
self.add_batch = types.MethodType(copyfunc(self.add_batch), self)
|
| 222 |
+
self.add = types.MethodType(copyfunc(self.add), self)
|
| 223 |
+
self.compute.__func__.__doc__ += self.info.inputs_description
|
| 224 |
+
self.add_batch.__func__.__doc__ += self.info.inputs_description
|
| 225 |
+
self.add.__func__.__doc__ += self.info.inputs_description
|
| 226 |
+
|
| 227 |
+
# self.arrow_schema = pa.schema(field for field in self.info.features.type)
|
| 228 |
+
self.buf_writer = None
|
| 229 |
+
self.writer = None
|
| 230 |
+
self.writer_batch_size = None
|
| 231 |
+
self.data = None
|
| 232 |
+
|
| 233 |
+
# This is the cache file we store our predictions/references in
|
| 234 |
+
# Keep it None for now so we can (cloud)pickle the object
|
| 235 |
+
self.cache_file_name = None
|
| 236 |
+
self.filelock = None
|
| 237 |
+
self.rendez_vous_lock = None
|
| 238 |
+
|
| 239 |
+
# This is all the cache files on which we have a lock when we are in a distributed setting
|
| 240 |
+
self.file_paths = None
|
| 241 |
+
self.filelocks = None
|
| 242 |
+
|
| 243 |
+
def __len__(self):
|
| 244 |
+
"""Return the number of examples (predictions or predictions/references pair)
|
| 245 |
+
currently stored in the metric's cache.
|
| 246 |
+
"""
|
| 247 |
+
return 0 if self.writer is None else len(self.writer)
|
| 248 |
+
|
| 249 |
+
def __repr__(self):
|
| 250 |
+
return (
|
| 251 |
+
f'Metric(name: "{self.name}", features: {self.features}, '
|
| 252 |
+
f'usage: """{self.inputs_description}""", '
|
| 253 |
+
f"stored examples: {len(self)})"
|
| 254 |
+
)
|
| 255 |
+
|
| 256 |
+
def _build_data_dir(self):
|
| 257 |
+
"""Path of this metric in cache_dir:
|
| 258 |
+
Will be:
|
| 259 |
+
self._data_dir_root/self.name/self.config_name/self.hash (if not none)/
|
| 260 |
+
If any of these element is missing or if ``with_version=False`` the corresponding subfolders are dropped.
|
| 261 |
+
"""
|
| 262 |
+
builder_data_dir = self._data_dir_root
|
| 263 |
+
builder_data_dir = os.path.join(builder_data_dir, self.name, self.config_name)
|
| 264 |
+
os.makedirs(builder_data_dir, exist_ok=True)
|
| 265 |
+
return builder_data_dir
|
| 266 |
+
|
| 267 |
+
def _create_cache_file(self, timeout=1) -> Tuple[str, FileLock]:
|
| 268 |
+
"""Create a new cache file. If the default cache file is used, we generated a new hash."""
|
| 269 |
+
file_path = os.path.join(self.data_dir, f"{self.experiment_id}-{self.num_process}-{self.process_id}.arrow")
|
| 270 |
+
filelock = None
|
| 271 |
+
for i in range(self.max_concurrent_cache_files):
|
| 272 |
+
filelock = FileLock(file_path + ".lock")
|
| 273 |
+
try:
|
| 274 |
+
filelock.acquire(timeout=timeout)
|
| 275 |
+
except Timeout:
|
| 276 |
+
# If we have reached the max number of attempts or we are not allow to find a free name (distributed setup)
|
| 277 |
+
# We raise an error
|
| 278 |
+
if self.num_process != 1:
|
| 279 |
+
raise ValueError(
|
| 280 |
+
f"Error in _create_cache_file: another metric instance is already using the local cache file at {file_path}. "
|
| 281 |
+
f"Please specify an experiment_id (currently: {self.experiment_id}) to avoid collision "
|
| 282 |
+
f"between distributed metric instances."
|
| 283 |
+
) from None
|
| 284 |
+
if i == self.max_concurrent_cache_files - 1:
|
| 285 |
+
raise ValueError(
|
| 286 |
+
f"Cannot acquire lock, too many metric instance are operating concurrently on this file system."
|
| 287 |
+
f"You should set a larger value of max_concurrent_cache_files when creating the metric "
|
| 288 |
+
f"(current value is {self.max_concurrent_cache_files})."
|
| 289 |
+
) from None
|
| 290 |
+
# In other cases (allow to find new file name + not yet at max num of attempts) we can try to sample a new hashing name.
|
| 291 |
+
file_uuid = str(uuid.uuid4())
|
| 292 |
+
file_path = os.path.join(
|
| 293 |
+
self.data_dir, f"{self.experiment_id}-{file_uuid}-{self.num_process}-{self.process_id}.arrow"
|
| 294 |
+
)
|
| 295 |
+
else:
|
| 296 |
+
break
|
| 297 |
+
|
| 298 |
+
return file_path, filelock
|
| 299 |
+
|
| 300 |
+
def _get_all_cache_files(self) -> Tuple[List[str], List[FileLock]]:
|
| 301 |
+
"""Get a lock on all the cache files in a distributed setup.
|
| 302 |
+
We wait for timeout second to let all the distributed node finish their tasks (default is 100 seconds).
|
| 303 |
+
"""
|
| 304 |
+
if self.num_process == 1:
|
| 305 |
+
if self.cache_file_name is None:
|
| 306 |
+
raise ValueError(
|
| 307 |
+
"Metric cache file doesn't exist. Please make sure that you call `add` or `add_batch` "
|
| 308 |
+
"at least once before calling `compute`."
|
| 309 |
+
)
|
| 310 |
+
file_paths = [self.cache_file_name]
|
| 311 |
+
else:
|
| 312 |
+
file_paths = [
|
| 313 |
+
os.path.join(self.data_dir, f"{self.experiment_id}-{self.num_process}-{process_id}.arrow")
|
| 314 |
+
for process_id in range(self.num_process)
|
| 315 |
+
]
|
| 316 |
+
|
| 317 |
+
# Let's acquire a lock on each process files to be sure they are finished writing
|
| 318 |
+
filelocks = []
|
| 319 |
+
for process_id, file_path in enumerate(file_paths):
|
| 320 |
+
if process_id == 0: # process 0 already has its lock file
|
| 321 |
+
filelocks.append(self.filelock)
|
| 322 |
+
else:
|
| 323 |
+
filelock = FileLock(file_path + ".lock")
|
| 324 |
+
try:
|
| 325 |
+
filelock.acquire(timeout=self.timeout)
|
| 326 |
+
except Timeout:
|
| 327 |
+
raise ValueError(
|
| 328 |
+
f"Cannot acquire lock on cached file {file_path} for process {process_id}."
|
| 329 |
+
) from None
|
| 330 |
+
else:
|
| 331 |
+
filelocks.append(filelock)
|
| 332 |
+
|
| 333 |
+
return file_paths, filelocks
|
| 334 |
+
|
| 335 |
+
def _check_all_processes_locks(self):
|
| 336 |
+
expected_lock_file_names = [
|
| 337 |
+
os.path.join(self.data_dir, f"{self.experiment_id}-{self.num_process}-{process_id}.arrow.lock")
|
| 338 |
+
for process_id in range(self.num_process)
|
| 339 |
+
]
|
| 340 |
+
for expected_lock_file_name in expected_lock_file_names:
|
| 341 |
+
nofilelock = FileFreeLock(expected_lock_file_name)
|
| 342 |
+
try:
|
| 343 |
+
nofilelock.acquire(timeout=self.timeout)
|
| 344 |
+
except Timeout:
|
| 345 |
+
raise ValueError(
|
| 346 |
+
f"Expected to find locked file {expected_lock_file_name} from process {self.process_id} but it doesn't exist."
|
| 347 |
+
) from None
|
| 348 |
+
else:
|
| 349 |
+
nofilelock.release()
|
| 350 |
+
|
| 351 |
+
def _check_rendez_vous(self):
|
| 352 |
+
expected_lock_file_name = os.path.join(self.data_dir, f"{self.experiment_id}-{self.num_process}-0.arrow.lock")
|
| 353 |
+
nofilelock = FileFreeLock(expected_lock_file_name)
|
| 354 |
+
try:
|
| 355 |
+
nofilelock.acquire(timeout=self.timeout)
|
| 356 |
+
except Timeout:
|
| 357 |
+
raise ValueError(
|
| 358 |
+
f"Expected to find locked file {expected_lock_file_name} from process {self.process_id} but it doesn't exist."
|
| 359 |
+
) from None
|
| 360 |
+
else:
|
| 361 |
+
nofilelock.release()
|
| 362 |
+
lock_file_name = os.path.join(self.data_dir, f"{self.experiment_id}-{self.num_process}-rdv.lock")
|
| 363 |
+
rendez_vous_lock = FileLock(lock_file_name)
|
| 364 |
+
try:
|
| 365 |
+
rendez_vous_lock.acquire(timeout=self.timeout)
|
| 366 |
+
except Timeout:
|
| 367 |
+
raise ValueError(f"Couldn't acquire lock on {lock_file_name} from process {self.process_id}.") from None
|
| 368 |
+
else:
|
| 369 |
+
rendez_vous_lock.release()
|
| 370 |
+
|
| 371 |
+
def _finalize(self):
|
| 372 |
+
"""Close all the writing process and load/gather the data
|
| 373 |
+
from all the nodes if main node or all_process is True.
|
| 374 |
+
"""
|
| 375 |
+
if self.writer is not None:
|
| 376 |
+
self.writer.finalize()
|
| 377 |
+
self.writer = None
|
| 378 |
+
# release the locks of the processes > 0 so that process 0 can lock them to read + delete the data
|
| 379 |
+
if self.filelock is not None and self.process_id > 0:
|
| 380 |
+
self.filelock.release()
|
| 381 |
+
|
| 382 |
+
if self.keep_in_memory:
|
| 383 |
+
# Read the predictions and references
|
| 384 |
+
reader = ArrowReader(path=self.data_dir, info=DatasetInfo(features=self.features))
|
| 385 |
+
self.data = Dataset.from_buffer(self.buf_writer.getvalue())
|
| 386 |
+
|
| 387 |
+
elif self.process_id == 0:
|
| 388 |
+
# Let's acquire a lock on each node files to be sure they are finished writing
|
| 389 |
+
file_paths, filelocks = self._get_all_cache_files()
|
| 390 |
+
|
| 391 |
+
# Read the predictions and references
|
| 392 |
+
try:
|
| 393 |
+
reader = ArrowReader(path="", info=DatasetInfo(features=self.features))
|
| 394 |
+
self.data = Dataset(**reader.read_files([{"filename": f} for f in file_paths]))
|
| 395 |
+
except FileNotFoundError:
|
| 396 |
+
raise ValueError(
|
| 397 |
+
"Error in finalize: another metric instance is already using the local cache file. "
|
| 398 |
+
"Please specify an experiment_id to avoid collision between distributed metric instances."
|
| 399 |
+
) from None
|
| 400 |
+
|
| 401 |
+
# Store file paths and locks and we will release/delete them after the computation.
|
| 402 |
+
self.file_paths = file_paths
|
| 403 |
+
self.filelocks = filelocks
|
| 404 |
+
|
| 405 |
+
def compute(self, *, predictions=None, references=None, **kwargs) -> Optional[dict]:
|
| 406 |
+
"""Compute the metrics.
|
| 407 |
+
|
| 408 |
+
Usage of positional arguments is not allowed to prevent mistakes.
|
| 409 |
+
|
| 410 |
+
Args:
|
| 411 |
+
predictions (list/array/tensor, optional): Predictions.
|
| 412 |
+
references (list/array/tensor, optional): References.
|
| 413 |
+
**kwargs (optional): Keyword arguments that will be forwarded to the metrics :meth:`_compute`
|
| 414 |
+
method (see details in the docstring).
|
| 415 |
+
|
| 416 |
+
Return:
|
| 417 |
+
dict or None
|
| 418 |
+
|
| 419 |
+
- Dictionary with the metrics if this metric is run on the main process (``process_id == 0``).
|
| 420 |
+
- None if the metric is not run on the main process (``process_id != 0``).
|
| 421 |
+
|
| 422 |
+
Example:
|
| 423 |
+
|
| 424 |
+
```py
|
| 425 |
+
>>> from datasets import load_metric
|
| 426 |
+
>>> metric = load_metric("accuracy")
|
| 427 |
+
>>> accuracy = metric.compute(predictions=model_prediction, references=labels)
|
| 428 |
+
```
|
| 429 |
+
"""
|
| 430 |
+
all_kwargs = {"predictions": predictions, "references": references, **kwargs}
|
| 431 |
+
if predictions is None and references is None:
|
| 432 |
+
missing_kwargs = {k: None for k in self.features if k not in all_kwargs}
|
| 433 |
+
all_kwargs.update(missing_kwargs)
|
| 434 |
+
else:
|
| 435 |
+
missing_inputs = [k for k in self.features if k not in all_kwargs]
|
| 436 |
+
if missing_inputs:
|
| 437 |
+
raise ValueError(
|
| 438 |
+
f"Metric inputs are missing: {missing_inputs}. All required inputs are {list(self.features)}"
|
| 439 |
+
)
|
| 440 |
+
inputs = {input_name: all_kwargs[input_name] for input_name in self.features}
|
| 441 |
+
compute_kwargs = {k: kwargs[k] for k in kwargs if k not in self.features}
|
| 442 |
+
|
| 443 |
+
if any(v is not None for v in inputs.values()):
|
| 444 |
+
self.add_batch(**inputs)
|
| 445 |
+
self._finalize()
|
| 446 |
+
|
| 447 |
+
self.cache_file_name = None
|
| 448 |
+
self.filelock = None
|
| 449 |
+
|
| 450 |
+
if self.process_id == 0:
|
| 451 |
+
self.data.set_format(type=self.info.format)
|
| 452 |
+
|
| 453 |
+
inputs = {input_name: self.data[input_name] for input_name in self.features}
|
| 454 |
+
with temp_seed(self.seed):
|
| 455 |
+
output = self._compute(**inputs, **compute_kwargs)
|
| 456 |
+
|
| 457 |
+
if self.buf_writer is not None:
|
| 458 |
+
self.buf_writer = None
|
| 459 |
+
del self.data
|
| 460 |
+
self.data = None
|
| 461 |
+
else:
|
| 462 |
+
# Release locks and delete all the cache files. Process 0 is released last.
|
| 463 |
+
for filelock, file_path in reversed(list(zip(self.filelocks, self.file_paths))):
|
| 464 |
+
logger.info(f"Removing {file_path}")
|
| 465 |
+
del self.data
|
| 466 |
+
self.data = None
|
| 467 |
+
del self.writer
|
| 468 |
+
self.writer = None
|
| 469 |
+
os.remove(file_path)
|
| 470 |
+
filelock.release()
|
| 471 |
+
|
| 472 |
+
return output
|
| 473 |
+
else:
|
| 474 |
+
return None
|
| 475 |
+
|
| 476 |
+
def add_batch(self, *, predictions=None, references=None, **kwargs):
|
| 477 |
+
"""Add a batch of predictions and references for the metric's stack.
|
| 478 |
+
|
| 479 |
+
Args:
|
| 480 |
+
predictions (list/array/tensor, optional): Predictions.
|
| 481 |
+
references (list/array/tensor, optional): References.
|
| 482 |
+
|
| 483 |
+
Example:
|
| 484 |
+
|
| 485 |
+
```py
|
| 486 |
+
>>> from datasets import load_metric
|
| 487 |
+
>>> metric = load_metric("accuracy")
|
| 488 |
+
>>> metric.add_batch(predictions=model_prediction, references=labels)
|
| 489 |
+
```
|
| 490 |
+
"""
|
| 491 |
+
bad_inputs = [input_name for input_name in kwargs if input_name not in self.features]
|
| 492 |
+
if bad_inputs:
|
| 493 |
+
raise ValueError(f"Bad inputs for metric: {bad_inputs}. All required inputs are {list(self.features)}")
|
| 494 |
+
batch = {"predictions": predictions, "references": references, **kwargs}
|
| 495 |
+
batch = {intput_name: batch[intput_name] for intput_name in self.features}
|
| 496 |
+
batch = self.info.features.encode_batch(batch)
|
| 497 |
+
if self.writer is None:
|
| 498 |
+
self._init_writer()
|
| 499 |
+
try:
|
| 500 |
+
self.writer.write_batch(batch)
|
| 501 |
+
except pa.ArrowInvalid:
|
| 502 |
+
if any(len(batch[c]) != len(next(iter(batch.values()))) for c in batch):
|
| 503 |
+
col0 = next(iter(batch))
|
| 504 |
+
bad_col = [c for c in batch if len(batch[c]) != len(batch[col0])][0]
|
| 505 |
+
error_msg = (
|
| 506 |
+
f"Mismatch in the number of {col0} ({len(batch[col0])}) and {bad_col} ({len(batch[bad_col])})"
|
| 507 |
+
)
|
| 508 |
+
elif sorted(self.features) != ["references", "predictions"]:
|
| 509 |
+
error_msg = f"Metric inputs don't match the expected format.\n" f"Expected format: {self.features},\n"
|
| 510 |
+
error_msg_inputs = ",\n".join(
|
| 511 |
+
f"Input {input_name}: {summarize_if_long_list(batch[input_name])}" for input_name in self.features
|
| 512 |
+
)
|
| 513 |
+
error_msg += error_msg_inputs
|
| 514 |
+
else:
|
| 515 |
+
error_msg = (
|
| 516 |
+
f"Predictions and/or references don't match the expected format.\n"
|
| 517 |
+
f"Expected format: {self.features},\n"
|
| 518 |
+
f"Input predictions: {summarize_if_long_list(predictions)},\n"
|
| 519 |
+
f"Input references: {summarize_if_long_list(references)}"
|
| 520 |
+
)
|
| 521 |
+
raise ValueError(error_msg) from None
|
| 522 |
+
|
| 523 |
+
def add(self, *, prediction=None, reference=None, **kwargs):
|
| 524 |
+
"""Add one prediction and reference for the metric's stack.
|
| 525 |
+
|
| 526 |
+
Args:
|
| 527 |
+
prediction (list/array/tensor, optional): Predictions.
|
| 528 |
+
reference (list/array/tensor, optional): References.
|
| 529 |
+
|
| 530 |
+
Example:
|
| 531 |
+
|
| 532 |
+
```py
|
| 533 |
+
>>> from datasets import load_metric
|
| 534 |
+
>>> metric = load_metric("accuracy")
|
| 535 |
+
>>> metric.add(predictions=model_predictions, references=labels)
|
| 536 |
+
```
|
| 537 |
+
"""
|
| 538 |
+
bad_inputs = [input_name for input_name in kwargs if input_name not in self.features]
|
| 539 |
+
if bad_inputs:
|
| 540 |
+
raise ValueError(f"Bad inputs for metric: {bad_inputs}. All required inputs are {list(self.features)}")
|
| 541 |
+
example = {"predictions": prediction, "references": reference, **kwargs}
|
| 542 |
+
example = {intput_name: example[intput_name] for intput_name in self.features}
|
| 543 |
+
example = self.info.features.encode_example(example)
|
| 544 |
+
if self.writer is None:
|
| 545 |
+
self._init_writer()
|
| 546 |
+
try:
|
| 547 |
+
self.writer.write(example)
|
| 548 |
+
except pa.ArrowInvalid:
|
| 549 |
+
error_msg = f"Metric inputs don't match the expected format.\n" f"Expected format: {self.features},\n"
|
| 550 |
+
error_msg_inputs = ",\n".join(
|
| 551 |
+
f"Input {input_name}: {summarize_if_long_list(example[input_name])}" for input_name in self.features
|
| 552 |
+
)
|
| 553 |
+
error_msg += error_msg_inputs
|
| 554 |
+
raise ValueError(error_msg) from None
|
| 555 |
+
|
| 556 |
+
def _init_writer(self, timeout=1):
|
| 557 |
+
if self.num_process > 1:
|
| 558 |
+
if self.process_id == 0:
|
| 559 |
+
file_path = os.path.join(self.data_dir, f"{self.experiment_id}-{self.num_process}-rdv.lock")
|
| 560 |
+
self.rendez_vous_lock = FileLock(file_path)
|
| 561 |
+
try:
|
| 562 |
+
self.rendez_vous_lock.acquire(timeout=timeout)
|
| 563 |
+
except TimeoutError:
|
| 564 |
+
raise ValueError(
|
| 565 |
+
f"Error in _init_writer: another metric instance is already using the local cache file at {file_path}. "
|
| 566 |
+
f"Please specify an experiment_id (currently: {self.experiment_id}) to avoid collision "
|
| 567 |
+
f"between distributed metric instances."
|
| 568 |
+
) from None
|
| 569 |
+
|
| 570 |
+
if self.keep_in_memory:
|
| 571 |
+
self.buf_writer = pa.BufferOutputStream()
|
| 572 |
+
self.writer = ArrowWriter(
|
| 573 |
+
features=self.info.features, stream=self.buf_writer, writer_batch_size=self.writer_batch_size
|
| 574 |
+
)
|
| 575 |
+
else:
|
| 576 |
+
self.buf_writer = None
|
| 577 |
+
|
| 578 |
+
# Get cache file name and lock it
|
| 579 |
+
if self.cache_file_name is None or self.filelock is None:
|
| 580 |
+
cache_file_name, filelock = self._create_cache_file() # get ready
|
| 581 |
+
self.cache_file_name = cache_file_name
|
| 582 |
+
self.filelock = filelock
|
| 583 |
+
|
| 584 |
+
self.writer = ArrowWriter(
|
| 585 |
+
features=self.info.features, path=self.cache_file_name, writer_batch_size=self.writer_batch_size
|
| 586 |
+
)
|
| 587 |
+
# Setup rendez-vous here if
|
| 588 |
+
if self.num_process > 1:
|
| 589 |
+
if self.process_id == 0:
|
| 590 |
+
self._check_all_processes_locks() # wait for everyone to be ready
|
| 591 |
+
self.rendez_vous_lock.release() # let everyone go
|
| 592 |
+
else:
|
| 593 |
+
self._check_rendez_vous() # wait for master to be ready and to let everyone go
|
| 594 |
+
|
| 595 |
+
def _info(self) -> MetricInfo:
|
| 596 |
+
"""Construct the MetricInfo object. See `MetricInfo` for details.
|
| 597 |
+
|
| 598 |
+
Warning: This function is only called once and the result is cached for all
|
| 599 |
+
following .info() calls.
|
| 600 |
+
|
| 601 |
+
Returns:
|
| 602 |
+
info: (MetricInfo) The metrics information
|
| 603 |
+
"""
|
| 604 |
+
raise NotImplementedError
|
| 605 |
+
|
| 606 |
+
def download_and_prepare(
|
| 607 |
+
self,
|
| 608 |
+
download_config: Optional[DownloadConfig] = None,
|
| 609 |
+
dl_manager: Optional[DownloadManager] = None,
|
| 610 |
+
):
|
| 611 |
+
"""Downloads and prepares dataset for reading.
|
| 612 |
+
|
| 613 |
+
Args:
|
| 614 |
+
download_config (:class:`DownloadConfig`, optional): Specific download configuration parameters.
|
| 615 |
+
dl_manager (:class:`DownloadManager`, optional): Specific download manager to use.
|
| 616 |
+
"""
|
| 617 |
+
if dl_manager is None:
|
| 618 |
+
if download_config is None:
|
| 619 |
+
download_config = DownloadConfig()
|
| 620 |
+
download_config.cache_dir = os.path.join(self.data_dir, "downloads")
|
| 621 |
+
download_config.force_download = False
|
| 622 |
+
|
| 623 |
+
dl_manager = DownloadManager(
|
| 624 |
+
dataset_name=self.name, download_config=download_config, data_dir=self.data_dir
|
| 625 |
+
)
|
| 626 |
+
|
| 627 |
+
self._download_and_prepare(dl_manager)
|
| 628 |
+
|
| 629 |
+
def _download_and_prepare(self, dl_manager):
|
| 630 |
+
"""Downloads and prepares resources for the metric.
|
| 631 |
+
|
| 632 |
+
This is the internal implementation to overwrite called when user calls
|
| 633 |
+
`download_and_prepare`. It should download all required resources for the metric.
|
| 634 |
+
|
| 635 |
+
Args:
|
| 636 |
+
dl_manager (:class:`DownloadManager`): `DownloadManager` used to download and cache data.
|
| 637 |
+
"""
|
| 638 |
+
return None
|
| 639 |
+
|
| 640 |
+
def _compute(self, *, predictions=None, references=None, **kwargs) -> Dict[str, Any]:
|
| 641 |
+
"""This method defines the common API for all the metrics in the library"""
|
| 642 |
+
raise NotImplementedError
|
| 643 |
+
|
| 644 |
+
def __del__(self):
|
| 645 |
+
if hasattr(self, "filelock") and self.filelock is not None:
|
| 646 |
+
self.filelock.release()
|
| 647 |
+
if hasattr(self, "rendez_vous_lock") and self.rendez_vous_lock is not None:
|
| 648 |
+
self.rendez_vous_lock.release()
|
| 649 |
+
if hasattr(self, "writer"): # in case it was already deleted
|
| 650 |
+
del self.writer
|
| 651 |
+
if hasattr(self, "data"): # in case it was already deleted
|
| 652 |
+
del self.data
|
evalkit_tf437/lib/python3.10/site-packages/datasets/naming.py
ADDED
|
@@ -0,0 +1,84 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors.
|
| 2 |
+
#
|
| 3 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
| 4 |
+
# you may not use this file except in compliance with the License.
|
| 5 |
+
# You may obtain a copy of the License at
|
| 6 |
+
#
|
| 7 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
| 8 |
+
#
|
| 9 |
+
# Unless required by applicable law or agreed to in writing, software
|
| 10 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
| 11 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| 12 |
+
# See the License for the specific language governing permissions and
|
| 13 |
+
# limitations under the License.
|
| 14 |
+
|
| 15 |
+
# Lint as: python3
|
| 16 |
+
"""Utilities for file names."""
|
| 17 |
+
|
| 18 |
+
import itertools
|
| 19 |
+
import os
|
| 20 |
+
import re
|
| 21 |
+
|
| 22 |
+
|
| 23 |
+
_uppercase_uppercase_re = re.compile(r"([A-Z]+)([A-Z][a-z])")
|
| 24 |
+
_lowercase_uppercase_re = re.compile(r"([a-z\d])([A-Z])")
|
| 25 |
+
|
| 26 |
+
_single_underscore_re = re.compile(r"(?<!_)_(?!_)")
|
| 27 |
+
_multiple_underscores_re = re.compile(r"(_{2,})")
|
| 28 |
+
|
| 29 |
+
_split_re = r"^\w+(\.\w+)*$"
|
| 30 |
+
|
| 31 |
+
INVALID_WINDOWS_CHARACTERS_IN_PATH = r"<>:/\|?*"
|
| 32 |
+
|
| 33 |
+
|
| 34 |
+
def camelcase_to_snakecase(name):
|
| 35 |
+
"""Convert camel-case string to snake-case."""
|
| 36 |
+
name = _uppercase_uppercase_re.sub(r"\1_\2", name)
|
| 37 |
+
name = _lowercase_uppercase_re.sub(r"\1_\2", name)
|
| 38 |
+
return name.lower()
|
| 39 |
+
|
| 40 |
+
|
| 41 |
+
def snakecase_to_camelcase(name):
|
| 42 |
+
"""Convert snake-case string to camel-case string."""
|
| 43 |
+
name = _single_underscore_re.split(name)
|
| 44 |
+
name = [_multiple_underscores_re.split(n) for n in name]
|
| 45 |
+
return "".join(n.capitalize() for n in itertools.chain.from_iterable(name) if n != "")
|
| 46 |
+
|
| 47 |
+
|
| 48 |
+
def filename_prefix_for_name(name):
|
| 49 |
+
if os.path.basename(name) != name:
|
| 50 |
+
raise ValueError(f"Should be a dataset name, not a path: {name}")
|
| 51 |
+
return camelcase_to_snakecase(name)
|
| 52 |
+
|
| 53 |
+
|
| 54 |
+
def filename_prefix_for_split(name, split):
|
| 55 |
+
if os.path.basename(name) != name:
|
| 56 |
+
raise ValueError(f"Should be a dataset name, not a path: {name}")
|
| 57 |
+
if not re.match(_split_re, split):
|
| 58 |
+
raise ValueError(f"Split name should match '{_split_re}'' but got '{split}'.")
|
| 59 |
+
return f"{filename_prefix_for_name(name)}-{split}"
|
| 60 |
+
|
| 61 |
+
|
| 62 |
+
def filepattern_for_dataset_split(dataset_name, split, data_dir, filetype_suffix=None):
|
| 63 |
+
prefix = filename_prefix_for_split(dataset_name, split)
|
| 64 |
+
if filetype_suffix:
|
| 65 |
+
prefix += f".{filetype_suffix}"
|
| 66 |
+
filepath = os.path.join(data_dir, prefix)
|
| 67 |
+
return f"{filepath}*"
|
| 68 |
+
|
| 69 |
+
|
| 70 |
+
def filenames_for_dataset_split(path, dataset_name, split, filetype_suffix=None, shard_lengths=None):
|
| 71 |
+
prefix = filename_prefix_for_split(dataset_name, split)
|
| 72 |
+
prefix = os.path.join(path, prefix)
|
| 73 |
+
|
| 74 |
+
if shard_lengths:
|
| 75 |
+
num_shards = len(shard_lengths)
|
| 76 |
+
filenames = [f"{prefix}-{shard_id:05d}-of-{num_shards:05d}" for shard_id in range(num_shards)]
|
| 77 |
+
if filetype_suffix:
|
| 78 |
+
filenames = [filename + f".{filetype_suffix}" for filename in filenames]
|
| 79 |
+
return filenames
|
| 80 |
+
else:
|
| 81 |
+
filename = prefix
|
| 82 |
+
if filetype_suffix:
|
| 83 |
+
filename += f".{filetype_suffix}"
|
| 84 |
+
return [filename]
|
evalkit_tf437/lib/python3.10/site-packages/datasets/search.py
ADDED
|
@@ -0,0 +1,779 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import importlib.util
|
| 2 |
+
import os
|
| 3 |
+
import tempfile
|
| 4 |
+
from pathlib import PurePath
|
| 5 |
+
from typing import TYPE_CHECKING, Dict, List, NamedTuple, Optional, Union
|
| 6 |
+
|
| 7 |
+
import fsspec
|
| 8 |
+
import numpy as np
|
| 9 |
+
|
| 10 |
+
from .utils import logging
|
| 11 |
+
from .utils import tqdm as hf_tqdm
|
| 12 |
+
|
| 13 |
+
|
| 14 |
+
if TYPE_CHECKING:
|
| 15 |
+
from .arrow_dataset import Dataset # noqa: F401
|
| 16 |
+
|
| 17 |
+
try:
|
| 18 |
+
from elasticsearch import Elasticsearch # noqa: F401
|
| 19 |
+
|
| 20 |
+
except ImportError:
|
| 21 |
+
pass
|
| 22 |
+
try:
|
| 23 |
+
import faiss # noqa: F401
|
| 24 |
+
|
| 25 |
+
except ImportError:
|
| 26 |
+
pass
|
| 27 |
+
|
| 28 |
+
_has_elasticsearch = importlib.util.find_spec("elasticsearch") is not None
|
| 29 |
+
_has_faiss = importlib.util.find_spec("faiss") is not None
|
| 30 |
+
|
| 31 |
+
|
| 32 |
+
logger = logging.get_logger(__name__)
|
| 33 |
+
|
| 34 |
+
|
| 35 |
+
class MissingIndex(Exception):
|
| 36 |
+
pass
|
| 37 |
+
|
| 38 |
+
|
| 39 |
+
class SearchResults(NamedTuple):
|
| 40 |
+
scores: List[float]
|
| 41 |
+
indices: List[int]
|
| 42 |
+
|
| 43 |
+
|
| 44 |
+
class BatchedSearchResults(NamedTuple):
|
| 45 |
+
total_scores: List[List[float]]
|
| 46 |
+
total_indices: List[List[int]]
|
| 47 |
+
|
| 48 |
+
|
| 49 |
+
class NearestExamplesResults(NamedTuple):
|
| 50 |
+
scores: List[float]
|
| 51 |
+
examples: dict
|
| 52 |
+
|
| 53 |
+
|
| 54 |
+
class BatchedNearestExamplesResults(NamedTuple):
|
| 55 |
+
total_scores: List[List[float]]
|
| 56 |
+
total_examples: List[dict]
|
| 57 |
+
|
| 58 |
+
|
| 59 |
+
class BaseIndex:
|
| 60 |
+
"""Base class for indexing"""
|
| 61 |
+
|
| 62 |
+
def search(self, query, k: int = 10, **kwargs) -> SearchResults:
|
| 63 |
+
"""
|
| 64 |
+
To implement.
|
| 65 |
+
This method has to return the scores and the indices of the retrieved examples given a certain query.
|
| 66 |
+
"""
|
| 67 |
+
raise NotImplementedError
|
| 68 |
+
|
| 69 |
+
def search_batch(self, queries, k: int = 10, **kwargs) -> BatchedSearchResults:
|
| 70 |
+
"""Find the nearest examples indices to the query.
|
| 71 |
+
|
| 72 |
+
Args:
|
| 73 |
+
queries (`Union[List[str], np.ndarray]`): The queries as a list of strings if `column` is a text index or as a numpy array if `column` is a vector index.
|
| 74 |
+
k (`int`): The number of examples to retrieve per query.
|
| 75 |
+
|
| 76 |
+
Ouput:
|
| 77 |
+
total_scores (`List[List[float]`): The retrieval scores of the retrieved examples per query.
|
| 78 |
+
total_indices (`List[List[int]]`): The indices of the retrieved examples per query.
|
| 79 |
+
"""
|
| 80 |
+
total_scores, total_indices = [], []
|
| 81 |
+
for query in queries:
|
| 82 |
+
scores, indices = self.search(query, k)
|
| 83 |
+
total_scores.append(scores)
|
| 84 |
+
total_indices.append(indices)
|
| 85 |
+
return BatchedSearchResults(total_scores, total_indices)
|
| 86 |
+
|
| 87 |
+
def save(self, file: Union[str, PurePath]):
|
| 88 |
+
"""Serialize the index on disk"""
|
| 89 |
+
raise NotImplementedError
|
| 90 |
+
|
| 91 |
+
@classmethod
|
| 92 |
+
def load(cls, file: Union[str, PurePath]) -> "BaseIndex":
|
| 93 |
+
"""Deserialize the index from disk"""
|
| 94 |
+
raise NotImplementedError
|
| 95 |
+
|
| 96 |
+
|
| 97 |
+
class ElasticSearchIndex(BaseIndex):
|
| 98 |
+
"""
|
| 99 |
+
Sparse index using Elasticsearch. It is used to index text and run queries based on BM25 similarity.
|
| 100 |
+
An Elasticsearch server needs to be accessible, and a python client is declared with
|
| 101 |
+
```
|
| 102 |
+
es_client = Elasticsearch([{'host': 'localhost', 'port': '9200'}])
|
| 103 |
+
```
|
| 104 |
+
for example.
|
| 105 |
+
"""
|
| 106 |
+
|
| 107 |
+
def __init__(
|
| 108 |
+
self,
|
| 109 |
+
host: Optional[str] = None,
|
| 110 |
+
port: Optional[int] = None,
|
| 111 |
+
es_client: Optional["Elasticsearch"] = None,
|
| 112 |
+
es_index_name: Optional[str] = None,
|
| 113 |
+
es_index_config: Optional[dict] = None,
|
| 114 |
+
):
|
| 115 |
+
if not _has_elasticsearch:
|
| 116 |
+
raise ImportError(
|
| 117 |
+
"You must install ElasticSearch to use ElasticSearchIndex. To do so you can run `pip install elasticsearch==7.7.1 for example`"
|
| 118 |
+
)
|
| 119 |
+
if es_client is not None and (host is not None or port is not None):
|
| 120 |
+
raise ValueError("Please specify either `es_client` or `(host, port)`, but not both.")
|
| 121 |
+
host = host or "localhost"
|
| 122 |
+
port = port or 9200
|
| 123 |
+
|
| 124 |
+
import elasticsearch.helpers # noqa: F401 - need this to properly load all the es features
|
| 125 |
+
from elasticsearch import Elasticsearch # noqa: F811
|
| 126 |
+
|
| 127 |
+
self.es_client = es_client if es_client is not None else Elasticsearch([{"host": host, "port": str(port)}])
|
| 128 |
+
self.es_index_name = (
|
| 129 |
+
es_index_name
|
| 130 |
+
if es_index_name is not None
|
| 131 |
+
else "huggingface_datasets_" + os.path.basename(tempfile.NamedTemporaryFile().name)
|
| 132 |
+
)
|
| 133 |
+
self.es_index_config = (
|
| 134 |
+
es_index_config
|
| 135 |
+
if es_index_config is not None
|
| 136 |
+
else {
|
| 137 |
+
"settings": {
|
| 138 |
+
"number_of_shards": 1,
|
| 139 |
+
"analysis": {"analyzer": {"stop_standard": {"type": "standard", " stopwords": "_english_"}}},
|
| 140 |
+
},
|
| 141 |
+
"mappings": {"properties": {"text": {"type": "text", "analyzer": "standard", "similarity": "BM25"}}},
|
| 142 |
+
}
|
| 143 |
+
)
|
| 144 |
+
|
| 145 |
+
def add_documents(self, documents: Union[List[str], "Dataset"], column: Optional[str] = None):
|
| 146 |
+
"""
|
| 147 |
+
Add documents to the index.
|
| 148 |
+
If the documents are inside a certain column, you can specify it using the `column` argument.
|
| 149 |
+
"""
|
| 150 |
+
index_name = self.es_index_name
|
| 151 |
+
index_config = self.es_index_config
|
| 152 |
+
self.es_client.indices.create(index=index_name, body=index_config)
|
| 153 |
+
number_of_docs = len(documents)
|
| 154 |
+
progress = hf_tqdm(unit="docs", total=number_of_docs)
|
| 155 |
+
successes = 0
|
| 156 |
+
|
| 157 |
+
def passage_generator():
|
| 158 |
+
if column is not None:
|
| 159 |
+
for i, example in enumerate(documents):
|
| 160 |
+
yield {"text": example[column], "_id": i}
|
| 161 |
+
else:
|
| 162 |
+
for i, example in enumerate(documents):
|
| 163 |
+
yield {"text": example, "_id": i}
|
| 164 |
+
|
| 165 |
+
# create the ES index
|
| 166 |
+
import elasticsearch as es
|
| 167 |
+
|
| 168 |
+
for ok, action in es.helpers.streaming_bulk(
|
| 169 |
+
client=self.es_client,
|
| 170 |
+
index=index_name,
|
| 171 |
+
actions=passage_generator(),
|
| 172 |
+
):
|
| 173 |
+
progress.update(1)
|
| 174 |
+
successes += ok
|
| 175 |
+
if successes != len(documents):
|
| 176 |
+
logger.warning(
|
| 177 |
+
f"Some documents failed to be added to ElasticSearch. Failures: {len(documents)-successes}/{len(documents)}"
|
| 178 |
+
)
|
| 179 |
+
logger.info(f"Indexed {successes:d} documents")
|
| 180 |
+
|
| 181 |
+
def search(self, query: str, k=10, **kwargs) -> SearchResults:
|
| 182 |
+
"""Find the nearest examples indices to the query.
|
| 183 |
+
|
| 184 |
+
Args:
|
| 185 |
+
query (`str`): The query as a string.
|
| 186 |
+
k (`int`): The number of examples to retrieve.
|
| 187 |
+
|
| 188 |
+
Ouput:
|
| 189 |
+
scores (`List[List[float]`): The retrieval scores of the retrieved examples.
|
| 190 |
+
indices (`List[List[int]]`): The indices of the retrieved examples.
|
| 191 |
+
"""
|
| 192 |
+
response = self.es_client.search(
|
| 193 |
+
index=self.es_index_name,
|
| 194 |
+
body={"query": {"multi_match": {"query": query, "fields": ["text"], "type": "cross_fields"}}, "size": k},
|
| 195 |
+
**kwargs,
|
| 196 |
+
)
|
| 197 |
+
hits = response["hits"]["hits"]
|
| 198 |
+
return SearchResults([hit["_score"] for hit in hits], [int(hit["_id"]) for hit in hits])
|
| 199 |
+
|
| 200 |
+
def search_batch(self, queries, k: int = 10, max_workers=10, **kwargs) -> BatchedSearchResults:
|
| 201 |
+
import concurrent.futures
|
| 202 |
+
|
| 203 |
+
total_scores, total_indices = [None] * len(queries), [None] * len(queries)
|
| 204 |
+
with concurrent.futures.ThreadPoolExecutor(max_workers=max_workers) as executor:
|
| 205 |
+
future_to_index = {executor.submit(self.search, query, k, **kwargs): i for i, query in enumerate(queries)}
|
| 206 |
+
for future in concurrent.futures.as_completed(future_to_index):
|
| 207 |
+
index = future_to_index[future]
|
| 208 |
+
results: SearchResults = future.result()
|
| 209 |
+
total_scores[index] = results.scores
|
| 210 |
+
total_indices[index] = results.indices
|
| 211 |
+
return BatchedSearchResults(total_indices=total_indices, total_scores=total_scores)
|
| 212 |
+
|
| 213 |
+
|
| 214 |
+
class FaissIndex(BaseIndex):
|
| 215 |
+
"""
|
| 216 |
+
Dense index using Faiss. It is used to index vectors.
|
| 217 |
+
Faiss is a library for efficient similarity search and clustering of dense vectors.
|
| 218 |
+
It contains algorithms that search in sets of vectors of any size, up to ones that possibly do not fit in RAM.
|
| 219 |
+
You can find more information about Faiss here:
|
| 220 |
+
- For index types and the string factory: https://github.com/facebookresearch/faiss/wiki/The-index-factory
|
| 221 |
+
- For GPU settings: https://github.com/facebookresearch/faiss/wiki/Faiss-on-the-GPU
|
| 222 |
+
"""
|
| 223 |
+
|
| 224 |
+
def __init__(
|
| 225 |
+
self,
|
| 226 |
+
device: Optional[Union[int, List[int]]] = None,
|
| 227 |
+
string_factory: Optional[str] = None,
|
| 228 |
+
metric_type: Optional[int] = None,
|
| 229 |
+
custom_index: Optional["faiss.Index"] = None,
|
| 230 |
+
):
|
| 231 |
+
"""
|
| 232 |
+
Create a Dense index using Faiss. You can specify `device` if you want to run it on GPU (`device` must be the GPU index).
|
| 233 |
+
You can find more information about Faiss here:
|
| 234 |
+
- For `string factory`: https://github.com/facebookresearch/faiss/wiki/The-index-factory
|
| 235 |
+
"""
|
| 236 |
+
if string_factory is not None and custom_index is not None:
|
| 237 |
+
raise ValueError("Please specify either `string_factory` or `custom_index` but not both.")
|
| 238 |
+
if device is not None and custom_index is not None:
|
| 239 |
+
raise ValueError(
|
| 240 |
+
"Cannot pass both 'custom_index' and 'device'. "
|
| 241 |
+
"Pass 'custom_index' already transferred to the target device instead."
|
| 242 |
+
)
|
| 243 |
+
self.device = device
|
| 244 |
+
self.string_factory = string_factory
|
| 245 |
+
self.metric_type = metric_type
|
| 246 |
+
self.faiss_index = custom_index
|
| 247 |
+
if not _has_faiss:
|
| 248 |
+
raise ImportError(
|
| 249 |
+
"You must install Faiss to use FaissIndex. To do so you can run `conda install -c pytorch faiss-cpu` or `conda install -c pytorch faiss-gpu`. "
|
| 250 |
+
"A community supported package is also available on pypi: `pip install faiss-cpu` or `pip install faiss-gpu`. "
|
| 251 |
+
"Note that pip may not have the latest version of FAISS, and thus, some of the latest features and bug fixes may not be available."
|
| 252 |
+
)
|
| 253 |
+
|
| 254 |
+
def add_vectors(
|
| 255 |
+
self,
|
| 256 |
+
vectors: Union[np.array, "Dataset"],
|
| 257 |
+
column: Optional[str] = None,
|
| 258 |
+
batch_size: int = 1000,
|
| 259 |
+
train_size: Optional[int] = None,
|
| 260 |
+
faiss_verbose: Optional[bool] = None,
|
| 261 |
+
):
|
| 262 |
+
"""
|
| 263 |
+
Add vectors to the index.
|
| 264 |
+
If the arrays are inside a certain column, you can specify it using the `column` argument.
|
| 265 |
+
"""
|
| 266 |
+
import faiss # noqa: F811
|
| 267 |
+
|
| 268 |
+
# Create index
|
| 269 |
+
if self.faiss_index is None:
|
| 270 |
+
size = len(vectors[0]) if column is None else len(vectors[0][column])
|
| 271 |
+
if self.string_factory is not None:
|
| 272 |
+
if self.metric_type is None:
|
| 273 |
+
index = faiss.index_factory(size, self.string_factory)
|
| 274 |
+
else:
|
| 275 |
+
index = faiss.index_factory(size, self.string_factory, self.metric_type)
|
| 276 |
+
else:
|
| 277 |
+
if self.metric_type is None:
|
| 278 |
+
index = faiss.IndexFlat(size)
|
| 279 |
+
else:
|
| 280 |
+
index = faiss.IndexFlat(size, self.metric_type)
|
| 281 |
+
|
| 282 |
+
self.faiss_index = self._faiss_index_to_device(index, self.device)
|
| 283 |
+
logger.info(f"Created faiss index of type {type(self.faiss_index)}")
|
| 284 |
+
|
| 285 |
+
# Set verbosity level
|
| 286 |
+
if faiss_verbose is not None:
|
| 287 |
+
self.faiss_index.verbose = faiss_verbose
|
| 288 |
+
if hasattr(self.faiss_index, "index") and self.faiss_index.index is not None:
|
| 289 |
+
self.faiss_index.index.verbose = faiss_verbose
|
| 290 |
+
if hasattr(self.faiss_index, "quantizer") and self.faiss_index.quantizer is not None:
|
| 291 |
+
self.faiss_index.quantizer.verbose = faiss_verbose
|
| 292 |
+
if hasattr(self.faiss_index, "clustering_index") and self.faiss_index.clustering_index is not None:
|
| 293 |
+
self.faiss_index.clustering_index.verbose = faiss_verbose
|
| 294 |
+
|
| 295 |
+
# Train
|
| 296 |
+
if train_size is not None:
|
| 297 |
+
train_vecs = vectors[:train_size] if column is None else vectors[:train_size][column]
|
| 298 |
+
logger.info(f"Training the index with the first {len(train_vecs)} vectors")
|
| 299 |
+
self.faiss_index.train(train_vecs)
|
| 300 |
+
else:
|
| 301 |
+
logger.info("Ignored the training step of the faiss index as `train_size` is None.")
|
| 302 |
+
|
| 303 |
+
# Add vectors
|
| 304 |
+
logger.info(f"Adding {len(vectors)} vectors to the faiss index")
|
| 305 |
+
for i in hf_tqdm(range(0, len(vectors), batch_size)):
|
| 306 |
+
vecs = vectors[i : i + batch_size] if column is None else vectors[i : i + batch_size][column]
|
| 307 |
+
self.faiss_index.add(vecs)
|
| 308 |
+
|
| 309 |
+
@staticmethod
|
| 310 |
+
def _faiss_index_to_device(index: "faiss.Index", device: Optional[Union[int, List[int]]] = None) -> "faiss.Index":
|
| 311 |
+
"""
|
| 312 |
+
Sends a faiss index to a device.
|
| 313 |
+
A device can either be a positive integer (GPU id), a negative integer (all GPUs),
|
| 314 |
+
or a list of positive integers (select GPUs to use), or `None` for CPU.
|
| 315 |
+
"""
|
| 316 |
+
|
| 317 |
+
# If device is not specified, then it runs on CPU.
|
| 318 |
+
if device is None:
|
| 319 |
+
return index
|
| 320 |
+
|
| 321 |
+
import faiss # noqa: F811
|
| 322 |
+
|
| 323 |
+
# If the device id is given as an integer
|
| 324 |
+
if isinstance(device, int):
|
| 325 |
+
# Positive integers are directly mapped to GPU ids
|
| 326 |
+
if device > -1:
|
| 327 |
+
faiss_res = faiss.StandardGpuResources()
|
| 328 |
+
index = faiss.index_cpu_to_gpu(faiss_res, device, index)
|
| 329 |
+
# And negative integers mean using all GPUs
|
| 330 |
+
else:
|
| 331 |
+
index = faiss.index_cpu_to_all_gpus(index)
|
| 332 |
+
# Device ids given as a list mean mapping to those devices specified.
|
| 333 |
+
elif isinstance(device, (list, tuple)):
|
| 334 |
+
index = faiss.index_cpu_to_gpus_list(index, gpus=list(device))
|
| 335 |
+
else:
|
| 336 |
+
raise TypeError(
|
| 337 |
+
f"The argument type: {type(device)} is not expected. "
|
| 338 |
+
+ "Please pass in either nothing, a positive int, a negative int, or a list of positive ints."
|
| 339 |
+
)
|
| 340 |
+
|
| 341 |
+
return index
|
| 342 |
+
|
| 343 |
+
def search(self, query: np.array, k=10, **kwargs) -> SearchResults:
|
| 344 |
+
"""Find the nearest examples indices to the query.
|
| 345 |
+
|
| 346 |
+
Args:
|
| 347 |
+
query (`np.array`): The query as a numpy array.
|
| 348 |
+
k (`int`): The number of examples to retrieve.
|
| 349 |
+
|
| 350 |
+
Ouput:
|
| 351 |
+
scores (`List[List[float]`): The retrieval scores of the retrieved examples.
|
| 352 |
+
indices (`List[List[int]]`): The indices of the retrieved examples.
|
| 353 |
+
"""
|
| 354 |
+
if len(query.shape) != 1 and (len(query.shape) != 2 or query.shape[0] != 1):
|
| 355 |
+
raise ValueError("Shape of query is incorrect, it has to be either a 1D array or 2D (1, N)")
|
| 356 |
+
|
| 357 |
+
queries = query.reshape(1, -1)
|
| 358 |
+
if not queries.flags.c_contiguous:
|
| 359 |
+
queries = np.asarray(queries, order="C")
|
| 360 |
+
scores, indices = self.faiss_index.search(queries, k, **kwargs)
|
| 361 |
+
return SearchResults(scores[0], indices[0].astype(int))
|
| 362 |
+
|
| 363 |
+
def search_batch(self, queries: np.array, k=10, **kwargs) -> BatchedSearchResults:
|
| 364 |
+
"""Find the nearest examples indices to the queries.
|
| 365 |
+
|
| 366 |
+
Args:
|
| 367 |
+
queries (`np.array`): The queries as a numpy array.
|
| 368 |
+
k (`int`): The number of examples to retrieve.
|
| 369 |
+
|
| 370 |
+
Ouput:
|
| 371 |
+
total_scores (`List[List[float]`): The retrieval scores of the retrieved examples per query.
|
| 372 |
+
total_indices (`List[List[int]]`): The indices of the retrieved examples per query.
|
| 373 |
+
"""
|
| 374 |
+
if len(queries.shape) != 2:
|
| 375 |
+
raise ValueError("Shape of query must be 2D")
|
| 376 |
+
if not queries.flags.c_contiguous:
|
| 377 |
+
queries = np.asarray(queries, order="C")
|
| 378 |
+
scores, indices = self.faiss_index.search(queries, k, **kwargs)
|
| 379 |
+
return BatchedSearchResults(scores, indices.astype(int))
|
| 380 |
+
|
| 381 |
+
def save(self, file: Union[str, PurePath], storage_options: Optional[Dict] = None):
|
| 382 |
+
"""Serialize the FaissIndex on disk"""
|
| 383 |
+
import faiss # noqa: F811
|
| 384 |
+
|
| 385 |
+
if self.device is not None and isinstance(self.device, (int, list, tuple)):
|
| 386 |
+
index = faiss.index_gpu_to_cpu(self.faiss_index)
|
| 387 |
+
else:
|
| 388 |
+
index = self.faiss_index
|
| 389 |
+
|
| 390 |
+
with fsspec.open(str(file), "wb", **(storage_options or {})) as f:
|
| 391 |
+
faiss.write_index(index, faiss.BufferedIOWriter(faiss.PyCallbackIOWriter(f.write)))
|
| 392 |
+
|
| 393 |
+
@classmethod
|
| 394 |
+
def load(
|
| 395 |
+
cls,
|
| 396 |
+
file: Union[str, PurePath],
|
| 397 |
+
device: Optional[Union[int, List[int]]] = None,
|
| 398 |
+
storage_options: Optional[Dict] = None,
|
| 399 |
+
) -> "FaissIndex":
|
| 400 |
+
"""Deserialize the FaissIndex from disk"""
|
| 401 |
+
import faiss # noqa: F811
|
| 402 |
+
|
| 403 |
+
# Instances of FaissIndex is essentially just a wrapper for faiss indices.
|
| 404 |
+
faiss_index = cls(device=device)
|
| 405 |
+
with fsspec.open(str(file), "rb", **(storage_options or {})) as f:
|
| 406 |
+
index = faiss.read_index(faiss.BufferedIOReader(faiss.PyCallbackIOReader(f.read)))
|
| 407 |
+
faiss_index.faiss_index = faiss_index._faiss_index_to_device(index, faiss_index.device)
|
| 408 |
+
return faiss_index
|
| 409 |
+
|
| 410 |
+
|
| 411 |
+
class IndexableMixin:
|
| 412 |
+
"""Add indexing features to `datasets.Dataset`"""
|
| 413 |
+
|
| 414 |
+
def __init__(self):
|
| 415 |
+
self._indexes: Dict[str, BaseIndex] = {}
|
| 416 |
+
|
| 417 |
+
def __len__(self):
|
| 418 |
+
raise NotImplementedError
|
| 419 |
+
|
| 420 |
+
def __getitem__(self, key):
|
| 421 |
+
raise NotImplementedError
|
| 422 |
+
|
| 423 |
+
def is_index_initialized(self, index_name: str) -> bool:
|
| 424 |
+
return index_name in self._indexes
|
| 425 |
+
|
| 426 |
+
def _check_index_is_initialized(self, index_name: str):
|
| 427 |
+
if not self.is_index_initialized(index_name):
|
| 428 |
+
raise MissingIndex(
|
| 429 |
+
f"Index with index_name '{index_name}' not initialized yet. Please make sure that you call `add_faiss_index` or `add_elasticsearch_index` first."
|
| 430 |
+
)
|
| 431 |
+
|
| 432 |
+
def list_indexes(self) -> List[str]:
|
| 433 |
+
"""List the `colindex_nameumns`/identifiers of all the attached indexes."""
|
| 434 |
+
return list(self._indexes)
|
| 435 |
+
|
| 436 |
+
def get_index(self, index_name: str) -> BaseIndex:
|
| 437 |
+
"""List the `index_name`/identifiers of all the attached indexes.
|
| 438 |
+
|
| 439 |
+
Args:
|
| 440 |
+
index_name (`str`): Index name.
|
| 441 |
+
|
| 442 |
+
Returns:
|
| 443 |
+
[`BaseIndex`]
|
| 444 |
+
"""
|
| 445 |
+
self._check_index_is_initialized(index_name)
|
| 446 |
+
return self._indexes[index_name]
|
| 447 |
+
|
| 448 |
+
def add_faiss_index(
|
| 449 |
+
self,
|
| 450 |
+
column: str,
|
| 451 |
+
index_name: Optional[str] = None,
|
| 452 |
+
device: Optional[Union[int, List[int]]] = None,
|
| 453 |
+
string_factory: Optional[str] = None,
|
| 454 |
+
metric_type: Optional[int] = None,
|
| 455 |
+
custom_index: Optional["faiss.Index"] = None,
|
| 456 |
+
batch_size: int = 1000,
|
| 457 |
+
train_size: Optional[int] = None,
|
| 458 |
+
faiss_verbose: bool = False,
|
| 459 |
+
):
|
| 460 |
+
"""Add a dense index using Faiss for fast retrieval.
|
| 461 |
+
The index is created using the vectors of the specified column.
|
| 462 |
+
You can specify `device` if you want to run it on GPU (`device` must be the GPU index, see more below).
|
| 463 |
+
You can find more information about Faiss here:
|
| 464 |
+
- For `string factory`: https://github.com/facebookresearch/faiss/wiki/The-index-factory
|
| 465 |
+
|
| 466 |
+
Args:
|
| 467 |
+
column (`str`): The column of the vectors to add to the index.
|
| 468 |
+
index_name (Optional `str`): The index_name/identifier of the index. This is the index_name that is used to call `.get_nearest` or `.search`.
|
| 469 |
+
By default it corresponds to `column`.
|
| 470 |
+
device (Optional `Union[int, List[int]]`): If positive integer, this is the index of the GPU to use. If negative integer, use all GPUs.
|
| 471 |
+
If a list of positive integers is passed in, run only on those GPUs. By default it uses the CPU.
|
| 472 |
+
string_factory (Optional `str`): This is passed to the index factory of Faiss to create the index. Default index class is IndexFlatIP.
|
| 473 |
+
metric_type (Optional `int`): Type of metric. Ex: `faiss.METRIC_INNER_PRODUCT` or `faiss.METRIC_L2`.
|
| 474 |
+
custom_index (Optional `faiss.Index`): Custom Faiss index that you already have instantiated and configured for your needs.
|
| 475 |
+
batch_size (Optional `int`): Size of the batch to use while adding vectors to the FaissIndex. Default value is 1000.
|
| 476 |
+
<Added version="2.4.0"/>
|
| 477 |
+
train_size (Optional `int`): If the index needs a training step, specifies how many vectors will be used to train the index.
|
| 478 |
+
faiss_verbose (`bool`, defaults to False): Enable the verbosity of the Faiss index.
|
| 479 |
+
"""
|
| 480 |
+
index_name = index_name if index_name is not None else column
|
| 481 |
+
faiss_index = FaissIndex(
|
| 482 |
+
device=device, string_factory=string_factory, metric_type=metric_type, custom_index=custom_index
|
| 483 |
+
)
|
| 484 |
+
faiss_index.add_vectors(
|
| 485 |
+
self, column=column, batch_size=batch_size, train_size=train_size, faiss_verbose=faiss_verbose
|
| 486 |
+
)
|
| 487 |
+
self._indexes[index_name] = faiss_index
|
| 488 |
+
|
| 489 |
+
def add_faiss_index_from_external_arrays(
|
| 490 |
+
self,
|
| 491 |
+
external_arrays: np.array,
|
| 492 |
+
index_name: str,
|
| 493 |
+
device: Optional[Union[int, List[int]]] = None,
|
| 494 |
+
string_factory: Optional[str] = None,
|
| 495 |
+
metric_type: Optional[int] = None,
|
| 496 |
+
custom_index: Optional["faiss.Index"] = None,
|
| 497 |
+
batch_size: int = 1000,
|
| 498 |
+
train_size: Optional[int] = None,
|
| 499 |
+
faiss_verbose: bool = False,
|
| 500 |
+
):
|
| 501 |
+
"""Add a dense index using Faiss for fast retrieval.
|
| 502 |
+
The index is created using the vectors of `external_arrays`.
|
| 503 |
+
You can specify `device` if you want to run it on GPU (`device` must be the GPU index).
|
| 504 |
+
You can find more information about Faiss here:
|
| 505 |
+
- For `string factory`: https://github.com/facebookresearch/faiss/wiki/The-index-factory
|
| 506 |
+
|
| 507 |
+
Args:
|
| 508 |
+
external_arrays (`np.array`): If you want to use arrays from outside the lib for the index, you can set `external_arrays`.
|
| 509 |
+
It will use `external_arrays` to create the Faiss index instead of the arrays in the given `column`.
|
| 510 |
+
index_name (`str`): The index_name/identifier of the index. This is the index_name that is used to call `.get_nearest` or `.search`.
|
| 511 |
+
device (Optional `Union[int, List[int]]`): If positive integer, this is the index of the GPU to use. If negative integer, use all GPUs.
|
| 512 |
+
If a list of positive integers is passed in, run only on those GPUs. By default it uses the CPU.
|
| 513 |
+
string_factory (Optional `str`): This is passed to the index factory of Faiss to create the index. Default index class is IndexFlatIP.
|
| 514 |
+
metric_type (Optional `int`): Type of metric. Ex: `faiss.METRIC_INNER_PRODUCT` or `faiss.METRIC_L2`.
|
| 515 |
+
custom_index (Optional `faiss.Index`): Custom Faiss index that you already have instantiated and configured for your needs.
|
| 516 |
+
batch_size (Optional `int`): Size of the batch to use while adding vectors to the FaissIndex. Default value is 1000.
|
| 517 |
+
<Added version="2.4.0"/>
|
| 518 |
+
train_size (Optional `int`): If the index needs a training step, specifies how many vectors will be used to train the index.
|
| 519 |
+
faiss_verbose (`bool`, defaults to False): Enable the verbosity of the Faiss index.
|
| 520 |
+
"""
|
| 521 |
+
faiss_index = FaissIndex(
|
| 522 |
+
device=device, string_factory=string_factory, metric_type=metric_type, custom_index=custom_index
|
| 523 |
+
)
|
| 524 |
+
faiss_index.add_vectors(
|
| 525 |
+
external_arrays, column=None, batch_size=batch_size, train_size=train_size, faiss_verbose=faiss_verbose
|
| 526 |
+
)
|
| 527 |
+
self._indexes[index_name] = faiss_index
|
| 528 |
+
|
| 529 |
+
def save_faiss_index(self, index_name: str, file: Union[str, PurePath], storage_options: Optional[Dict] = None):
|
| 530 |
+
"""Save a FaissIndex on disk.
|
| 531 |
+
|
| 532 |
+
Args:
|
| 533 |
+
index_name (`str`): The index_name/identifier of the index. This is the index_name that is used to call `.get_nearest` or `.search`.
|
| 534 |
+
file (`str`): The path to the serialized faiss index on disk or remote URI (e.g. `"s3://my-bucket/index.faiss"`).
|
| 535 |
+
storage_options (`dict`, *optional*):
|
| 536 |
+
Key/value pairs to be passed on to the file-system backend, if any.
|
| 537 |
+
|
| 538 |
+
<Added version="2.11.0"/>
|
| 539 |
+
|
| 540 |
+
"""
|
| 541 |
+
index = self.get_index(index_name)
|
| 542 |
+
if not isinstance(index, FaissIndex):
|
| 543 |
+
raise ValueError(f"Index '{index_name}' is not a FaissIndex but a '{type(index)}'")
|
| 544 |
+
index.save(file, storage_options=storage_options)
|
| 545 |
+
logger.info(f"Saved FaissIndex {index_name} at {file}")
|
| 546 |
+
|
| 547 |
+
def load_faiss_index(
|
| 548 |
+
self,
|
| 549 |
+
index_name: str,
|
| 550 |
+
file: Union[str, PurePath],
|
| 551 |
+
device: Optional[Union[int, List[int]]] = None,
|
| 552 |
+
storage_options: Optional[Dict] = None,
|
| 553 |
+
):
|
| 554 |
+
"""Load a FaissIndex from disk.
|
| 555 |
+
|
| 556 |
+
If you want to do additional configurations, you can have access to the faiss index object by doing
|
| 557 |
+
`.get_index(index_name).faiss_index` to make it fit your needs.
|
| 558 |
+
|
| 559 |
+
Args:
|
| 560 |
+
index_name (`str`): The index_name/identifier of the index. This is the index_name that is used to
|
| 561 |
+
call `.get_nearest` or `.search`.
|
| 562 |
+
file (`str`): The path to the serialized faiss index on disk or remote URI (e.g. `"s3://my-bucket/index.faiss"`).
|
| 563 |
+
device (Optional `Union[int, List[int]]`): If positive integer, this is the index of the GPU to use. If negative integer, use all GPUs.
|
| 564 |
+
If a list of positive integers is passed in, run only on those GPUs. By default it uses the CPU.
|
| 565 |
+
storage_options (`dict`, *optional*):
|
| 566 |
+
Key/value pairs to be passed on to the file-system backend, if any.
|
| 567 |
+
|
| 568 |
+
<Added version="2.11.0"/>
|
| 569 |
+
|
| 570 |
+
"""
|
| 571 |
+
index = FaissIndex.load(file, device=device, storage_options=storage_options)
|
| 572 |
+
if index.faiss_index.ntotal != len(self):
|
| 573 |
+
raise ValueError(
|
| 574 |
+
f"Index size should match Dataset size, but Index '{index_name}' at {file} has {index.faiss_index.ntotal} elements while the dataset has {len(self)} examples."
|
| 575 |
+
)
|
| 576 |
+
self._indexes[index_name] = index
|
| 577 |
+
logger.info(f"Loaded FaissIndex {index_name} from {file}")
|
| 578 |
+
|
| 579 |
+
def add_elasticsearch_index(
|
| 580 |
+
self,
|
| 581 |
+
column: str,
|
| 582 |
+
index_name: Optional[str] = None,
|
| 583 |
+
host: Optional[str] = None,
|
| 584 |
+
port: Optional[int] = None,
|
| 585 |
+
es_client: Optional["Elasticsearch"] = None,
|
| 586 |
+
es_index_name: Optional[str] = None,
|
| 587 |
+
es_index_config: Optional[dict] = None,
|
| 588 |
+
):
|
| 589 |
+
"""Add a text index using ElasticSearch for fast retrieval.
|
| 590 |
+
|
| 591 |
+
Args:
|
| 592 |
+
column (`str`): The column of the documents to add to the index.
|
| 593 |
+
index_name (Optional `str`): The index_name/identifier of the index. This is the index name that is used to call `.get_nearest` or `.search`.
|
| 594 |
+
By default it corresponds to `column`.
|
| 595 |
+
host (Optional `str`, defaults to localhost):
|
| 596 |
+
host of where ElasticSearch is running
|
| 597 |
+
port (Optional `str`, defaults to 9200):
|
| 598 |
+
port of where ElasticSearch is running
|
| 599 |
+
es_client (Optional `elasticsearch.Elasticsearch`):
|
| 600 |
+
The elasticsearch client used to create the index if host and port are None.
|
| 601 |
+
es_index_name (Optional `str`): The elasticsearch index name used to create the index.
|
| 602 |
+
es_index_config (Optional `dict`):
|
| 603 |
+
The configuration of the elasticsearch index.
|
| 604 |
+
Default config is:
|
| 605 |
+
|
| 606 |
+
Config::
|
| 607 |
+
|
| 608 |
+
{
|
| 609 |
+
"settings": {
|
| 610 |
+
"number_of_shards": 1,
|
| 611 |
+
"analysis": {"analyzer": {"stop_standard": {"type": "standard", " stopwords": "_english_"}}},
|
| 612 |
+
},
|
| 613 |
+
"mappings": {
|
| 614 |
+
"properties": {
|
| 615 |
+
"text": {
|
| 616 |
+
"type": "text",
|
| 617 |
+
"analyzer": "standard",
|
| 618 |
+
"similarity": "BM25"
|
| 619 |
+
},
|
| 620 |
+
}
|
| 621 |
+
},
|
| 622 |
+
}
|
| 623 |
+
"""
|
| 624 |
+
index_name = index_name if index_name is not None else column
|
| 625 |
+
es_index = ElasticSearchIndex(
|
| 626 |
+
host=host, port=port, es_client=es_client, es_index_name=es_index_name, es_index_config=es_index_config
|
| 627 |
+
)
|
| 628 |
+
es_index.add_documents(self, column=column)
|
| 629 |
+
self._indexes[index_name] = es_index
|
| 630 |
+
|
| 631 |
+
def load_elasticsearch_index(
|
| 632 |
+
self,
|
| 633 |
+
index_name: str,
|
| 634 |
+
es_index_name: str,
|
| 635 |
+
host: Optional[str] = None,
|
| 636 |
+
port: Optional[int] = None,
|
| 637 |
+
es_client: Optional["Elasticsearch"] = None,
|
| 638 |
+
es_index_config: Optional[dict] = None,
|
| 639 |
+
):
|
| 640 |
+
"""Load an existing text index using ElasticSearch for fast retrieval.
|
| 641 |
+
|
| 642 |
+
Args:
|
| 643 |
+
index_name (`str`):
|
| 644 |
+
The `index_name`/identifier of the index. This is the index name that is used to call `get_nearest` or `search`.
|
| 645 |
+
es_index_name (`str`):
|
| 646 |
+
The name of elasticsearch index to load.
|
| 647 |
+
host (`str`, *optional*, defaults to `localhost`):
|
| 648 |
+
Host of where ElasticSearch is running.
|
| 649 |
+
port (`str`, *optional*, defaults to `9200`):
|
| 650 |
+
Port of where ElasticSearch is running.
|
| 651 |
+
es_client (`elasticsearch.Elasticsearch`, *optional*):
|
| 652 |
+
The elasticsearch client used to create the index if host and port are `None`.
|
| 653 |
+
es_index_config (`dict`, *optional*):
|
| 654 |
+
The configuration of the elasticsearch index.
|
| 655 |
+
Default config is:
|
| 656 |
+
```
|
| 657 |
+
{
|
| 658 |
+
"settings": {
|
| 659 |
+
"number_of_shards": 1,
|
| 660 |
+
"analysis": {"analyzer": {"stop_standard": {"type": "standard", " stopwords": "_english_"}}},
|
| 661 |
+
},
|
| 662 |
+
"mappings": {
|
| 663 |
+
"properties": {
|
| 664 |
+
"text": {
|
| 665 |
+
"type": "text",
|
| 666 |
+
"analyzer": "standard",
|
| 667 |
+
"similarity": "BM25"
|
| 668 |
+
},
|
| 669 |
+
}
|
| 670 |
+
},
|
| 671 |
+
}
|
| 672 |
+
```
|
| 673 |
+
"""
|
| 674 |
+
self._indexes[index_name] = ElasticSearchIndex(
|
| 675 |
+
host=host, port=port, es_client=es_client, es_index_name=es_index_name, es_index_config=es_index_config
|
| 676 |
+
)
|
| 677 |
+
|
| 678 |
+
def drop_index(self, index_name: str):
|
| 679 |
+
"""Drop the index with the specified column.
|
| 680 |
+
|
| 681 |
+
Args:
|
| 682 |
+
index_name (`str`):
|
| 683 |
+
The `index_name`/identifier of the index.
|
| 684 |
+
"""
|
| 685 |
+
del self._indexes[index_name]
|
| 686 |
+
|
| 687 |
+
def search(self, index_name: str, query: Union[str, np.array], k: int = 10, **kwargs) -> SearchResults:
|
| 688 |
+
"""Find the nearest examples indices in the dataset to the query.
|
| 689 |
+
|
| 690 |
+
Args:
|
| 691 |
+
index_name (`str`):
|
| 692 |
+
The name/identifier of the index.
|
| 693 |
+
query (`Union[str, np.ndarray]`):
|
| 694 |
+
The query as a string if `index_name` is a text index or as a numpy array if `index_name` is a vector index.
|
| 695 |
+
k (`int`):
|
| 696 |
+
The number of examples to retrieve.
|
| 697 |
+
|
| 698 |
+
Returns:
|
| 699 |
+
`(scores, indices)`:
|
| 700 |
+
A tuple of `(scores, indices)` where:
|
| 701 |
+
- **scores** (`List[List[float]`): the retrieval scores from either FAISS (`IndexFlatL2` by default) or ElasticSearch of the retrieved examples
|
| 702 |
+
- **indices** (`List[List[int]]`): the indices of the retrieved examples
|
| 703 |
+
"""
|
| 704 |
+
self._check_index_is_initialized(index_name)
|
| 705 |
+
return self._indexes[index_name].search(query, k, **kwargs)
|
| 706 |
+
|
| 707 |
+
def search_batch(
|
| 708 |
+
self, index_name: str, queries: Union[List[str], np.array], k: int = 10, **kwargs
|
| 709 |
+
) -> BatchedSearchResults:
|
| 710 |
+
"""Find the nearest examples indices in the dataset to the query.
|
| 711 |
+
|
| 712 |
+
Args:
|
| 713 |
+
index_name (`str`):
|
| 714 |
+
The `index_name`/identifier of the index.
|
| 715 |
+
queries (`Union[List[str], np.ndarray]`):
|
| 716 |
+
The queries as a list of strings if `index_name` is a text index or as a numpy array if `index_name` is a vector index.
|
| 717 |
+
k (`int`):
|
| 718 |
+
The number of examples to retrieve per query.
|
| 719 |
+
|
| 720 |
+
Returns:
|
| 721 |
+
`(total_scores, total_indices)`:
|
| 722 |
+
A tuple of `(total_scores, total_indices)` where:
|
| 723 |
+
- **total_scores** (`List[List[float]`): the retrieval scores from either FAISS (`IndexFlatL2` by default) or ElasticSearch of the retrieved examples per query
|
| 724 |
+
- **total_indices** (`List[List[int]]`): the indices of the retrieved examples per query
|
| 725 |
+
"""
|
| 726 |
+
self._check_index_is_initialized(index_name)
|
| 727 |
+
return self._indexes[index_name].search_batch(queries, k, **kwargs)
|
| 728 |
+
|
| 729 |
+
def get_nearest_examples(
|
| 730 |
+
self, index_name: str, query: Union[str, np.array], k: int = 10, **kwargs
|
| 731 |
+
) -> NearestExamplesResults:
|
| 732 |
+
"""Find the nearest examples in the dataset to the query.
|
| 733 |
+
|
| 734 |
+
Args:
|
| 735 |
+
index_name (`str`):
|
| 736 |
+
The index_name/identifier of the index.
|
| 737 |
+
query (`Union[str, np.ndarray]`):
|
| 738 |
+
The query as a string if `index_name` is a text index or as a numpy array if `index_name` is a vector index.
|
| 739 |
+
k (`int`):
|
| 740 |
+
The number of examples to retrieve.
|
| 741 |
+
|
| 742 |
+
Returns:
|
| 743 |
+
`(scores, examples)`:
|
| 744 |
+
A tuple of `(scores, examples)` where:
|
| 745 |
+
- **scores** (`List[float]`): the retrieval scores from either FAISS (`IndexFlatL2` by default) or ElasticSearch of the retrieved examples
|
| 746 |
+
- **examples** (`dict`): the retrieved examples
|
| 747 |
+
"""
|
| 748 |
+
self._check_index_is_initialized(index_name)
|
| 749 |
+
scores, indices = self.search(index_name, query, k, **kwargs)
|
| 750 |
+
top_indices = [i for i in indices if i >= 0]
|
| 751 |
+
return NearestExamplesResults(scores[: len(top_indices)], self[top_indices])
|
| 752 |
+
|
| 753 |
+
def get_nearest_examples_batch(
|
| 754 |
+
self, index_name: str, queries: Union[List[str], np.array], k: int = 10, **kwargs
|
| 755 |
+
) -> BatchedNearestExamplesResults:
|
| 756 |
+
"""Find the nearest examples in the dataset to the query.
|
| 757 |
+
|
| 758 |
+
Args:
|
| 759 |
+
index_name (`str`):
|
| 760 |
+
The `index_name`/identifier of the index.
|
| 761 |
+
queries (`Union[List[str], np.ndarray]`):
|
| 762 |
+
The queries as a list of strings if `index_name` is a text index or as a numpy array if `index_name` is a vector index.
|
| 763 |
+
k (`int`):
|
| 764 |
+
The number of examples to retrieve per query.
|
| 765 |
+
|
| 766 |
+
Returns:
|
| 767 |
+
`(total_scores, total_examples)`:
|
| 768 |
+
A tuple of `(total_scores, total_examples)` where:
|
| 769 |
+
- **total_scores** (`List[List[float]`): the retrieval scores from either FAISS (`IndexFlatL2` by default) or ElasticSearch of the retrieved examples per query
|
| 770 |
+
- **total_examples** (`List[dict]`): the retrieved examples per query
|
| 771 |
+
"""
|
| 772 |
+
self._check_index_is_initialized(index_name)
|
| 773 |
+
total_scores, total_indices = self.search_batch(index_name, queries, k, **kwargs)
|
| 774 |
+
total_scores = [
|
| 775 |
+
scores_i[: len([i for i in indices_i if i >= 0])]
|
| 776 |
+
for scores_i, indices_i in zip(total_scores, total_indices)
|
| 777 |
+
]
|
| 778 |
+
total_samples = [self[[i for i in indices if i >= 0]] for indices in total_indices]
|
| 779 |
+
return BatchedNearestExamplesResults(total_scores, total_samples)
|
evalkit_tf437/lib/python3.10/site-packages/datasets/splits.py
ADDED
|
@@ -0,0 +1,635 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors.
|
| 2 |
+
#
|
| 3 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
| 4 |
+
# you may not use this file except in compliance with the License.
|
| 5 |
+
# You may obtain a copy of the License at
|
| 6 |
+
#
|
| 7 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
| 8 |
+
#
|
| 9 |
+
# Unless required by applicable law or agreed to in writing, software
|
| 10 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
| 11 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| 12 |
+
# See the License for the specific language governing permissions and
|
| 13 |
+
# limitations under the License.
|
| 14 |
+
|
| 15 |
+
# Lint as: python3
|
| 16 |
+
"""Splits related API."""
|
| 17 |
+
|
| 18 |
+
import abc
|
| 19 |
+
import collections
|
| 20 |
+
import copy
|
| 21 |
+
import dataclasses
|
| 22 |
+
import re
|
| 23 |
+
from dataclasses import dataclass
|
| 24 |
+
from typing import Dict, List, Optional, Union
|
| 25 |
+
|
| 26 |
+
from .arrow_reader import FileInstructions, make_file_instructions
|
| 27 |
+
from .naming import _split_re
|
| 28 |
+
from .utils.py_utils import NonMutableDict, asdict
|
| 29 |
+
|
| 30 |
+
|
| 31 |
+
@dataclass
|
| 32 |
+
class SplitInfo:
|
| 33 |
+
name: str = dataclasses.field(default="", metadata={"include_in_asdict_even_if_is_default": True})
|
| 34 |
+
num_bytes: int = dataclasses.field(default=0, metadata={"include_in_asdict_even_if_is_default": True})
|
| 35 |
+
num_examples: int = dataclasses.field(default=0, metadata={"include_in_asdict_even_if_is_default": True})
|
| 36 |
+
shard_lengths: Optional[List[int]] = None
|
| 37 |
+
|
| 38 |
+
# Deprecated
|
| 39 |
+
# For backward compatibility, this field needs to always be included in files like
|
| 40 |
+
# dataset_infos.json and dataset_info.json files
|
| 41 |
+
# To do so, we always include it in the output of datasets.utils.py_utils.asdict(split_info)
|
| 42 |
+
dataset_name: Optional[str] = dataclasses.field(
|
| 43 |
+
default=None, metadata={"include_in_asdict_even_if_is_default": True}
|
| 44 |
+
)
|
| 45 |
+
|
| 46 |
+
@property
|
| 47 |
+
def file_instructions(self):
|
| 48 |
+
"""Returns the list of dict(filename, take, skip)."""
|
| 49 |
+
# `self.dataset_name` is assigned in `SplitDict.add()`.
|
| 50 |
+
instructions = make_file_instructions(
|
| 51 |
+
name=self.dataset_name,
|
| 52 |
+
split_infos=[self],
|
| 53 |
+
instruction=str(self.name),
|
| 54 |
+
)
|
| 55 |
+
return instructions.file_instructions
|
| 56 |
+
|
| 57 |
+
|
| 58 |
+
@dataclass
|
| 59 |
+
class SubSplitInfo:
|
| 60 |
+
"""Wrapper around a sub split info.
|
| 61 |
+
This class expose info on the subsplit:
|
| 62 |
+
```
|
| 63 |
+
ds, info = datasets.load_dataset(..., split='train[75%:]', with_info=True)
|
| 64 |
+
info.splits['train[75%:]'].num_examples
|
| 65 |
+
```
|
| 66 |
+
"""
|
| 67 |
+
|
| 68 |
+
instructions: FileInstructions
|
| 69 |
+
|
| 70 |
+
@property
|
| 71 |
+
def num_examples(self):
|
| 72 |
+
"""Returns the number of example in the subsplit."""
|
| 73 |
+
return self.instructions.num_examples
|
| 74 |
+
|
| 75 |
+
@property
|
| 76 |
+
def file_instructions(self):
|
| 77 |
+
"""Returns the list of dict(filename, take, skip)."""
|
| 78 |
+
return self.instructions.file_instructions
|
| 79 |
+
|
| 80 |
+
|
| 81 |
+
class SplitBase(metaclass=abc.ABCMeta):
|
| 82 |
+
# pylint: disable=line-too-long
|
| 83 |
+
"""Abstract base class for Split compositionality.
|
| 84 |
+
|
| 85 |
+
See the
|
| 86 |
+
[guide on splits](../loading#slice-splits)
|
| 87 |
+
for more information.
|
| 88 |
+
|
| 89 |
+
There are three parts to the composition:
|
| 90 |
+
1) The splits are composed (defined, merged, split,...) together before
|
| 91 |
+
calling the `.as_dataset()` function. This is done with the `__add__`,
|
| 92 |
+
`__getitem__`, which return a tree of `SplitBase` (whose leaf
|
| 93 |
+
are the `NamedSplit` objects)
|
| 94 |
+
|
| 95 |
+
```
|
| 96 |
+
split = datasets.Split.TRAIN + datasets.Split.TEST.subsplit(datasets.percent[:50])
|
| 97 |
+
```
|
| 98 |
+
|
| 99 |
+
2) The `SplitBase` is forwarded to the `.as_dataset()` function
|
| 100 |
+
to be resolved into actual read instruction. This is done by the
|
| 101 |
+
`.get_read_instruction()` method which takes the real dataset splits
|
| 102 |
+
(name, number of shards,...) and parse the tree to return a
|
| 103 |
+
`SplitReadInstruction()` object
|
| 104 |
+
|
| 105 |
+
```
|
| 106 |
+
read_instruction = split.get_read_instruction(self.info.splits)
|
| 107 |
+
```
|
| 108 |
+
|
| 109 |
+
3) The `SplitReadInstruction` is then used in the `tf.data.Dataset` pipeline
|
| 110 |
+
to define which files to read and how to skip examples within file.
|
| 111 |
+
|
| 112 |
+
"""
|
| 113 |
+
|
| 114 |
+
# pylint: enable=line-too-long
|
| 115 |
+
|
| 116 |
+
@abc.abstractmethod
|
| 117 |
+
def get_read_instruction(self, split_dict):
|
| 118 |
+
"""Parse the descriptor tree and compile all read instructions together.
|
| 119 |
+
|
| 120 |
+
Args:
|
| 121 |
+
split_dict: `dict`, The `dict[split_name, SplitInfo]` of the dataset
|
| 122 |
+
|
| 123 |
+
Returns:
|
| 124 |
+
split_read_instruction: `SplitReadInstruction`
|
| 125 |
+
"""
|
| 126 |
+
raise NotImplementedError("Abstract method")
|
| 127 |
+
|
| 128 |
+
def __eq__(self, other):
|
| 129 |
+
"""Equality: datasets.Split.TRAIN == 'train'."""
|
| 130 |
+
if isinstance(other, (NamedSplit, str)):
|
| 131 |
+
return False
|
| 132 |
+
raise NotImplementedError("Equality is not implemented between merged/sub splits.")
|
| 133 |
+
|
| 134 |
+
def __ne__(self, other):
|
| 135 |
+
"""InEquality: datasets.Split.TRAIN != 'test'."""
|
| 136 |
+
return not self.__eq__(other)
|
| 137 |
+
|
| 138 |
+
def __add__(self, other):
|
| 139 |
+
"""Merging: datasets.Split.TRAIN + datasets.Split.TEST."""
|
| 140 |
+
return _SplitMerged(self, other)
|
| 141 |
+
|
| 142 |
+
def subsplit(self, arg=None, k=None, percent=None, weighted=None): # pylint: disable=redefined-outer-name
|
| 143 |
+
"""Divides this split into subsplits.
|
| 144 |
+
|
| 145 |
+
There are 3 ways to define subsplits, which correspond to the 3
|
| 146 |
+
arguments `k` (get `k` even subsplits), `percent` (get a slice of the
|
| 147 |
+
dataset with `datasets.percent`), and `weighted` (get subsplits with proportions
|
| 148 |
+
specified by `weighted`).
|
| 149 |
+
|
| 150 |
+
Example::
|
| 151 |
+
|
| 152 |
+
```
|
| 153 |
+
# 50% train, 50% test
|
| 154 |
+
train, test = split.subsplit(k=2)
|
| 155 |
+
# 50% train, 25% test, 25% validation
|
| 156 |
+
train, test, validation = split.subsplit(weighted=[2, 1, 1])
|
| 157 |
+
# Extract last 20%
|
| 158 |
+
subsplit = split.subsplit(datasets.percent[-20:])
|
| 159 |
+
```
|
| 160 |
+
|
| 161 |
+
Warning: k and weighted will be converted into percent which mean that
|
| 162 |
+
values below the percent will be rounded up or down. The final split may be
|
| 163 |
+
bigger to deal with remainders. For instance:
|
| 164 |
+
|
| 165 |
+
```
|
| 166 |
+
train, test, valid = split.subsplit(k=3) # 33%, 33%, 34%
|
| 167 |
+
s1, s2, s3, s4 = split.subsplit(weighted=[2, 2, 1, 1]) # 33%, 33%, 16%, 18%
|
| 168 |
+
```
|
| 169 |
+
|
| 170 |
+
Args:
|
| 171 |
+
arg: If no kwargs are given, `arg` will be interpreted as one of
|
| 172 |
+
`k`, `percent`, or `weighted` depending on the type.
|
| 173 |
+
For example:
|
| 174 |
+
```
|
| 175 |
+
split.subsplit(10) # Equivalent to split.subsplit(k=10)
|
| 176 |
+
split.subsplit(datasets.percent[:-20]) # percent=datasets.percent[:-20]
|
| 177 |
+
split.subsplit([1, 1, 2]) # weighted=[1, 1, 2]
|
| 178 |
+
```
|
| 179 |
+
k: `int` If set, subdivide the split into `k` equal parts.
|
| 180 |
+
percent: `datasets.percent slice`, return a single subsplit corresponding to
|
| 181 |
+
a slice of the original split. For example:
|
| 182 |
+
`split.subsplit(datasets.percent[-20:]) # Last 20% of the dataset`.
|
| 183 |
+
weighted: `list[int]`, return a list of subsplits whose proportions match
|
| 184 |
+
the normalized sum of the list. For example:
|
| 185 |
+
`split.subsplit(weighted=[1, 1, 2]) # 25%, 25%, 50%`.
|
| 186 |
+
|
| 187 |
+
Returns:
|
| 188 |
+
A subsplit or list of subsplits extracted from this split object.
|
| 189 |
+
"""
|
| 190 |
+
# Note that the percent kwargs redefine the outer name datasets.percent. This
|
| 191 |
+
# is done for consistency (.subsplit(percent=datasets.percent[:40]))
|
| 192 |
+
if sum(bool(x) for x in (arg, k, percent, weighted)) != 1:
|
| 193 |
+
raise ValueError("Only one argument of subsplit should be set.")
|
| 194 |
+
|
| 195 |
+
# Auto deduce k
|
| 196 |
+
if isinstance(arg, int):
|
| 197 |
+
k = arg
|
| 198 |
+
elif isinstance(arg, slice):
|
| 199 |
+
percent = arg
|
| 200 |
+
elif isinstance(arg, list):
|
| 201 |
+
weighted = arg
|
| 202 |
+
|
| 203 |
+
if not (k or percent or weighted):
|
| 204 |
+
raise ValueError(
|
| 205 |
+
f"Invalid split argument {arg}. Only list, slice and int supported. "
|
| 206 |
+
"One of k, weighted or percent should be set to a non empty value."
|
| 207 |
+
)
|
| 208 |
+
|
| 209 |
+
def assert_slices_coverage(slices):
|
| 210 |
+
# Ensure that the expended slices cover all percents.
|
| 211 |
+
assert sum((list(range(*s.indices(100))) for s in slices), []) == list(range(100))
|
| 212 |
+
|
| 213 |
+
if k:
|
| 214 |
+
if not 0 < k <= 100:
|
| 215 |
+
raise ValueError(f"Subsplit k should be between 0 and 100, got {k}")
|
| 216 |
+
shift = 100 // k
|
| 217 |
+
slices = [slice(i * shift, (i + 1) * shift) for i in range(k)]
|
| 218 |
+
# Round up last element to ensure all elements are taken
|
| 219 |
+
slices[-1] = slice(slices[-1].start, 100)
|
| 220 |
+
# Internal check to ensure full coverage
|
| 221 |
+
assert_slices_coverage(slices)
|
| 222 |
+
return tuple(_SubSplit(self, s) for s in slices)
|
| 223 |
+
elif percent:
|
| 224 |
+
return _SubSplit(self, percent)
|
| 225 |
+
elif weighted:
|
| 226 |
+
# Normalize the weighted sum
|
| 227 |
+
total = sum(weighted)
|
| 228 |
+
weighted = [100 * x // total for x in weighted]
|
| 229 |
+
# Create the slice for each of the elements
|
| 230 |
+
start = 0
|
| 231 |
+
stop = 0
|
| 232 |
+
slices = []
|
| 233 |
+
for v in weighted:
|
| 234 |
+
stop += v
|
| 235 |
+
slices.append(slice(start, stop))
|
| 236 |
+
start = stop
|
| 237 |
+
# Round up last element to ensure all elements are taken
|
| 238 |
+
slices[-1] = slice(slices[-1].start, 100)
|
| 239 |
+
# Internal check to ensure full coverage
|
| 240 |
+
assert_slices_coverage(slices)
|
| 241 |
+
return tuple(_SubSplit(self, s) for s in slices)
|
| 242 |
+
else:
|
| 243 |
+
# Should not be possible
|
| 244 |
+
raise ValueError("Could not determine the split")
|
| 245 |
+
|
| 246 |
+
|
| 247 |
+
# 2 requirements:
|
| 248 |
+
# 1. datasets.percent be sliceable
|
| 249 |
+
# 2. datasets.percent be documented
|
| 250 |
+
#
|
| 251 |
+
# Instances are not documented, so we want datasets.percent to be a class, but to
|
| 252 |
+
# have it be sliceable, we need this metaclass.
|
| 253 |
+
class PercentSliceMeta(type):
|
| 254 |
+
def __getitem__(cls, slice_value):
|
| 255 |
+
if not isinstance(slice_value, slice):
|
| 256 |
+
raise ValueError(f"datasets.percent should only be called with slice, not {slice_value}")
|
| 257 |
+
return slice_value
|
| 258 |
+
|
| 259 |
+
|
| 260 |
+
class PercentSlice(metaclass=PercentSliceMeta):
|
| 261 |
+
# pylint: disable=line-too-long
|
| 262 |
+
"""Syntactic sugar for defining slice subsplits: `datasets.percent[75:-5]`.
|
| 263 |
+
|
| 264 |
+
See the
|
| 265 |
+
[guide on splits](../loading#slice-splits)
|
| 266 |
+
for more information.
|
| 267 |
+
"""
|
| 268 |
+
|
| 269 |
+
# pylint: enable=line-too-long
|
| 270 |
+
pass
|
| 271 |
+
|
| 272 |
+
|
| 273 |
+
percent = PercentSlice # pylint: disable=invalid-name
|
| 274 |
+
|
| 275 |
+
|
| 276 |
+
class _SplitMerged(SplitBase):
|
| 277 |
+
"""Represent two split descriptors merged together."""
|
| 278 |
+
|
| 279 |
+
def __init__(self, split1, split2):
|
| 280 |
+
self._split1 = split1
|
| 281 |
+
self._split2 = split2
|
| 282 |
+
|
| 283 |
+
def get_read_instruction(self, split_dict):
|
| 284 |
+
read_instruction1 = self._split1.get_read_instruction(split_dict)
|
| 285 |
+
read_instruction2 = self._split2.get_read_instruction(split_dict)
|
| 286 |
+
return read_instruction1 + read_instruction2
|
| 287 |
+
|
| 288 |
+
def __repr__(self):
|
| 289 |
+
return f"({repr(self._split1)} + {repr(self._split2)})"
|
| 290 |
+
|
| 291 |
+
|
| 292 |
+
class _SubSplit(SplitBase):
|
| 293 |
+
"""Represent a sub split of a split descriptor."""
|
| 294 |
+
|
| 295 |
+
def __init__(self, split, slice_value):
|
| 296 |
+
self._split = split
|
| 297 |
+
self._slice_value = slice_value
|
| 298 |
+
|
| 299 |
+
def get_read_instruction(self, split_dict):
|
| 300 |
+
return self._split.get_read_instruction(split_dict)[self._slice_value]
|
| 301 |
+
|
| 302 |
+
def __repr__(self):
|
| 303 |
+
slice_str = "{start}:{stop}"
|
| 304 |
+
if self._slice_value.step is not None:
|
| 305 |
+
slice_str += ":{step}"
|
| 306 |
+
slice_str = slice_str.format(
|
| 307 |
+
start="" if self._slice_value.start is None else self._slice_value.start,
|
| 308 |
+
stop="" if self._slice_value.stop is None else self._slice_value.stop,
|
| 309 |
+
step=self._slice_value.step,
|
| 310 |
+
)
|
| 311 |
+
return f"{repr(self._split)}(datasets.percent[{slice_str}])"
|
| 312 |
+
|
| 313 |
+
|
| 314 |
+
class NamedSplit(SplitBase):
|
| 315 |
+
"""Descriptor corresponding to a named split (train, test, ...).
|
| 316 |
+
|
| 317 |
+
Example:
|
| 318 |
+
Each descriptor can be composed with other using addition or slice:
|
| 319 |
+
|
| 320 |
+
```py
|
| 321 |
+
split = datasets.Split.TRAIN.subsplit(datasets.percent[0:25]) + datasets.Split.TEST
|
| 322 |
+
```
|
| 323 |
+
|
| 324 |
+
The resulting split will correspond to 25% of the train split merged with
|
| 325 |
+
100% of the test split.
|
| 326 |
+
|
| 327 |
+
A split cannot be added twice, so the following will fail:
|
| 328 |
+
|
| 329 |
+
```py
|
| 330 |
+
split = (
|
| 331 |
+
datasets.Split.TRAIN.subsplit(datasets.percent[:25]) +
|
| 332 |
+
datasets.Split.TRAIN.subsplit(datasets.percent[75:])
|
| 333 |
+
) # Error
|
| 334 |
+
split = datasets.Split.TEST + datasets.Split.ALL # Error
|
| 335 |
+
```
|
| 336 |
+
|
| 337 |
+
The slices can be applied only one time. So the following are valid:
|
| 338 |
+
|
| 339 |
+
```py
|
| 340 |
+
split = (
|
| 341 |
+
datasets.Split.TRAIN.subsplit(datasets.percent[:25]) +
|
| 342 |
+
datasets.Split.TEST.subsplit(datasets.percent[:50])
|
| 343 |
+
)
|
| 344 |
+
split = (datasets.Split.TRAIN + datasets.Split.TEST).subsplit(datasets.percent[:50])
|
| 345 |
+
```
|
| 346 |
+
|
| 347 |
+
But this is not valid:
|
| 348 |
+
|
| 349 |
+
```py
|
| 350 |
+
train = datasets.Split.TRAIN
|
| 351 |
+
test = datasets.Split.TEST
|
| 352 |
+
split = train.subsplit(datasets.percent[:25]).subsplit(datasets.percent[:25])
|
| 353 |
+
split = (train.subsplit(datasets.percent[:25]) + test).subsplit(datasets.percent[:50])
|
| 354 |
+
```
|
| 355 |
+
"""
|
| 356 |
+
|
| 357 |
+
def __init__(self, name):
|
| 358 |
+
self._name = name
|
| 359 |
+
split_names_from_instruction = [split_instruction.split("[")[0] for split_instruction in name.split("+")]
|
| 360 |
+
for split_name in split_names_from_instruction:
|
| 361 |
+
if not re.match(_split_re, split_name):
|
| 362 |
+
raise ValueError(f"Split name should match '{_split_re}' but got '{split_name}'.")
|
| 363 |
+
|
| 364 |
+
def __str__(self):
|
| 365 |
+
return self._name
|
| 366 |
+
|
| 367 |
+
def __repr__(self):
|
| 368 |
+
return f"NamedSplit({self._name!r})"
|
| 369 |
+
|
| 370 |
+
def __eq__(self, other):
|
| 371 |
+
"""Equality: datasets.Split.TRAIN == 'train'."""
|
| 372 |
+
if isinstance(other, NamedSplit):
|
| 373 |
+
return self._name == other._name # pylint: disable=protected-access
|
| 374 |
+
elif isinstance(other, SplitBase):
|
| 375 |
+
return False
|
| 376 |
+
elif isinstance(other, str): # Other should be string
|
| 377 |
+
return self._name == other
|
| 378 |
+
else:
|
| 379 |
+
raise ValueError(f"Equality not supported between split {self} and {other}")
|
| 380 |
+
|
| 381 |
+
def __lt__(self, other):
|
| 382 |
+
return self._name < other._name # pylint: disable=protected-access
|
| 383 |
+
|
| 384 |
+
def __hash__(self):
|
| 385 |
+
return hash(self._name)
|
| 386 |
+
|
| 387 |
+
def get_read_instruction(self, split_dict):
|
| 388 |
+
return SplitReadInstruction(split_dict[self._name])
|
| 389 |
+
|
| 390 |
+
|
| 391 |
+
class NamedSplitAll(NamedSplit):
|
| 392 |
+
"""Split corresponding to the union of all defined dataset splits."""
|
| 393 |
+
|
| 394 |
+
def __init__(self):
|
| 395 |
+
super().__init__("all")
|
| 396 |
+
|
| 397 |
+
def __repr__(self):
|
| 398 |
+
return "NamedSplitAll()"
|
| 399 |
+
|
| 400 |
+
def get_read_instruction(self, split_dict):
|
| 401 |
+
# Merge all dataset split together
|
| 402 |
+
read_instructions = [SplitReadInstruction(s) for s in split_dict.values()]
|
| 403 |
+
return sum(read_instructions, SplitReadInstruction())
|
| 404 |
+
|
| 405 |
+
|
| 406 |
+
class Split:
|
| 407 |
+
# pylint: disable=line-too-long
|
| 408 |
+
"""`Enum` for dataset splits.
|
| 409 |
+
|
| 410 |
+
Datasets are typically split into different subsets to be used at various
|
| 411 |
+
stages of training and evaluation.
|
| 412 |
+
|
| 413 |
+
- `TRAIN`: the training data.
|
| 414 |
+
- `VALIDATION`: the validation data. If present, this is typically used as
|
| 415 |
+
evaluation data while iterating on a model (e.g. changing hyperparameters,
|
| 416 |
+
model architecture, etc.).
|
| 417 |
+
- `TEST`: the testing data. This is the data to report metrics on. Typically
|
| 418 |
+
you do not want to use this during model iteration as you may overfit to it.
|
| 419 |
+
- `ALL`: the union of all defined dataset splits.
|
| 420 |
+
|
| 421 |
+
All splits, including compositions inherit from `datasets.SplitBase`.
|
| 422 |
+
|
| 423 |
+
See the [guide](../load_hub#splits) on splits for more information.
|
| 424 |
+
|
| 425 |
+
Example:
|
| 426 |
+
|
| 427 |
+
```py
|
| 428 |
+
>>> datasets.SplitGenerator(
|
| 429 |
+
... name=datasets.Split.TRAIN,
|
| 430 |
+
... gen_kwargs={"split_key": "train", "files": dl_manager.download_and extract(url)},
|
| 431 |
+
... ),
|
| 432 |
+
... datasets.SplitGenerator(
|
| 433 |
+
... name=datasets.Split.VALIDATION,
|
| 434 |
+
... gen_kwargs={"split_key": "validation", "files": dl_manager.download_and extract(url)},
|
| 435 |
+
... ),
|
| 436 |
+
... datasets.SplitGenerator(
|
| 437 |
+
... name=datasets.Split.TEST,
|
| 438 |
+
... gen_kwargs={"split_key": "test", "files": dl_manager.download_and extract(url)},
|
| 439 |
+
... )
|
| 440 |
+
```
|
| 441 |
+
"""
|
| 442 |
+
|
| 443 |
+
# pylint: enable=line-too-long
|
| 444 |
+
TRAIN = NamedSplit("train")
|
| 445 |
+
TEST = NamedSplit("test")
|
| 446 |
+
VALIDATION = NamedSplit("validation")
|
| 447 |
+
ALL = NamedSplitAll()
|
| 448 |
+
|
| 449 |
+
def __new__(cls, name):
|
| 450 |
+
"""Create a custom split with datasets.Split('custom_name')."""
|
| 451 |
+
return NamedSplitAll() if name == "all" else NamedSplit(name)
|
| 452 |
+
|
| 453 |
+
|
| 454 |
+
# Similar to SplitInfo, but contain an additional slice info
|
| 455 |
+
SlicedSplitInfo = collections.namedtuple(
|
| 456 |
+
"SlicedSplitInfo",
|
| 457 |
+
[
|
| 458 |
+
"split_info",
|
| 459 |
+
"slice_value",
|
| 460 |
+
],
|
| 461 |
+
) # noqa: E231
|
| 462 |
+
|
| 463 |
+
|
| 464 |
+
class SplitReadInstruction:
|
| 465 |
+
"""Object containing the reading instruction for the dataset.
|
| 466 |
+
|
| 467 |
+
Similarly to `SplitDescriptor` nodes, this object can be composed with itself,
|
| 468 |
+
but the resolution happens instantaneously, instead of keeping track of the
|
| 469 |
+
tree, such as all instructions are compiled and flattened in a single
|
| 470 |
+
SplitReadInstruction object containing the list of files and slice to use.
|
| 471 |
+
|
| 472 |
+
Once resolved, the instructions can be accessed with:
|
| 473 |
+
|
| 474 |
+
```
|
| 475 |
+
read_instructions.get_list_sliced_split_info() # List of splits to use
|
| 476 |
+
```
|
| 477 |
+
|
| 478 |
+
"""
|
| 479 |
+
|
| 480 |
+
def __init__(self, split_info=None):
|
| 481 |
+
self._splits = NonMutableDict(error_msg="Overlap between splits. Split {key} has been added with " "itself.")
|
| 482 |
+
|
| 483 |
+
if split_info:
|
| 484 |
+
self.add(SlicedSplitInfo(split_info=split_info, slice_value=None))
|
| 485 |
+
|
| 486 |
+
def add(self, sliced_split):
|
| 487 |
+
"""Add a SlicedSplitInfo the read instructions."""
|
| 488 |
+
# TODO(epot): Check that the number of examples per shard % 100 == 0
|
| 489 |
+
# Otherwise the slices value may be unbalanced and not exactly reflect the
|
| 490 |
+
# requested slice.
|
| 491 |
+
self._splits[sliced_split.split_info.name] = sliced_split
|
| 492 |
+
|
| 493 |
+
def __add__(self, other):
|
| 494 |
+
"""Merging split together."""
|
| 495 |
+
# Will raise error if a split has already be added (NonMutableDict)
|
| 496 |
+
# TODO(epot): If a split is already added but there is no overlap between
|
| 497 |
+
# the slices, should merge the slices (ex: [:10] + [80:])
|
| 498 |
+
split_instruction = SplitReadInstruction()
|
| 499 |
+
split_instruction._splits.update(self._splits) # pylint: disable=protected-access
|
| 500 |
+
split_instruction._splits.update(other._splits) # pylint: disable=protected-access
|
| 501 |
+
return split_instruction
|
| 502 |
+
|
| 503 |
+
def __getitem__(self, slice_value):
|
| 504 |
+
"""Sub-splits."""
|
| 505 |
+
# Will raise an error if a split has already been sliced
|
| 506 |
+
split_instruction = SplitReadInstruction()
|
| 507 |
+
for v in self._splits.values():
|
| 508 |
+
if v.slice_value is not None:
|
| 509 |
+
raise ValueError(f"Trying to slice Split {v.split_info.name} which has already been sliced")
|
| 510 |
+
v = v._asdict()
|
| 511 |
+
v["slice_value"] = slice_value
|
| 512 |
+
split_instruction.add(SlicedSplitInfo(**v))
|
| 513 |
+
return split_instruction
|
| 514 |
+
|
| 515 |
+
def get_list_sliced_split_info(self):
|
| 516 |
+
return list(self._splits.values())
|
| 517 |
+
|
| 518 |
+
|
| 519 |
+
class SplitDict(dict):
|
| 520 |
+
"""Split info object."""
|
| 521 |
+
|
| 522 |
+
def __init__(self, *args, dataset_name=None, **kwargs):
|
| 523 |
+
super().__init__(*args, **kwargs)
|
| 524 |
+
self.dataset_name = dataset_name
|
| 525 |
+
|
| 526 |
+
def __getitem__(self, key: Union[SplitBase, str]):
|
| 527 |
+
# 1st case: The key exists: `info.splits['train']`
|
| 528 |
+
if str(key) in self:
|
| 529 |
+
return super().__getitem__(str(key))
|
| 530 |
+
# 2nd case: Uses instructions: `info.splits['train[50%]']`
|
| 531 |
+
else:
|
| 532 |
+
instructions = make_file_instructions(
|
| 533 |
+
name=self.dataset_name,
|
| 534 |
+
split_infos=self.values(),
|
| 535 |
+
instruction=key,
|
| 536 |
+
)
|
| 537 |
+
return SubSplitInfo(instructions)
|
| 538 |
+
|
| 539 |
+
def __setitem__(self, key: Union[SplitBase, str], value: SplitInfo):
|
| 540 |
+
if key != value.name:
|
| 541 |
+
raise ValueError(f"Cannot add elem. (key mismatch: '{key}' != '{value.name}')")
|
| 542 |
+
super().__setitem__(key, value)
|
| 543 |
+
|
| 544 |
+
def add(self, split_info: SplitInfo):
|
| 545 |
+
"""Add the split info."""
|
| 546 |
+
if split_info.name in self:
|
| 547 |
+
raise ValueError(f"Split {split_info.name} already present")
|
| 548 |
+
split_info.dataset_name = self.dataset_name
|
| 549 |
+
super().__setitem__(split_info.name, split_info)
|
| 550 |
+
|
| 551 |
+
@property
|
| 552 |
+
def total_num_examples(self):
|
| 553 |
+
"""Return the total number of examples."""
|
| 554 |
+
return sum(s.num_examples for s in self.values())
|
| 555 |
+
|
| 556 |
+
@classmethod
|
| 557 |
+
def from_split_dict(cls, split_infos: Union[List, Dict], dataset_name: Optional[str] = None):
|
| 558 |
+
"""Returns a new SplitDict initialized from a Dict or List of `split_infos`."""
|
| 559 |
+
if isinstance(split_infos, dict):
|
| 560 |
+
split_infos = list(split_infos.values())
|
| 561 |
+
|
| 562 |
+
if dataset_name is None:
|
| 563 |
+
dataset_name = split_infos[0].get("dataset_name") if split_infos else None
|
| 564 |
+
|
| 565 |
+
split_dict = cls(dataset_name=dataset_name)
|
| 566 |
+
|
| 567 |
+
for split_info in split_infos:
|
| 568 |
+
if isinstance(split_info, dict):
|
| 569 |
+
split_info = SplitInfo(**split_info)
|
| 570 |
+
split_dict.add(split_info)
|
| 571 |
+
|
| 572 |
+
return split_dict
|
| 573 |
+
|
| 574 |
+
def to_split_dict(self):
|
| 575 |
+
"""Returns a list of SplitInfo protos that we have."""
|
| 576 |
+
out = []
|
| 577 |
+
for split_name, split_info in self.items():
|
| 578 |
+
split_info = copy.deepcopy(split_info)
|
| 579 |
+
split_info.name = split_name
|
| 580 |
+
out.append(split_info)
|
| 581 |
+
return out
|
| 582 |
+
|
| 583 |
+
def copy(self):
|
| 584 |
+
return SplitDict.from_split_dict(self.to_split_dict(), self.dataset_name)
|
| 585 |
+
|
| 586 |
+
def _to_yaml_list(self) -> list:
|
| 587 |
+
out = [asdict(s) for s in self.to_split_dict()]
|
| 588 |
+
# we don't need the shard lengths in YAML, since it depends on max_shard_size and num_proc
|
| 589 |
+
for split_info_dict in out:
|
| 590 |
+
split_info_dict.pop("shard_lengths", None)
|
| 591 |
+
# we don't need the dataset_name attribute that is deprecated
|
| 592 |
+
for split_info_dict in out:
|
| 593 |
+
split_info_dict.pop("dataset_name", None)
|
| 594 |
+
return out
|
| 595 |
+
|
| 596 |
+
@classmethod
|
| 597 |
+
def _from_yaml_list(cls, yaml_data: list) -> "SplitDict":
|
| 598 |
+
return cls.from_split_dict(yaml_data)
|
| 599 |
+
|
| 600 |
+
|
| 601 |
+
@dataclass
|
| 602 |
+
class SplitGenerator:
|
| 603 |
+
"""Defines the split information for the generator.
|
| 604 |
+
|
| 605 |
+
This should be used as returned value of
|
| 606 |
+
`GeneratorBasedBuilder._split_generators`.
|
| 607 |
+
See `GeneratorBasedBuilder._split_generators` for more info and example
|
| 608 |
+
of usage.
|
| 609 |
+
|
| 610 |
+
Args:
|
| 611 |
+
name (`str`):
|
| 612 |
+
Name of the `Split` for which the generator will
|
| 613 |
+
create the examples.
|
| 614 |
+
**gen_kwargs (additional keyword arguments):
|
| 615 |
+
Keyword arguments to forward to the `DatasetBuilder._generate_examples` method
|
| 616 |
+
of the builder.
|
| 617 |
+
|
| 618 |
+
Example:
|
| 619 |
+
|
| 620 |
+
```py
|
| 621 |
+
>>> datasets.SplitGenerator(
|
| 622 |
+
... name=datasets.Split.TRAIN,
|
| 623 |
+
... gen_kwargs={"split_key": "train", "files": dl_manager.download_and_extract(url)},
|
| 624 |
+
... )
|
| 625 |
+
```
|
| 626 |
+
"""
|
| 627 |
+
|
| 628 |
+
name: str
|
| 629 |
+
gen_kwargs: Dict = dataclasses.field(default_factory=dict)
|
| 630 |
+
split_info: SplitInfo = dataclasses.field(init=False)
|
| 631 |
+
|
| 632 |
+
def __post_init__(self):
|
| 633 |
+
self.name = str(self.name) # Make sure we convert NamedSplits in strings
|
| 634 |
+
NamedSplit(self.name) # check that it's a valid split name
|
| 635 |
+
self.split_info = SplitInfo(name=self.name)
|
evalkit_tf437/lib/python3.10/site-packages/datasets/streaming.py
ADDED
|
@@ -0,0 +1,140 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import importlib
|
| 2 |
+
import inspect
|
| 3 |
+
from functools import wraps
|
| 4 |
+
from typing import TYPE_CHECKING, Optional
|
| 5 |
+
|
| 6 |
+
from .download.download_config import DownloadConfig
|
| 7 |
+
from .download.streaming_download_manager import (
|
| 8 |
+
xbasename,
|
| 9 |
+
xdirname,
|
| 10 |
+
xet_parse,
|
| 11 |
+
xexists,
|
| 12 |
+
xgetsize,
|
| 13 |
+
xglob,
|
| 14 |
+
xgzip_open,
|
| 15 |
+
xisdir,
|
| 16 |
+
xisfile,
|
| 17 |
+
xjoin,
|
| 18 |
+
xlistdir,
|
| 19 |
+
xnumpy_load,
|
| 20 |
+
xopen,
|
| 21 |
+
xpandas_read_csv,
|
| 22 |
+
xpandas_read_excel,
|
| 23 |
+
xPath,
|
| 24 |
+
xpyarrow_parquet_read_table,
|
| 25 |
+
xrelpath,
|
| 26 |
+
xsio_loadmat,
|
| 27 |
+
xsplit,
|
| 28 |
+
xsplitext,
|
| 29 |
+
xwalk,
|
| 30 |
+
xxml_dom_minidom_parse,
|
| 31 |
+
)
|
| 32 |
+
from .utils.logging import get_logger
|
| 33 |
+
from .utils.patching import patch_submodule
|
| 34 |
+
from .utils.py_utils import get_imports
|
| 35 |
+
|
| 36 |
+
|
| 37 |
+
logger = get_logger(__name__)
|
| 38 |
+
|
| 39 |
+
|
| 40 |
+
if TYPE_CHECKING:
|
| 41 |
+
from .builder import DatasetBuilder
|
| 42 |
+
|
| 43 |
+
|
| 44 |
+
def extend_module_for_streaming(module_path, download_config: Optional[DownloadConfig] = None):
|
| 45 |
+
"""Extend the module to support streaming.
|
| 46 |
+
|
| 47 |
+
We patch some functions in the module to use `fsspec` to support data streaming:
|
| 48 |
+
- We use `fsspec.open` to open and read remote files. We patch the module function:
|
| 49 |
+
- `open`
|
| 50 |
+
- We use the "::" hop separator to join paths and navigate remote compressed/archive files. We patch the module
|
| 51 |
+
functions:
|
| 52 |
+
- `os.path.join`
|
| 53 |
+
- `pathlib.Path.joinpath` and `pathlib.Path.__truediv__` (called when using the "/" operator)
|
| 54 |
+
|
| 55 |
+
The patched functions are replaced with custom functions defined to work with the
|
| 56 |
+
:class:`~download.streaming_download_manager.StreamingDownloadManager`.
|
| 57 |
+
|
| 58 |
+
Args:
|
| 59 |
+
module_path: Path to the module to be extended.
|
| 60 |
+
download_config : mainly use use_auth_token or storage_options to support different platforms and auth types.
|
| 61 |
+
"""
|
| 62 |
+
|
| 63 |
+
module = importlib.import_module(module_path)
|
| 64 |
+
|
| 65 |
+
# TODO(QL): always update the module to add subsequent new authentication without removing old ones
|
| 66 |
+
if hasattr(module, "_patched_for_streaming") and module._patched_for_streaming:
|
| 67 |
+
if isinstance(module._patched_for_streaming, DownloadConfig):
|
| 68 |
+
module._patched_for_streaming.token = download_config.token
|
| 69 |
+
module._patched_for_streaming.storage_options = download_config.storage_options
|
| 70 |
+
return
|
| 71 |
+
|
| 72 |
+
def wrap_auth(function):
|
| 73 |
+
@wraps(function)
|
| 74 |
+
def wrapper(*args, **kwargs):
|
| 75 |
+
return function(*args, download_config=download_config, **kwargs)
|
| 76 |
+
|
| 77 |
+
wrapper._decorator_name_ = "wrap_auth"
|
| 78 |
+
return wrapper
|
| 79 |
+
|
| 80 |
+
# open files in a streaming fashion
|
| 81 |
+
patch_submodule(module, "open", wrap_auth(xopen)).start()
|
| 82 |
+
patch_submodule(module, "os.listdir", wrap_auth(xlistdir)).start()
|
| 83 |
+
patch_submodule(module, "os.walk", wrap_auth(xwalk)).start()
|
| 84 |
+
patch_submodule(module, "glob.glob", wrap_auth(xglob)).start()
|
| 85 |
+
# allow to navigate in remote zip files
|
| 86 |
+
patch_submodule(module, "os.path.join", xjoin).start()
|
| 87 |
+
patch_submodule(module, "os.path.dirname", xdirname).start()
|
| 88 |
+
patch_submodule(module, "os.path.basename", xbasename).start()
|
| 89 |
+
patch_submodule(module, "os.path.relpath", xrelpath).start()
|
| 90 |
+
patch_submodule(module, "os.path.split", xsplit).start()
|
| 91 |
+
patch_submodule(module, "os.path.splitext", xsplitext).start()
|
| 92 |
+
# allow checks on paths
|
| 93 |
+
patch_submodule(module, "os.path.exists", wrap_auth(xexists)).start()
|
| 94 |
+
patch_submodule(module, "os.path.isdir", wrap_auth(xisdir)).start()
|
| 95 |
+
patch_submodule(module, "os.path.isfile", wrap_auth(xisfile)).start()
|
| 96 |
+
patch_submodule(module, "os.path.getsize", wrap_auth(xgetsize)).start()
|
| 97 |
+
patch_submodule(module, "pathlib.Path", xPath).start()
|
| 98 |
+
# file readers
|
| 99 |
+
patch_submodule(module, "gzip.open", wrap_auth(xgzip_open)).start()
|
| 100 |
+
patch_submodule(module, "numpy.load", wrap_auth(xnumpy_load)).start()
|
| 101 |
+
patch_submodule(module, "pandas.read_csv", wrap_auth(xpandas_read_csv), attrs=["__version__"]).start()
|
| 102 |
+
patch_submodule(module, "pandas.read_excel", wrap_auth(xpandas_read_excel), attrs=["__version__"]).start()
|
| 103 |
+
patch_submodule(module, "scipy.io.loadmat", wrap_auth(xsio_loadmat), attrs=["__version__"]).start()
|
| 104 |
+
patch_submodule(module, "xml.etree.ElementTree.parse", wrap_auth(xet_parse)).start()
|
| 105 |
+
patch_submodule(module, "xml.dom.minidom.parse", wrap_auth(xxml_dom_minidom_parse)).start()
|
| 106 |
+
# pyarrow: do not patch pyarrow attribute in packaged modules
|
| 107 |
+
if not module.__name__.startswith("datasets.packaged_modules."):
|
| 108 |
+
patch_submodule(module, "pyarrow.parquet.read_table", wrap_auth(xpyarrow_parquet_read_table)).start()
|
| 109 |
+
module._patched_for_streaming = download_config
|
| 110 |
+
|
| 111 |
+
|
| 112 |
+
def extend_dataset_builder_for_streaming(builder: "DatasetBuilder"):
|
| 113 |
+
"""Extend the dataset builder module and the modules imported by it to support streaming.
|
| 114 |
+
|
| 115 |
+
Args:
|
| 116 |
+
builder (:class:`DatasetBuilder`): Dataset builder instance.
|
| 117 |
+
"""
|
| 118 |
+
# this extends the open and os.path.join functions for data streaming
|
| 119 |
+
download_config = DownloadConfig(storage_options=builder.storage_options, token=builder.token)
|
| 120 |
+
extend_module_for_streaming(builder.__module__, download_config=download_config)
|
| 121 |
+
# if needed, we also have to extend additional internal imports (like wmt14 -> wmt_utils)
|
| 122 |
+
if not builder.__module__.startswith("datasets."): # check that it's not a packaged builder like csv
|
| 123 |
+
for imports in get_imports(inspect.getfile(builder.__class__)):
|
| 124 |
+
if imports[0] == "internal":
|
| 125 |
+
internal_import_name = imports[1]
|
| 126 |
+
internal_module_name = ".".join(builder.__module__.split(".")[:-1] + [internal_import_name])
|
| 127 |
+
extend_module_for_streaming(internal_module_name, download_config=download_config)
|
| 128 |
+
|
| 129 |
+
# builders can inherit from other builders that might use streaming functionality
|
| 130 |
+
# (for example, ImageFolder and AudioFolder inherit from FolderBuilder which implements examples generation)
|
| 131 |
+
# but these parents builders are not patched automatically as they are not instantiated, so we patch them here
|
| 132 |
+
from .builder import DatasetBuilder
|
| 133 |
+
|
| 134 |
+
parent_builder_modules = [
|
| 135 |
+
cls.__module__
|
| 136 |
+
for cls in type(builder).__mro__[1:] # make sure it's not the same module we've already patched
|
| 137 |
+
if issubclass(cls, DatasetBuilder) and cls.__module__ != DatasetBuilder.__module__
|
| 138 |
+
] # check it's not a standard builder from datasets.builder
|
| 139 |
+
for module in parent_builder_modules:
|
| 140 |
+
extend_module_for_streaming(module, download_config=download_config)
|
evalkit_tf437/lib/python3.10/site-packages/datasets/table.py
ADDED
|
@@ -0,0 +1,2360 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import copy
|
| 2 |
+
import os
|
| 3 |
+
from functools import partial
|
| 4 |
+
from itertools import groupby
|
| 5 |
+
from typing import TYPE_CHECKING, Callable, Iterator, List, Optional, Tuple, TypeVar, Union
|
| 6 |
+
|
| 7 |
+
import numpy as np
|
| 8 |
+
import pyarrow as pa
|
| 9 |
+
import pyarrow.compute as pc
|
| 10 |
+
import pyarrow.types
|
| 11 |
+
|
| 12 |
+
from . import config
|
| 13 |
+
from .utils.logging import get_logger
|
| 14 |
+
|
| 15 |
+
|
| 16 |
+
if TYPE_CHECKING:
|
| 17 |
+
from .features.features import Features, FeatureType
|
| 18 |
+
|
| 19 |
+
|
| 20 |
+
logger = get_logger(__name__)
|
| 21 |
+
|
| 22 |
+
|
| 23 |
+
def inject_arrow_table_documentation(arrow_table_method):
|
| 24 |
+
def wrapper(fn):
|
| 25 |
+
fn.__doc__ = arrow_table_method.__doc__ + (fn.__doc__ if fn.__doc__ is not None else "")
|
| 26 |
+
fn.__doc__ = fn.__doc__.replace("pyarrow.Table", "Table")
|
| 27 |
+
if hasattr(arrow_table_method, "__annotations__"):
|
| 28 |
+
fn.__annotations__ = arrow_table_method.__annotations__
|
| 29 |
+
return fn
|
| 30 |
+
|
| 31 |
+
return wrapper
|
| 32 |
+
|
| 33 |
+
|
| 34 |
+
def _in_memory_arrow_table_from_file(filename: str) -> pa.Table:
|
| 35 |
+
in_memory_stream = pa.input_stream(filename)
|
| 36 |
+
opened_stream = pa.ipc.open_stream(in_memory_stream)
|
| 37 |
+
pa_table = opened_stream.read_all()
|
| 38 |
+
return pa_table
|
| 39 |
+
|
| 40 |
+
|
| 41 |
+
def _in_memory_arrow_table_from_buffer(buffer: pa.Buffer) -> pa.Table:
|
| 42 |
+
stream = pa.BufferReader(buffer)
|
| 43 |
+
opened_stream = pa.ipc.open_stream(stream)
|
| 44 |
+
table = opened_stream.read_all()
|
| 45 |
+
return table
|
| 46 |
+
|
| 47 |
+
|
| 48 |
+
def _memory_mapped_record_batch_reader_from_file(filename: str) -> pa.RecordBatchStreamReader:
|
| 49 |
+
memory_mapped_stream = pa.memory_map(filename)
|
| 50 |
+
return pa.ipc.open_stream(memory_mapped_stream)
|
| 51 |
+
|
| 52 |
+
|
| 53 |
+
def read_schema_from_file(filename: str) -> pa.Schema:
|
| 54 |
+
"""
|
| 55 |
+
Infer arrow table schema from file without loading whole file into memory.
|
| 56 |
+
Usefull especially while having very big files.
|
| 57 |
+
"""
|
| 58 |
+
with pa.memory_map(filename) as memory_mapped_stream:
|
| 59 |
+
schema = pa.ipc.open_stream(memory_mapped_stream).schema
|
| 60 |
+
return schema
|
| 61 |
+
|
| 62 |
+
|
| 63 |
+
def _memory_mapped_arrow_table_from_file(filename: str) -> pa.Table:
|
| 64 |
+
opened_stream = _memory_mapped_record_batch_reader_from_file(filename)
|
| 65 |
+
pa_table = opened_stream.read_all()
|
| 66 |
+
return pa_table
|
| 67 |
+
|
| 68 |
+
|
| 69 |
+
def _deepcopy(x, memo: dict):
|
| 70 |
+
"""deepcopy a regular class instance"""
|
| 71 |
+
cls = x.__class__
|
| 72 |
+
result = cls.__new__(cls)
|
| 73 |
+
memo[id(x)] = result
|
| 74 |
+
for k, v in x.__dict__.items():
|
| 75 |
+
setattr(result, k, copy.deepcopy(v, memo))
|
| 76 |
+
return result
|
| 77 |
+
|
| 78 |
+
|
| 79 |
+
def _interpolation_search(arr: List[int], x: int) -> int:
|
| 80 |
+
"""
|
| 81 |
+
Return the position i of a sorted array so that arr[i] <= x < arr[i+1]
|
| 82 |
+
|
| 83 |
+
Args:
|
| 84 |
+
arr (`List[int]`): non-empty sorted list of integers
|
| 85 |
+
x (`int`): query
|
| 86 |
+
|
| 87 |
+
Returns:
|
| 88 |
+
`int`: the position i so that arr[i] <= x < arr[i+1]
|
| 89 |
+
|
| 90 |
+
Raises:
|
| 91 |
+
`IndexError`: if the array is empty or if the query is outside the array values
|
| 92 |
+
"""
|
| 93 |
+
i, j = 0, len(arr) - 1
|
| 94 |
+
while i < j and arr[i] <= x < arr[j]:
|
| 95 |
+
k = i + ((j - i) * (x - arr[i]) // (arr[j] - arr[i]))
|
| 96 |
+
if arr[k] <= x < arr[k + 1]:
|
| 97 |
+
return k
|
| 98 |
+
elif arr[k] < x:
|
| 99 |
+
i, j = k + 1, j
|
| 100 |
+
else:
|
| 101 |
+
i, j = i, k
|
| 102 |
+
raise IndexError(f"Invalid query '{x}' for size {arr[-1] if len(arr) else 'none'}.")
|
| 103 |
+
|
| 104 |
+
|
| 105 |
+
class IndexedTableMixin:
|
| 106 |
+
def __init__(self, table: pa.Table):
|
| 107 |
+
self._schema: pa.Schema = table.schema
|
| 108 |
+
self._batches: List[pa.RecordBatch] = [
|
| 109 |
+
recordbatch for recordbatch in table.to_batches() if len(recordbatch) > 0
|
| 110 |
+
]
|
| 111 |
+
self._offsets: np.ndarray = np.cumsum([0] + [len(b) for b in self._batches], dtype=np.int64)
|
| 112 |
+
|
| 113 |
+
def fast_gather(self, indices: Union[List[int], np.ndarray]) -> pa.Table:
|
| 114 |
+
"""
|
| 115 |
+
Create a pa.Table by gathering the records at the records at the specified indices. Should be faster
|
| 116 |
+
than pa.concat_tables(table.fast_slice(int(i) % table.num_rows, 1) for i in indices) since NumPy can compute
|
| 117 |
+
the binary searches in parallel, highly optimized C
|
| 118 |
+
"""
|
| 119 |
+
if not len(indices):
|
| 120 |
+
raise ValueError("Indices must be non-empty")
|
| 121 |
+
batch_indices = np.searchsorted(self._offsets, indices, side="right") - 1
|
| 122 |
+
return pa.Table.from_batches(
|
| 123 |
+
[
|
| 124 |
+
self._batches[batch_idx].slice(i - self._offsets[batch_idx], 1)
|
| 125 |
+
for batch_idx, i in zip(batch_indices, indices)
|
| 126 |
+
],
|
| 127 |
+
schema=self._schema,
|
| 128 |
+
)
|
| 129 |
+
|
| 130 |
+
def fast_slice(self, offset=0, length=None) -> pa.Table:
|
| 131 |
+
"""
|
| 132 |
+
Slice the Table using interpolation search.
|
| 133 |
+
The behavior is the same as `pyarrow.Table.slice` but it's significantly faster.
|
| 134 |
+
|
| 135 |
+
Interpolation search is used to find the start and end indexes of the batches we want to keep.
|
| 136 |
+
The batches to keep are then concatenated to form the sliced Table.
|
| 137 |
+
"""
|
| 138 |
+
if offset < 0:
|
| 139 |
+
raise IndexError("Offset must be non-negative")
|
| 140 |
+
elif offset >= self._offsets[-1] or (length is not None and length <= 0):
|
| 141 |
+
return pa.Table.from_batches([], schema=self._schema)
|
| 142 |
+
i = _interpolation_search(self._offsets, offset)
|
| 143 |
+
if length is None or length + offset >= self._offsets[-1]:
|
| 144 |
+
batches = self._batches[i:]
|
| 145 |
+
batches[0] = batches[0].slice(offset - self._offsets[i])
|
| 146 |
+
else:
|
| 147 |
+
j = _interpolation_search(self._offsets, offset + length - 1)
|
| 148 |
+
batches = self._batches[i : j + 1]
|
| 149 |
+
batches[-1] = batches[-1].slice(0, offset + length - self._offsets[j])
|
| 150 |
+
batches[0] = batches[0].slice(offset - self._offsets[i])
|
| 151 |
+
return pa.Table.from_batches(batches, schema=self._schema)
|
| 152 |
+
|
| 153 |
+
|
| 154 |
+
class Table(IndexedTableMixin):
|
| 155 |
+
"""
|
| 156 |
+
Wraps a pyarrow Table by using composition.
|
| 157 |
+
This is the base class for `InMemoryTable`, `MemoryMappedTable` and `ConcatenationTable`.
|
| 158 |
+
|
| 159 |
+
It implements all the basic attributes/methods of the pyarrow Table class except
|
| 160 |
+
the Table transforms: `slice, filter, flatten, combine_chunks, cast, add_column,
|
| 161 |
+
append_column, remove_column, set_column, rename_columns` and `drop`.
|
| 162 |
+
|
| 163 |
+
The implementation of these methods differs for the subclasses.
|
| 164 |
+
"""
|
| 165 |
+
|
| 166 |
+
def __init__(self, table: pa.Table):
|
| 167 |
+
super().__init__(table)
|
| 168 |
+
self.table = table
|
| 169 |
+
|
| 170 |
+
def __deepcopy__(self, memo: dict):
|
| 171 |
+
# arrow tables are immutable, so there's no need to copy self.table
|
| 172 |
+
# moreover calling deepcopy on a pyarrow table seems to make pa.total_allocated_bytes() decrease for some reason
|
| 173 |
+
# by adding it to the memo, self.table won't be copied
|
| 174 |
+
memo[id(self.table)] = self.table
|
| 175 |
+
# same for the recordbatches used by the index
|
| 176 |
+
memo[id(self._batches)] = list(self._batches)
|
| 177 |
+
return _deepcopy(self, memo)
|
| 178 |
+
|
| 179 |
+
def validate(self, *args, **kwargs):
|
| 180 |
+
"""
|
| 181 |
+
Perform validation checks. An exception is raised if validation fails.
|
| 182 |
+
|
| 183 |
+
By default only cheap validation checks are run. Pass `full=True`
|
| 184 |
+
for thorough validation checks (potentially `O(n)`).
|
| 185 |
+
|
| 186 |
+
Args:
|
| 187 |
+
full (`bool`, defaults to `False`):
|
| 188 |
+
If `True`, run expensive checks, otherwise cheap checks only.
|
| 189 |
+
|
| 190 |
+
Raises:
|
| 191 |
+
`pa.lib.ArrowInvalid`: if validation fails
|
| 192 |
+
"""
|
| 193 |
+
return self.table.validate(*args, **kwargs)
|
| 194 |
+
|
| 195 |
+
def equals(self, *args, **kwargs):
|
| 196 |
+
"""
|
| 197 |
+
Check if contents of two tables are equal.
|
| 198 |
+
|
| 199 |
+
Args:
|
| 200 |
+
other ([`~datasets.table.Table`]):
|
| 201 |
+
Table to compare against.
|
| 202 |
+
check_metadata `bool`, defaults to `False`):
|
| 203 |
+
Whether schema metadata equality should be checked as well.
|
| 204 |
+
|
| 205 |
+
Returns:
|
| 206 |
+
`bool`
|
| 207 |
+
"""
|
| 208 |
+
args = tuple(arg.table if isinstance(arg, Table) else arg for arg in args)
|
| 209 |
+
kwargs = {k: v.table if isinstance(v, Table) else v for k, v in kwargs}
|
| 210 |
+
return self.table.equals(*args, **kwargs)
|
| 211 |
+
|
| 212 |
+
def to_batches(self, *args, **kwargs):
|
| 213 |
+
"""
|
| 214 |
+
Convert Table to list of (contiguous) `RecordBatch` objects.
|
| 215 |
+
|
| 216 |
+
Args:
|
| 217 |
+
max_chunksize (`int`, defaults to `None`):
|
| 218 |
+
Maximum size for `RecordBatch` chunks. Individual chunks may be
|
| 219 |
+
smaller depending on the chunk layout of individual columns.
|
| 220 |
+
|
| 221 |
+
Returns:
|
| 222 |
+
`List[pyarrow.RecordBatch]`
|
| 223 |
+
"""
|
| 224 |
+
return self.table.to_batches(*args, **kwargs)
|
| 225 |
+
|
| 226 |
+
def to_pydict(self, *args, **kwargs):
|
| 227 |
+
"""
|
| 228 |
+
Convert the Table to a `dict` or `OrderedDict`.
|
| 229 |
+
|
| 230 |
+
Returns:
|
| 231 |
+
`dict`
|
| 232 |
+
"""
|
| 233 |
+
return self.table.to_pydict(*args, **kwargs)
|
| 234 |
+
|
| 235 |
+
def to_pylist(self, *args, **kwargs):
|
| 236 |
+
"""
|
| 237 |
+
Convert the Table to a list
|
| 238 |
+
|
| 239 |
+
Returns:
|
| 240 |
+
`list`
|
| 241 |
+
"""
|
| 242 |
+
return self.table.to_pylist(*args, **kwargs)
|
| 243 |
+
|
| 244 |
+
def to_pandas(self, *args, **kwargs):
|
| 245 |
+
"""
|
| 246 |
+
Convert to a pandas-compatible NumPy array or DataFrame, as appropriate.
|
| 247 |
+
|
| 248 |
+
Args:
|
| 249 |
+
memory_pool (`MemoryPool`, defaults to `None`):
|
| 250 |
+
Arrow MemoryPool to use for allocations. Uses the default memory
|
| 251 |
+
pool is not passed.
|
| 252 |
+
strings_to_categorical (`bool`, defaults to `False`):
|
| 253 |
+
Encode string (UTF8) and binary types to `pandas.Categorical`.
|
| 254 |
+
categories (`list`, defaults to `empty`):
|
| 255 |
+
List of fields that should be returned as `pandas.Categorical`. Only
|
| 256 |
+
applies to table-like data structures.
|
| 257 |
+
zero_copy_only (`bool`, defaults to `False`):
|
| 258 |
+
Raise an `ArrowException` if this function call would require copying
|
| 259 |
+
the underlying data.
|
| 260 |
+
integer_object_nulls (`bool`, defaults to `False`):
|
| 261 |
+
Cast integers with nulls to objects.
|
| 262 |
+
date_as_object (`bool`, defaults to `True`):
|
| 263 |
+
Cast dates to objects. If `False`, convert to `datetime64[ns]` dtype.
|
| 264 |
+
timestamp_as_object (`bool`, defaults to `False`):
|
| 265 |
+
Cast non-nanosecond timestamps (`np.datetime64`) to objects. This is
|
| 266 |
+
useful if you have timestamps that don't fit in the normal date
|
| 267 |
+
range of nanosecond timestamps (1678 CE-2262 CE).
|
| 268 |
+
If `False`, all timestamps are converted to `datetime64[ns]` dtype.
|
| 269 |
+
use_threads (`bool`, defaults to `True`):
|
| 270 |
+
Whether to parallelize the conversion using multiple threads.
|
| 271 |
+
deduplicate_objects (`bool`, defaults to `False`):
|
| 272 |
+
Do not create multiple copies Python objects when created, to save
|
| 273 |
+
on memory use. Conversion will be slower.
|
| 274 |
+
ignore_metadata (`bool`, defaults to `False`):
|
| 275 |
+
If `True`, do not use the 'pandas' metadata to reconstruct the
|
| 276 |
+
DataFrame index, if present.
|
| 277 |
+
safe (`bool`, defaults to `True`):
|
| 278 |
+
For certain data types, a cast is needed in order to store the
|
| 279 |
+
data in a pandas DataFrame or Series (e.g. timestamps are always
|
| 280 |
+
stored as nanoseconds in pandas). This option controls whether it
|
| 281 |
+
is a safe cast or not.
|
| 282 |
+
split_blocks (`bool`, defaults to `False`):
|
| 283 |
+
If `True`, generate one internal "block" for each column when
|
| 284 |
+
creating a pandas.DataFrame from a `RecordBatch` or `Table`. While this
|
| 285 |
+
can temporarily reduce memory note that various pandas operations
|
| 286 |
+
can trigger "consolidation" which may balloon memory use.
|
| 287 |
+
self_destruct (`bool`, defaults to `False`):
|
| 288 |
+
EXPERIMENTAL: If `True`, attempt to deallocate the originating Arrow
|
| 289 |
+
memory while converting the Arrow object to pandas. If you use the
|
| 290 |
+
object after calling `to_pandas` with this option it will crash your
|
| 291 |
+
program.
|
| 292 |
+
types_mapper (`function`, defaults to `None`):
|
| 293 |
+
A function mapping a pyarrow DataType to a pandas `ExtensionDtype`.
|
| 294 |
+
This can be used to override the default pandas type for conversion
|
| 295 |
+
of built-in pyarrow types or in absence of `pandas_metadata` in the
|
| 296 |
+
Table schema. The function receives a pyarrow DataType and is
|
| 297 |
+
expected to return a pandas `ExtensionDtype` or `None` if the
|
| 298 |
+
default conversion should be used for that type. If you have
|
| 299 |
+
a dictionary mapping, you can pass `dict.get` as function.
|
| 300 |
+
|
| 301 |
+
Returns:
|
| 302 |
+
`pandas.Series` or `pandas.DataFrame`: `pandas.Series` or `pandas.DataFrame` depending on type of object
|
| 303 |
+
"""
|
| 304 |
+
return self.table.to_pandas(*args, **kwargs)
|
| 305 |
+
|
| 306 |
+
def to_string(self, *args, **kwargs):
|
| 307 |
+
return self.table.to_string(*args, **kwargs)
|
| 308 |
+
|
| 309 |
+
def to_reader(self, max_chunksize: Optional[int] = None):
|
| 310 |
+
"""
|
| 311 |
+
Convert the Table to a RecordBatchReader.
|
| 312 |
+
|
| 313 |
+
Note that this method is zero-copy, it merely exposes the same data under a different API.
|
| 314 |
+
|
| 315 |
+
Args:
|
| 316 |
+
max_chunksize (`int`, defaults to `None`)
|
| 317 |
+
Maximum size for RecordBatch chunks. Individual chunks may be smaller depending
|
| 318 |
+
on the chunk layout of individual columns.
|
| 319 |
+
|
| 320 |
+
Returns:
|
| 321 |
+
`pyarrow.RecordBatchReader`
|
| 322 |
+
"""
|
| 323 |
+
return self.table.to_reader(max_chunksize=max_chunksize)
|
| 324 |
+
|
| 325 |
+
def field(self, *args, **kwargs):
|
| 326 |
+
"""
|
| 327 |
+
Select a schema field by its column name or numeric index.
|
| 328 |
+
|
| 329 |
+
Args:
|
| 330 |
+
i (`Union[int, str]`):
|
| 331 |
+
The index or name of the field to retrieve.
|
| 332 |
+
|
| 333 |
+
Returns:
|
| 334 |
+
`pyarrow.Field`
|
| 335 |
+
"""
|
| 336 |
+
return self.table.field(*args, **kwargs)
|
| 337 |
+
|
| 338 |
+
def column(self, *args, **kwargs):
|
| 339 |
+
"""
|
| 340 |
+
Select a column by its column name, or numeric index.
|
| 341 |
+
|
| 342 |
+
Args:
|
| 343 |
+
i (`Union[int, str]`):
|
| 344 |
+
The index or name of the column to retrieve.
|
| 345 |
+
|
| 346 |
+
Returns:
|
| 347 |
+
`pyarrow.ChunkedArray`
|
| 348 |
+
"""
|
| 349 |
+
return self.table.column(*args, **kwargs)
|
| 350 |
+
|
| 351 |
+
def itercolumns(self, *args, **kwargs):
|
| 352 |
+
"""
|
| 353 |
+
Iterator over all columns in their numerical order.
|
| 354 |
+
|
| 355 |
+
Yields:
|
| 356 |
+
`pyarrow.ChunkedArray`
|
| 357 |
+
"""
|
| 358 |
+
return self.table.itercolumns(*args, **kwargs)
|
| 359 |
+
|
| 360 |
+
@property
|
| 361 |
+
def schema(self):
|
| 362 |
+
"""
|
| 363 |
+
Schema of the table and its columns.
|
| 364 |
+
|
| 365 |
+
Returns:
|
| 366 |
+
`pyarrow.Schema`
|
| 367 |
+
"""
|
| 368 |
+
return self.table.schema
|
| 369 |
+
|
| 370 |
+
@property
|
| 371 |
+
def columns(self):
|
| 372 |
+
"""
|
| 373 |
+
List of all columns in numerical order.
|
| 374 |
+
|
| 375 |
+
Returns:
|
| 376 |
+
`List[pa.ChunkedArray]`
|
| 377 |
+
"""
|
| 378 |
+
return self.table.columns
|
| 379 |
+
|
| 380 |
+
@property
|
| 381 |
+
def num_columns(self):
|
| 382 |
+
"""
|
| 383 |
+
Number of columns in this table.
|
| 384 |
+
|
| 385 |
+
Returns:
|
| 386 |
+
int
|
| 387 |
+
"""
|
| 388 |
+
return self.table.num_columns
|
| 389 |
+
|
| 390 |
+
@property
|
| 391 |
+
def num_rows(self):
|
| 392 |
+
"""
|
| 393 |
+
Number of rows in this table.
|
| 394 |
+
|
| 395 |
+
Due to the definition of a table, all columns have the same number of
|
| 396 |
+
rows.
|
| 397 |
+
|
| 398 |
+
Returns:
|
| 399 |
+
int
|
| 400 |
+
"""
|
| 401 |
+
return self.table.num_rows
|
| 402 |
+
|
| 403 |
+
@property
|
| 404 |
+
def shape(self):
|
| 405 |
+
"""
|
| 406 |
+
Dimensions of the table: (#rows, #columns).
|
| 407 |
+
|
| 408 |
+
Returns:
|
| 409 |
+
`(int, int)`: Number of rows and number of columns.
|
| 410 |
+
"""
|
| 411 |
+
return self.table.shape
|
| 412 |
+
|
| 413 |
+
@property
|
| 414 |
+
def nbytes(self):
|
| 415 |
+
"""
|
| 416 |
+
Total number of bytes consumed by the elements of the table.
|
| 417 |
+
"""
|
| 418 |
+
return self.table.nbytes
|
| 419 |
+
|
| 420 |
+
@property
|
| 421 |
+
def column_names(self):
|
| 422 |
+
"""
|
| 423 |
+
Names of the table's columns.
|
| 424 |
+
"""
|
| 425 |
+
return self.table.column_names
|
| 426 |
+
|
| 427 |
+
def __eq__(self, other):
|
| 428 |
+
return self.equals(other)
|
| 429 |
+
|
| 430 |
+
def __getitem__(self, i):
|
| 431 |
+
return self.table[i]
|
| 432 |
+
|
| 433 |
+
def __len__(self):
|
| 434 |
+
return len(self.table)
|
| 435 |
+
|
| 436 |
+
def __repr__(self):
|
| 437 |
+
return self.table.__repr__().replace("pyarrow.Table", self.__class__.__name__)
|
| 438 |
+
|
| 439 |
+
def __str__(self):
|
| 440 |
+
return self.table.__str__().replace("pyarrow.Table", self.__class__.__name__)
|
| 441 |
+
|
| 442 |
+
def slice(self, *args, **kwargs):
|
| 443 |
+
"""
|
| 444 |
+
Compute zero-copy slice of this Table.
|
| 445 |
+
|
| 446 |
+
Args:
|
| 447 |
+
offset (`int`, defaults to `0`):
|
| 448 |
+
Offset from start of table to slice.
|
| 449 |
+
length (`int`, defaults to `None`):
|
| 450 |
+
Length of slice (default is until end of table starting from
|
| 451 |
+
offset).
|
| 452 |
+
|
| 453 |
+
Returns:
|
| 454 |
+
`datasets.table.Table`
|
| 455 |
+
"""
|
| 456 |
+
raise NotImplementedError()
|
| 457 |
+
|
| 458 |
+
def filter(self, *args, **kwargs):
|
| 459 |
+
"""
|
| 460 |
+
Select records from a Table. See `pyarrow.compute.filter` for full usage.
|
| 461 |
+
"""
|
| 462 |
+
raise NotImplementedError()
|
| 463 |
+
|
| 464 |
+
def flatten(self, *args, **kwargs):
|
| 465 |
+
"""
|
| 466 |
+
Flatten this Table. Each column with a struct type is flattened
|
| 467 |
+
into one column per struct field. Other columns are left unchanged.
|
| 468 |
+
|
| 469 |
+
Args:
|
| 470 |
+
memory_pool (`MemoryPool`, defaults to `None`):
|
| 471 |
+
For memory allocations, if required, otherwise use default pool.
|
| 472 |
+
|
| 473 |
+
Returns:
|
| 474 |
+
`datasets.table.Table`
|
| 475 |
+
"""
|
| 476 |
+
raise NotImplementedError()
|
| 477 |
+
|
| 478 |
+
def combine_chunks(self, *args, **kwargs):
|
| 479 |
+
"""
|
| 480 |
+
Make a new table by combining the chunks this table has.
|
| 481 |
+
|
| 482 |
+
All the underlying chunks in the `ChunkedArray` of each column are
|
| 483 |
+
concatenated into zero or one chunk.
|
| 484 |
+
|
| 485 |
+
Args:
|
| 486 |
+
memory_pool (`MemoryPool`, defaults to `None`):
|
| 487 |
+
For memory allocations, if required, otherwise use default pool.
|
| 488 |
+
|
| 489 |
+
Returns:
|
| 490 |
+
`datasets.table.Table`
|
| 491 |
+
"""
|
| 492 |
+
raise NotImplementedError()
|
| 493 |
+
|
| 494 |
+
def cast(self, *args, **kwargs):
|
| 495 |
+
"""
|
| 496 |
+
Cast table values to another schema.
|
| 497 |
+
|
| 498 |
+
Args:
|
| 499 |
+
target_schema (`Schema`):
|
| 500 |
+
Schema to cast to, the names and order of fields must match.
|
| 501 |
+
safe (`bool`, defaults to `True`):
|
| 502 |
+
Check for overflows or other unsafe conversions.
|
| 503 |
+
|
| 504 |
+
Returns:
|
| 505 |
+
`datasets.table.Table`
|
| 506 |
+
"""
|
| 507 |
+
raise NotImplementedError()
|
| 508 |
+
|
| 509 |
+
def replace_schema_metadata(self, *args, **kwargs):
|
| 510 |
+
"""
|
| 511 |
+
EXPERIMENTAL: Create shallow copy of table by replacing schema
|
| 512 |
+
key-value metadata with the indicated new metadata (which may be None,
|
| 513 |
+
which deletes any existing metadata
|
| 514 |
+
|
| 515 |
+
Args:
|
| 516 |
+
metadata (`dict`, defaults to `None`):
|
| 517 |
+
|
| 518 |
+
Returns:
|
| 519 |
+
`datasets.table.Table`: shallow_copy
|
| 520 |
+
"""
|
| 521 |
+
raise NotImplementedError()
|
| 522 |
+
|
| 523 |
+
def add_column(self, *args, **kwargs):
|
| 524 |
+
"""
|
| 525 |
+
Add column to Table at position.
|
| 526 |
+
|
| 527 |
+
A new table is returned with the column added, the original table
|
| 528 |
+
object is left unchanged.
|
| 529 |
+
|
| 530 |
+
Args:
|
| 531 |
+
i (`int`):
|
| 532 |
+
Index to place the column at.
|
| 533 |
+
field_ (`Union[str, pyarrow.Field]`):
|
| 534 |
+
If a string is passed then the type is deduced from the column
|
| 535 |
+
data.
|
| 536 |
+
column (`Union[pyarrow.Array, List[pyarrow.Array]]`):
|
| 537 |
+
Column data.
|
| 538 |
+
|
| 539 |
+
Returns:
|
| 540 |
+
`datasets.table.Table`: New table with the passed column added.
|
| 541 |
+
"""
|
| 542 |
+
raise NotImplementedError()
|
| 543 |
+
|
| 544 |
+
def append_column(self, *args, **kwargs):
|
| 545 |
+
"""
|
| 546 |
+
Append column at end of columns.
|
| 547 |
+
|
| 548 |
+
Args:
|
| 549 |
+
field_ (`Union[str, pyarrow.Field]`):
|
| 550 |
+
If a string is passed then the type is deduced from the column
|
| 551 |
+
data.
|
| 552 |
+
column (`Union[pyarrow.Array, List[pyarrow.Array]]`):
|
| 553 |
+
Column data.
|
| 554 |
+
|
| 555 |
+
Returns:
|
| 556 |
+
`datasets.table.Table`: New table with the passed column added.
|
| 557 |
+
"""
|
| 558 |
+
raise NotImplementedError()
|
| 559 |
+
|
| 560 |
+
def remove_column(self, *args, **kwargs):
|
| 561 |
+
"""
|
| 562 |
+
Create new Table with the indicated column removed.
|
| 563 |
+
|
| 564 |
+
Args:
|
| 565 |
+
i (`int`):
|
| 566 |
+
Index of column to remove.
|
| 567 |
+
|
| 568 |
+
Returns:
|
| 569 |
+
`datasets.table.Table`: New table without the column.
|
| 570 |
+
"""
|
| 571 |
+
raise NotImplementedError()
|
| 572 |
+
|
| 573 |
+
def set_column(self, *args, **kwargs):
|
| 574 |
+
"""
|
| 575 |
+
Replace column in Table at position.
|
| 576 |
+
|
| 577 |
+
Args:
|
| 578 |
+
i (`int`):
|
| 579 |
+
Index to place the column at.
|
| 580 |
+
field_ (`Union[str, pyarrow.Field]`):
|
| 581 |
+
If a string is passed then the type is deduced from the column
|
| 582 |
+
data.
|
| 583 |
+
column (`Union[pyarrow.Array, List[pyarrow.Array]]`):
|
| 584 |
+
Column data.
|
| 585 |
+
|
| 586 |
+
Returns:
|
| 587 |
+
`datasets.table.Table`: New table with the passed column set.
|
| 588 |
+
"""
|
| 589 |
+
raise NotImplementedError()
|
| 590 |
+
|
| 591 |
+
def rename_columns(self, *args, **kwargs):
|
| 592 |
+
"""
|
| 593 |
+
Create new table with columns renamed to provided names.
|
| 594 |
+
"""
|
| 595 |
+
raise NotImplementedError()
|
| 596 |
+
|
| 597 |
+
def drop(self, *args, **kwargs):
|
| 598 |
+
"""
|
| 599 |
+
Drop one or more columns and return a new table.
|
| 600 |
+
|
| 601 |
+
Args:
|
| 602 |
+
columns (`List[str]`):
|
| 603 |
+
List of field names referencing existing columns.
|
| 604 |
+
|
| 605 |
+
Raises:
|
| 606 |
+
`KeyError` : if any of the passed columns name are not existing.
|
| 607 |
+
|
| 608 |
+
Returns:
|
| 609 |
+
`datasets.table.Table`: New table without the columns.
|
| 610 |
+
"""
|
| 611 |
+
raise NotImplementedError()
|
| 612 |
+
|
| 613 |
+
def select(self, *args, **kwargs):
|
| 614 |
+
"""
|
| 615 |
+
Select columns of the table.
|
| 616 |
+
|
| 617 |
+
Returns a new table with the specified columns, and metadata preserved.
|
| 618 |
+
|
| 619 |
+
Args:
|
| 620 |
+
columns (:obj:`Union[List[str], List[int]]`):
|
| 621 |
+
The column names or integer indices to select.
|
| 622 |
+
|
| 623 |
+
Returns:
|
| 624 |
+
`datasets.table.Table`: table with only a subset of the columns
|
| 625 |
+
"""
|
| 626 |
+
raise NotImplementedError()
|
| 627 |
+
|
| 628 |
+
|
| 629 |
+
class TableBlock(Table):
|
| 630 |
+
"""
|
| 631 |
+
`TableBlock` is the allowed class inside a `ConcanetationTable`.
|
| 632 |
+
Only `MemoryMappedTable` and `InMemoryTable` are `TableBlock`.
|
| 633 |
+
This is because we don't want a `ConcanetationTable` made out of other `ConcanetationTables`.
|
| 634 |
+
"""
|
| 635 |
+
|
| 636 |
+
pass
|
| 637 |
+
|
| 638 |
+
|
| 639 |
+
class InMemoryTable(TableBlock):
|
| 640 |
+
"""
|
| 641 |
+
The table is said in-memory when it is loaded into the user's RAM.
|
| 642 |
+
|
| 643 |
+
Pickling it does copy all the data using memory.
|
| 644 |
+
Its implementation is simple and uses the underlying pyarrow Table methods directly.
|
| 645 |
+
|
| 646 |
+
This is different from the `MemoryMapped` table, for which pickling doesn't copy all the
|
| 647 |
+
data in memory. For a `MemoryMapped`, unpickling instead reloads the table from the disk.
|
| 648 |
+
|
| 649 |
+
`InMemoryTable` must be used when data fit in memory, while `MemoryMapped` are reserved for
|
| 650 |
+
data bigger than memory or when you want the memory footprint of your application to
|
| 651 |
+
stay low.
|
| 652 |
+
"""
|
| 653 |
+
|
| 654 |
+
@classmethod
|
| 655 |
+
def from_file(cls, filename: str):
|
| 656 |
+
table = _in_memory_arrow_table_from_file(filename)
|
| 657 |
+
return cls(table)
|
| 658 |
+
|
| 659 |
+
@classmethod
|
| 660 |
+
def from_buffer(cls, buffer: pa.Buffer):
|
| 661 |
+
table = _in_memory_arrow_table_from_buffer(buffer)
|
| 662 |
+
return cls(table)
|
| 663 |
+
|
| 664 |
+
@classmethod
|
| 665 |
+
def from_pandas(cls, *args, **kwargs):
|
| 666 |
+
"""
|
| 667 |
+
Convert pandas.DataFrame to an Arrow Table.
|
| 668 |
+
|
| 669 |
+
The column types in the resulting Arrow Table are inferred from the
|
| 670 |
+
dtypes of the pandas.Series in the DataFrame. In the case of non-object
|
| 671 |
+
Series, the NumPy dtype is translated to its Arrow equivalent. In the
|
| 672 |
+
case of `object`, we need to guess the datatype by looking at the
|
| 673 |
+
Python objects in this Series.
|
| 674 |
+
|
| 675 |
+
Be aware that Series of the `object` dtype don't carry enough
|
| 676 |
+
information to always lead to a meaningful Arrow type. In the case that
|
| 677 |
+
we cannot infer a type, e.g. because the DataFrame is of length 0 or
|
| 678 |
+
the Series only contains `None/nan` objects, the type is set to
|
| 679 |
+
null. This behavior can be avoided by constructing an explicit schema
|
| 680 |
+
and passing it to this function.
|
| 681 |
+
|
| 682 |
+
Args:
|
| 683 |
+
df (`pandas.DataFrame`):
|
| 684 |
+
schema (`pyarrow.Schema`, *optional*):
|
| 685 |
+
The expected schema of the Arrow Table. This can be used to
|
| 686 |
+
indicate the type of columns if we cannot infer it automatically.
|
| 687 |
+
If passed, the output will have exactly this schema. Columns
|
| 688 |
+
specified in the schema that are not found in the DataFrame columns
|
| 689 |
+
or its index will raise an error. Additional columns or index
|
| 690 |
+
levels in the DataFrame which are not specified in the schema will
|
| 691 |
+
be ignored.
|
| 692 |
+
preserve_index (`bool`, *optional*):
|
| 693 |
+
Whether to store the index as an additional column in the resulting
|
| 694 |
+
`Table`. The default of None will store the index as a column,
|
| 695 |
+
except for RangeIndex which is stored as metadata only. Use
|
| 696 |
+
`preserve_index=True` to force it to be stored as a column.
|
| 697 |
+
nthreads (`int`, defaults to `None` (may use up to system CPU count threads))
|
| 698 |
+
If greater than 1, convert columns to Arrow in parallel using
|
| 699 |
+
indicated number of threads.
|
| 700 |
+
columns (`List[str]`, *optional*):
|
| 701 |
+
List of column to be converted. If `None`, use all columns.
|
| 702 |
+
safe (`bool`, defaults to `True`):
|
| 703 |
+
Check for overflows or other unsafe conversions,
|
| 704 |
+
|
| 705 |
+
Returns:
|
| 706 |
+
`datasets.table.Table`:
|
| 707 |
+
|
| 708 |
+
Examples:
|
| 709 |
+
```python
|
| 710 |
+
>>> import pandas as pd
|
| 711 |
+
>>> import pyarrow as pa
|
| 712 |
+
>>> df = pd.DataFrame({
|
| 713 |
+
... 'int': [1, 2],
|
| 714 |
+
... 'str': ['a', 'b']
|
| 715 |
+
... })
|
| 716 |
+
>>> pa.Table.from_pandas(df)
|
| 717 |
+
<pyarrow.lib.Table object at 0x7f05d1fb1b40>
|
| 718 |
+
```
|
| 719 |
+
"""
|
| 720 |
+
return cls(pa.Table.from_pandas(*args, **kwargs))
|
| 721 |
+
|
| 722 |
+
@classmethod
|
| 723 |
+
def from_arrays(cls, *args, **kwargs):
|
| 724 |
+
"""
|
| 725 |
+
Construct a Table from Arrow arrays.
|
| 726 |
+
|
| 727 |
+
Args:
|
| 728 |
+
arrays (`List[Union[pyarrow.Array, pyarrow.ChunkedArray]]`):
|
| 729 |
+
Equal-length arrays that should form the table.
|
| 730 |
+
names (`List[str]`, *optional*):
|
| 731 |
+
Names for the table columns. If not passed, schema must be passed.
|
| 732 |
+
schema (`Schema`, defaults to `None`):
|
| 733 |
+
Schema for the created table. If not passed, names must be passed.
|
| 734 |
+
metadata (`Union[dict, Mapping]`, defaults to `None`):
|
| 735 |
+
Optional metadata for the schema (if inferred).
|
| 736 |
+
|
| 737 |
+
Returns:
|
| 738 |
+
`datasets.table.Table`
|
| 739 |
+
"""
|
| 740 |
+
return cls(pa.Table.from_arrays(*args, **kwargs))
|
| 741 |
+
|
| 742 |
+
@classmethod
|
| 743 |
+
def from_pydict(cls, *args, **kwargs):
|
| 744 |
+
"""
|
| 745 |
+
Construct a Table from Arrow arrays or columns.
|
| 746 |
+
|
| 747 |
+
Args:
|
| 748 |
+
mapping (`Union[dict, Mapping]`):
|
| 749 |
+
A mapping of strings to Arrays or Python lists.
|
| 750 |
+
schema (`Schema`, defaults to `None`):
|
| 751 |
+
If not passed, will be inferred from the Mapping values
|
| 752 |
+
metadata (`Union[dict, Mapping]`, defaults to `None`):
|
| 753 |
+
Optional metadata for the schema (if inferred).
|
| 754 |
+
|
| 755 |
+
Returns:
|
| 756 |
+
`datasets.table.Table`
|
| 757 |
+
"""
|
| 758 |
+
return cls(pa.Table.from_pydict(*args, **kwargs))
|
| 759 |
+
|
| 760 |
+
@classmethod
|
| 761 |
+
def from_pylist(cls, mapping, *args, **kwargs):
|
| 762 |
+
"""
|
| 763 |
+
Construct a Table from list of rows / dictionaries.
|
| 764 |
+
|
| 765 |
+
Args:
|
| 766 |
+
mapping (`List[dict]`):
|
| 767 |
+
A mapping of strings to row values.
|
| 768 |
+
schema (`Schema`, defaults to `None`):
|
| 769 |
+
If not passed, will be inferred from the Mapping values
|
| 770 |
+
metadata (`Union[dict, Mapping]`, defaults to `None`):
|
| 771 |
+
Optional metadata for the schema (if inferred).
|
| 772 |
+
|
| 773 |
+
Returns:
|
| 774 |
+
`datasets.table.Table`
|
| 775 |
+
"""
|
| 776 |
+
return cls(pa.Table.from_pylist(mapping, *args, **kwargs))
|
| 777 |
+
|
| 778 |
+
@classmethod
|
| 779 |
+
def from_batches(cls, *args, **kwargs):
|
| 780 |
+
"""
|
| 781 |
+
Construct a Table from a sequence or iterator of Arrow `RecordBatches`.
|
| 782 |
+
|
| 783 |
+
Args:
|
| 784 |
+
batches (`Union[Sequence[pyarrow.RecordBatch], Iterator[pyarrow.RecordBatch]]`):
|
| 785 |
+
Sequence of `RecordBatch` to be converted, all schemas must be equal.
|
| 786 |
+
schema (`Schema`, defaults to `None`):
|
| 787 |
+
If not passed, will be inferred from the first `RecordBatch`.
|
| 788 |
+
|
| 789 |
+
Returns:
|
| 790 |
+
`datasets.table.Table`:
|
| 791 |
+
"""
|
| 792 |
+
return cls(pa.Table.from_batches(*args, **kwargs))
|
| 793 |
+
|
| 794 |
+
def slice(self, offset=0, length=None):
|
| 795 |
+
"""
|
| 796 |
+
Compute zero-copy slice of this Table.
|
| 797 |
+
|
| 798 |
+
Args:
|
| 799 |
+
offset (`int`, defaults to `0`):
|
| 800 |
+
Offset from start of table to slice.
|
| 801 |
+
length (`int`, defaults to `None`):
|
| 802 |
+
Length of slice (default is until end of table starting from
|
| 803 |
+
offset).
|
| 804 |
+
|
| 805 |
+
Returns:
|
| 806 |
+
`datasets.table.Table`
|
| 807 |
+
"""
|
| 808 |
+
# Use fast slicing here
|
| 809 |
+
return InMemoryTable(self.fast_slice(offset=offset, length=length))
|
| 810 |
+
|
| 811 |
+
def filter(self, *args, **kwargs):
|
| 812 |
+
"""
|
| 813 |
+
Select records from a Table. See `pyarrow.compute.filter` for full usage.
|
| 814 |
+
"""
|
| 815 |
+
return InMemoryTable(self.table.filter(*args, **kwargs))
|
| 816 |
+
|
| 817 |
+
def flatten(self, *args, **kwargs):
|
| 818 |
+
"""
|
| 819 |
+
Flatten this Table. Each column with a struct type is flattened
|
| 820 |
+
into one column per struct field. Other columns are left unchanged.
|
| 821 |
+
|
| 822 |
+
Args:
|
| 823 |
+
memory_pool (`MemoryPool`, defaults to `None`):
|
| 824 |
+
For memory allocations, if required, otherwise use default pool.
|
| 825 |
+
|
| 826 |
+
Returns:
|
| 827 |
+
`datasets.table.Table`
|
| 828 |
+
"""
|
| 829 |
+
return InMemoryTable(table_flatten(self.table, *args, **kwargs))
|
| 830 |
+
|
| 831 |
+
def combine_chunks(self, *args, **kwargs):
|
| 832 |
+
"""
|
| 833 |
+
Make a new table by combining the chunks this table has.
|
| 834 |
+
|
| 835 |
+
All the underlying chunks in the `ChunkedArray` of each column are
|
| 836 |
+
concatenated into zero or one chunk.
|
| 837 |
+
|
| 838 |
+
Args:
|
| 839 |
+
memory_pool (`MemoryPool`, defaults to `None`):
|
| 840 |
+
For memory allocations, if required, otherwise use default pool.
|
| 841 |
+
|
| 842 |
+
Returns:
|
| 843 |
+
`datasets.table.Table`
|
| 844 |
+
"""
|
| 845 |
+
return InMemoryTable(self.table.combine_chunks(*args, **kwargs))
|
| 846 |
+
|
| 847 |
+
def cast(self, *args, **kwargs):
|
| 848 |
+
"""
|
| 849 |
+
Cast table values to another schema.
|
| 850 |
+
|
| 851 |
+
Args:
|
| 852 |
+
target_schema (`Schema`):
|
| 853 |
+
Schema to cast to, the names and order of fields must match.
|
| 854 |
+
safe (`bool`, defaults to `True`):
|
| 855 |
+
Check for overflows or other unsafe conversions.
|
| 856 |
+
|
| 857 |
+
Returns:
|
| 858 |
+
`datasets.table.Table`
|
| 859 |
+
"""
|
| 860 |
+
return InMemoryTable(table_cast(self.table, *args, **kwargs))
|
| 861 |
+
|
| 862 |
+
def replace_schema_metadata(self, *args, **kwargs):
|
| 863 |
+
"""
|
| 864 |
+
EXPERIMENTAL: Create shallow copy of table by replacing schema
|
| 865 |
+
key-value metadata with the indicated new metadata (which may be `None`,
|
| 866 |
+
which deletes any existing metadata).
|
| 867 |
+
|
| 868 |
+
Args:
|
| 869 |
+
metadata (`dict`, defaults to `None`):
|
| 870 |
+
|
| 871 |
+
Returns:
|
| 872 |
+
`datasets.table.Table`: shallow_copy
|
| 873 |
+
"""
|
| 874 |
+
return InMemoryTable(self.table.replace_schema_metadata(*args, **kwargs))
|
| 875 |
+
|
| 876 |
+
def add_column(self, *args, **kwargs):
|
| 877 |
+
"""
|
| 878 |
+
Add column to Table at position.
|
| 879 |
+
|
| 880 |
+
A new table is returned with the column added, the original table
|
| 881 |
+
object is left unchanged.
|
| 882 |
+
|
| 883 |
+
Args:
|
| 884 |
+
i (`int`):
|
| 885 |
+
Index to place the column at.
|
| 886 |
+
field_ (`Union[str, pyarrow.Field]`):
|
| 887 |
+
If a string is passed then the type is deduced from the column
|
| 888 |
+
data.
|
| 889 |
+
column (`Union[pyarrow.Array, List[pyarrow.Array]]`):
|
| 890 |
+
Column data.
|
| 891 |
+
|
| 892 |
+
Returns:
|
| 893 |
+
`datasets.table.Table`: New table with the passed column added.
|
| 894 |
+
"""
|
| 895 |
+
return InMemoryTable(self.table.add_column(*args, **kwargs))
|
| 896 |
+
|
| 897 |
+
def append_column(self, *args, **kwargs):
|
| 898 |
+
"""
|
| 899 |
+
Append column at end of columns.
|
| 900 |
+
|
| 901 |
+
Args:
|
| 902 |
+
field_ (`Union[str, pyarrow.Field]`):
|
| 903 |
+
If a string is passed then the type is deduced from the column
|
| 904 |
+
data.
|
| 905 |
+
column (`Union[pyarrow.Array, List[pyarrow.Array]]`):
|
| 906 |
+
Column data.
|
| 907 |
+
|
| 908 |
+
Returns:
|
| 909 |
+
`datasets.table.Table`:
|
| 910 |
+
New table with the passed column added.
|
| 911 |
+
"""
|
| 912 |
+
return InMemoryTable(self.table.append_column(*args, **kwargs))
|
| 913 |
+
|
| 914 |
+
def remove_column(self, *args, **kwargs):
|
| 915 |
+
"""
|
| 916 |
+
Create new Table with the indicated column removed.
|
| 917 |
+
|
| 918 |
+
Args:
|
| 919 |
+
i (`int`):
|
| 920 |
+
Index of column to remove.
|
| 921 |
+
|
| 922 |
+
Returns:
|
| 923 |
+
`datasets.table.Table`:
|
| 924 |
+
New table without the column.
|
| 925 |
+
"""
|
| 926 |
+
return InMemoryTable(self.table.remove_column(*args, **kwargs))
|
| 927 |
+
|
| 928 |
+
def set_column(self, *args, **kwargs):
|
| 929 |
+
"""
|
| 930 |
+
Replace column in Table at position.
|
| 931 |
+
|
| 932 |
+
Args:
|
| 933 |
+
i (`int`):
|
| 934 |
+
Index to place the column at.
|
| 935 |
+
field_ (`Union[str, pyarrow.Field]`):
|
| 936 |
+
If a string is passed then the type is deduced from the column
|
| 937 |
+
data.
|
| 938 |
+
column (`Union[pyarrow.Array, List[pyarrow.Array]]`):
|
| 939 |
+
Column data.
|
| 940 |
+
|
| 941 |
+
Returns:
|
| 942 |
+
`datasets.table.Table`:
|
| 943 |
+
New table with the passed column set.
|
| 944 |
+
"""
|
| 945 |
+
return InMemoryTable(self.table.set_column(*args, **kwargs))
|
| 946 |
+
|
| 947 |
+
def rename_columns(self, *args, **kwargs):
|
| 948 |
+
"""
|
| 949 |
+
Create new table with columns renamed to provided names.
|
| 950 |
+
"""
|
| 951 |
+
return InMemoryTable(self.table.rename_columns(*args, **kwargs))
|
| 952 |
+
|
| 953 |
+
def drop(self, *args, **kwargs):
|
| 954 |
+
"""
|
| 955 |
+
Drop one or more columns and return a new table.
|
| 956 |
+
|
| 957 |
+
Args:
|
| 958 |
+
columns (`List[str]`):
|
| 959 |
+
List of field names referencing existing columns.
|
| 960 |
+
|
| 961 |
+
Raises:
|
| 962 |
+
`KeyError` : if any of the passed columns name are not existing.
|
| 963 |
+
|
| 964 |
+
Returns:
|
| 965 |
+
`datasets.table.Table`:
|
| 966 |
+
New table without the columns.
|
| 967 |
+
"""
|
| 968 |
+
return InMemoryTable(self.table.drop(*args, **kwargs))
|
| 969 |
+
|
| 970 |
+
def select(self, *args, **kwargs):
|
| 971 |
+
"""
|
| 972 |
+
Select columns of the table.
|
| 973 |
+
|
| 974 |
+
Returns a new table with the specified columns, and metadata preserved.
|
| 975 |
+
|
| 976 |
+
Args:
|
| 977 |
+
columns (:obj:`Union[List[str], List[int]]`):
|
| 978 |
+
The column names or integer indices to select.
|
| 979 |
+
|
| 980 |
+
Returns:
|
| 981 |
+
:class:`datasets.table.Table`: New table with the specified columns, and metadata preserved.
|
| 982 |
+
"""
|
| 983 |
+
return InMemoryTable(self.table.select(*args, **kwargs))
|
| 984 |
+
|
| 985 |
+
|
| 986 |
+
# The MemoryMappedTable needs replays to properly reload tables from the disk
|
| 987 |
+
Replay = Tuple[str, tuple, dict]
|
| 988 |
+
|
| 989 |
+
|
| 990 |
+
class MemoryMappedTable(TableBlock):
|
| 991 |
+
"""
|
| 992 |
+
The table is said memory mapped when it doesn't use the user's RAM but loads the data
|
| 993 |
+
from the disk instead.
|
| 994 |
+
|
| 995 |
+
Pickling it doesn't copy the data into memory.
|
| 996 |
+
Instead, only the path to the memory mapped arrow file is pickled, as well as the list
|
| 997 |
+
of transforms to "replay" when reloading the table from the disk.
|
| 998 |
+
|
| 999 |
+
Its implementation requires to store an history of all the transforms that were applied
|
| 1000 |
+
to the underlying pyarrow Table, so that they can be "replayed" when reloading the Table
|
| 1001 |
+
from the disk.
|
| 1002 |
+
|
| 1003 |
+
This is different from the `InMemoryTable` table, for which pickling does copy all the
|
| 1004 |
+
data in memory.
|
| 1005 |
+
|
| 1006 |
+
`InMemoryTable` must be used when data fit in memory, while `MemoryMapped` are reserved for
|
| 1007 |
+
data bigger than memory or when you want the memory footprint of your application to
|
| 1008 |
+
stay low.
|
| 1009 |
+
"""
|
| 1010 |
+
|
| 1011 |
+
def __init__(self, table: pa.Table, path: str, replays: Optional[List[Replay]] = None):
|
| 1012 |
+
super().__init__(table)
|
| 1013 |
+
self.path = os.path.abspath(path)
|
| 1014 |
+
self.replays: List[Replay] = replays if replays is not None else []
|
| 1015 |
+
|
| 1016 |
+
@classmethod
|
| 1017 |
+
def from_file(cls, filename: str, replays=None):
|
| 1018 |
+
table = _memory_mapped_arrow_table_from_file(filename)
|
| 1019 |
+
table = cls._apply_replays(table, replays)
|
| 1020 |
+
return cls(table, filename, replays)
|
| 1021 |
+
|
| 1022 |
+
def __getstate__(self):
|
| 1023 |
+
return {"path": self.path, "replays": self.replays}
|
| 1024 |
+
|
| 1025 |
+
def __setstate__(self, state):
|
| 1026 |
+
path = state["path"]
|
| 1027 |
+
replays = state["replays"]
|
| 1028 |
+
table = _memory_mapped_arrow_table_from_file(path)
|
| 1029 |
+
table = self._apply_replays(table, replays)
|
| 1030 |
+
MemoryMappedTable.__init__(self, table, path=path, replays=replays)
|
| 1031 |
+
|
| 1032 |
+
@staticmethod
|
| 1033 |
+
def _apply_replays(table: pa.Table, replays: Optional[List[Replay]] = None) -> pa.Table:
|
| 1034 |
+
if replays is not None:
|
| 1035 |
+
for name, args, kwargs in replays:
|
| 1036 |
+
if name == "cast":
|
| 1037 |
+
table = table_cast(table, *args, **kwargs)
|
| 1038 |
+
elif name == "flatten":
|
| 1039 |
+
table = table_flatten(table, *args, **kwargs)
|
| 1040 |
+
else:
|
| 1041 |
+
table = getattr(table, name)(*args, **kwargs)
|
| 1042 |
+
return table
|
| 1043 |
+
|
| 1044 |
+
def _append_replay(self, replay: Replay) -> List[Replay]:
|
| 1045 |
+
replays = copy.deepcopy(self.replays)
|
| 1046 |
+
replays.append(replay)
|
| 1047 |
+
return replays
|
| 1048 |
+
|
| 1049 |
+
def slice(self, offset=0, length=None):
|
| 1050 |
+
"""
|
| 1051 |
+
Compute zero-copy slice of this Table.
|
| 1052 |
+
|
| 1053 |
+
Args:
|
| 1054 |
+
offset (`int`, defaults to `0`):
|
| 1055 |
+
Offset from start of table to slice.
|
| 1056 |
+
length (`int`, defaults to `None`):
|
| 1057 |
+
Length of slice (default is until end of table starting from
|
| 1058 |
+
offset).
|
| 1059 |
+
|
| 1060 |
+
Returns:
|
| 1061 |
+
`datasets.table.Table`
|
| 1062 |
+
"""
|
| 1063 |
+
replay = ("slice", (offset, length), {})
|
| 1064 |
+
replays = self._append_replay(replay)
|
| 1065 |
+
# Use fast slicing here
|
| 1066 |
+
return MemoryMappedTable(self.fast_slice(offset=offset, length=length), self.path, replays)
|
| 1067 |
+
|
| 1068 |
+
def filter(self, *args, **kwargs):
|
| 1069 |
+
"""
|
| 1070 |
+
Select records from a Table. See `pyarrow.compute.filter` for full usage.
|
| 1071 |
+
"""
|
| 1072 |
+
replay = ("filter", copy.deepcopy(args), copy.deepcopy(kwargs))
|
| 1073 |
+
replays = self._append_replay(replay)
|
| 1074 |
+
return MemoryMappedTable(self.table.filter(*args, **kwargs), self.path, replays)
|
| 1075 |
+
|
| 1076 |
+
def flatten(self, *args, **kwargs):
|
| 1077 |
+
"""
|
| 1078 |
+
Flatten this Table. Each column with a struct type is flattened
|
| 1079 |
+
into one column per struct field. Other columns are left unchanged.
|
| 1080 |
+
|
| 1081 |
+
Args:
|
| 1082 |
+
memory_pool (`MemoryPool`, defaults to `None`):
|
| 1083 |
+
For memory allocations, if required, otherwise use default pool.
|
| 1084 |
+
|
| 1085 |
+
Returns:
|
| 1086 |
+
`datasets.table.Table`
|
| 1087 |
+
"""
|
| 1088 |
+
replay = ("flatten", copy.deepcopy(args), copy.deepcopy(kwargs))
|
| 1089 |
+
replays = self._append_replay(replay)
|
| 1090 |
+
return MemoryMappedTable(table_flatten(self.table, *args, **kwargs), self.path, replays)
|
| 1091 |
+
|
| 1092 |
+
def combine_chunks(self, *args, **kwargs):
|
| 1093 |
+
"""
|
| 1094 |
+
Make a new table by combining the chunks this table has.
|
| 1095 |
+
|
| 1096 |
+
All the underlying chunks in the ChunkedArray of each column are
|
| 1097 |
+
concatenated into zero or one chunk.
|
| 1098 |
+
|
| 1099 |
+
Args:
|
| 1100 |
+
memory_pool (`MemoryPool`, defaults to `None`):
|
| 1101 |
+
For memory allocations, if required, otherwise use default pool.
|
| 1102 |
+
|
| 1103 |
+
Returns:
|
| 1104 |
+
`datasets.table.Table`
|
| 1105 |
+
"""
|
| 1106 |
+
replay = ("combine_chunks", copy.deepcopy(args), copy.deepcopy(kwargs))
|
| 1107 |
+
replays = self._append_replay(replay)
|
| 1108 |
+
return MemoryMappedTable(self.table.combine_chunks(*args, **kwargs), self.path, replays)
|
| 1109 |
+
|
| 1110 |
+
def cast(self, *args, **kwargs):
|
| 1111 |
+
"""
|
| 1112 |
+
Cast table values to another schema
|
| 1113 |
+
|
| 1114 |
+
Args:
|
| 1115 |
+
target_schema (`Schema`):
|
| 1116 |
+
Schema to cast to, the names and order of fields must match.
|
| 1117 |
+
safe (`bool`, defaults to `True`):
|
| 1118 |
+
Check for overflows or other unsafe conversions.
|
| 1119 |
+
|
| 1120 |
+
Returns:
|
| 1121 |
+
`datasets.table.Table`
|
| 1122 |
+
"""
|
| 1123 |
+
replay = ("cast", copy.deepcopy(args), copy.deepcopy(kwargs))
|
| 1124 |
+
replays = self._append_replay(replay)
|
| 1125 |
+
return MemoryMappedTable(table_cast(self.table, *args, **kwargs), self.path, replays)
|
| 1126 |
+
|
| 1127 |
+
def replace_schema_metadata(self, *args, **kwargs):
|
| 1128 |
+
"""
|
| 1129 |
+
EXPERIMENTAL: Create shallow copy of table by replacing schema
|
| 1130 |
+
key-value metadata with the indicated new metadata (which may be None,
|
| 1131 |
+
which deletes any existing metadata.
|
| 1132 |
+
|
| 1133 |
+
Args:
|
| 1134 |
+
metadata (`dict`, defaults to `None`):
|
| 1135 |
+
|
| 1136 |
+
Returns:
|
| 1137 |
+
`datasets.table.Table`: shallow_copy
|
| 1138 |
+
"""
|
| 1139 |
+
replay = ("replace_schema_metadata", copy.deepcopy(args), copy.deepcopy(kwargs))
|
| 1140 |
+
replays = self._append_replay(replay)
|
| 1141 |
+
return MemoryMappedTable(self.table.replace_schema_metadata(*args, **kwargs), self.path, replays)
|
| 1142 |
+
|
| 1143 |
+
def add_column(self, *args, **kwargs):
|
| 1144 |
+
"""
|
| 1145 |
+
Add column to Table at position.
|
| 1146 |
+
|
| 1147 |
+
A new table is returned with the column added, the original table
|
| 1148 |
+
object is left unchanged.
|
| 1149 |
+
|
| 1150 |
+
Args:
|
| 1151 |
+
i (`int`):
|
| 1152 |
+
Index to place the column at.
|
| 1153 |
+
field_ (`Union[str, pyarrow.Field]`):
|
| 1154 |
+
If a string is passed then the type is deduced from the column
|
| 1155 |
+
data.
|
| 1156 |
+
column (`Union[pyarrow.Array, List[pyarrow.Array]]`):
|
| 1157 |
+
Column data.
|
| 1158 |
+
|
| 1159 |
+
Returns:
|
| 1160 |
+
`datasets.table.Table`: New table with the passed column added.
|
| 1161 |
+
"""
|
| 1162 |
+
replay = ("add_column", copy.deepcopy(args), copy.deepcopy(kwargs))
|
| 1163 |
+
replays = self._append_replay(replay)
|
| 1164 |
+
return MemoryMappedTable(self.table.add_column(*args, **kwargs), self.path, replays)
|
| 1165 |
+
|
| 1166 |
+
def append_column(self, *args, **kwargs):
|
| 1167 |
+
"""
|
| 1168 |
+
Append column at end of columns.
|
| 1169 |
+
|
| 1170 |
+
Args:
|
| 1171 |
+
field_ (`Union[str, pyarrow.Field]`):
|
| 1172 |
+
If a string is passed then the type is deduced from the column
|
| 1173 |
+
data.
|
| 1174 |
+
column (`Union[pyarrow.Array, List[pyarrow.Array]]`):
|
| 1175 |
+
Column data.
|
| 1176 |
+
|
| 1177 |
+
Returns:
|
| 1178 |
+
`datasets.table.Table`:
|
| 1179 |
+
New table with the passed column added.
|
| 1180 |
+
"""
|
| 1181 |
+
replay = ("append_column", copy.deepcopy(args), copy.deepcopy(kwargs))
|
| 1182 |
+
replays = self._append_replay(replay)
|
| 1183 |
+
return MemoryMappedTable(self.table.append_column(*args, **kwargs), self.path, replays)
|
| 1184 |
+
|
| 1185 |
+
def remove_column(self, *args, **kwargs):
|
| 1186 |
+
"""
|
| 1187 |
+
Create new Table with the indicated column removed.
|
| 1188 |
+
|
| 1189 |
+
Args:
|
| 1190 |
+
i (`int`):
|
| 1191 |
+
Index of column to remove.
|
| 1192 |
+
|
| 1193 |
+
Returns:
|
| 1194 |
+
`datasets.table.Table`:
|
| 1195 |
+
New table without the column.
|
| 1196 |
+
"""
|
| 1197 |
+
replay = ("remove_column", copy.deepcopy(args), copy.deepcopy(kwargs))
|
| 1198 |
+
replays = self._append_replay(replay)
|
| 1199 |
+
return MemoryMappedTable(self.table.remove_column(*args, **kwargs), self.path, replays)
|
| 1200 |
+
|
| 1201 |
+
def set_column(self, *args, **kwargs):
|
| 1202 |
+
"""
|
| 1203 |
+
Replace column in Table at position.
|
| 1204 |
+
|
| 1205 |
+
Args:
|
| 1206 |
+
i (`int`):
|
| 1207 |
+
Index to place the column at.
|
| 1208 |
+
field_ (`Union[str, pyarrow.Field]`):
|
| 1209 |
+
If a string is passed then the type is deduced from the column
|
| 1210 |
+
data.
|
| 1211 |
+
column (`Union[pyarrow.Array, List[pyarrow.Array]]`):
|
| 1212 |
+
Column data.
|
| 1213 |
+
|
| 1214 |
+
Returns:
|
| 1215 |
+
`datasets.table.Table`:
|
| 1216 |
+
New table with the passed column set.
|
| 1217 |
+
"""
|
| 1218 |
+
replay = ("set_column", copy.deepcopy(args), copy.deepcopy(kwargs))
|
| 1219 |
+
replays = self._append_replay(replay)
|
| 1220 |
+
return MemoryMappedTable(self.table.set_column(*args, **kwargs), self.path, replays)
|
| 1221 |
+
|
| 1222 |
+
def rename_columns(self, *args, **kwargs):
|
| 1223 |
+
"""
|
| 1224 |
+
Create new table with columns renamed to provided names.
|
| 1225 |
+
"""
|
| 1226 |
+
replay = ("rename_columns", copy.deepcopy(args), copy.deepcopy(kwargs))
|
| 1227 |
+
replays = self._append_replay(replay)
|
| 1228 |
+
return MemoryMappedTable(self.table.rename_columns(*args, **kwargs), self.path, replays)
|
| 1229 |
+
|
| 1230 |
+
def drop(self, *args, **kwargs):
|
| 1231 |
+
"""
|
| 1232 |
+
Drop one or more columns and return a new table.
|
| 1233 |
+
|
| 1234 |
+
Args:
|
| 1235 |
+
columns (`List[str]`):
|
| 1236 |
+
List of field names referencing existing columns.
|
| 1237 |
+
|
| 1238 |
+
Raises:
|
| 1239 |
+
`KeyError` : if any of the passed columns name are not existing.
|
| 1240 |
+
|
| 1241 |
+
Returns:
|
| 1242 |
+
`datasets.table.Table`:
|
| 1243 |
+
New table without the columns.
|
| 1244 |
+
"""
|
| 1245 |
+
replay = ("drop", copy.deepcopy(args), copy.deepcopy(kwargs))
|
| 1246 |
+
replays = self._append_replay(replay)
|
| 1247 |
+
return MemoryMappedTable(self.table.drop(*args, **kwargs), self.path, replays)
|
| 1248 |
+
|
| 1249 |
+
def select(self, *args, **kwargs):
|
| 1250 |
+
"""
|
| 1251 |
+
Select columns of the table.
|
| 1252 |
+
|
| 1253 |
+
Returns a new table with the specified columns, and metadata preserved.
|
| 1254 |
+
|
| 1255 |
+
Args:
|
| 1256 |
+
columns (:obj:`Union[List[str], List[int]]`):
|
| 1257 |
+
The column names or integer indices to select.
|
| 1258 |
+
|
| 1259 |
+
Returns:
|
| 1260 |
+
:class:`datasets.table.Table`: New table with the specified columns, and metadata preserved.
|
| 1261 |
+
"""
|
| 1262 |
+
replay = ("select", copy.deepcopy(args), copy.deepcopy(kwargs))
|
| 1263 |
+
replays = self._append_replay(replay)
|
| 1264 |
+
return MemoryMappedTable(self.table.select(*args, **kwargs), self.path, replays)
|
| 1265 |
+
|
| 1266 |
+
|
| 1267 |
+
# A ConcatenationTable is the concatenation of several tables.
|
| 1268 |
+
# The ``blocks`` attributes stores a list of list of blocks.
|
| 1269 |
+
# The first axis concatenates the tables along the axis 0 (it appends rows),
|
| 1270 |
+
# while the second axis concatenates tables along the axis 1 (it appends columns).
|
| 1271 |
+
TableBlockContainer = TypeVar("TableBlockContainer", TableBlock, List[TableBlock], List[List[TableBlock]])
|
| 1272 |
+
|
| 1273 |
+
|
| 1274 |
+
class ConcatenationTable(Table):
|
| 1275 |
+
"""
|
| 1276 |
+
The table comes from the concatenation of several tables called blocks.
|
| 1277 |
+
It enables concatenation on both axis 0 (append rows) and axis 1 (append columns).
|
| 1278 |
+
|
| 1279 |
+
The underlying tables are called "blocks" and can be either `InMemoryTable`
|
| 1280 |
+
or `MemoryMappedTable` objects.
|
| 1281 |
+
This allows to combine tables that come from memory or that are memory mapped.
|
| 1282 |
+
When a `ConcatenationTable` is pickled, then each block is pickled:
|
| 1283 |
+
- the `InMemoryTable` objects are pickled by copying all the data in memory.
|
| 1284 |
+
- the MemoryMappedTable objects are pickled without copying the data into memory.
|
| 1285 |
+
Instead, only the path to the memory mapped arrow file is pickled, as well as the list
|
| 1286 |
+
of transforms to "replays" when reloading the table from the disk.
|
| 1287 |
+
|
| 1288 |
+
Its implementation requires to store each block separately.
|
| 1289 |
+
The `blocks` attributes stores a list of list of blocks.
|
| 1290 |
+
The first axis concatenates the tables along the axis 0 (it appends rows),
|
| 1291 |
+
while the second axis concatenates tables along the axis 1 (it appends columns).
|
| 1292 |
+
|
| 1293 |
+
If some columns are missing when concatenating on axis 0, they are filled with null values.
|
| 1294 |
+
This is done using `pyarrow.concat_tables(tables, promote=True)`.
|
| 1295 |
+
|
| 1296 |
+
You can access the fully combined table by accessing the `ConcatenationTable.table` attribute,
|
| 1297 |
+
and the blocks by accessing the `ConcatenationTable.blocks` attribute.
|
| 1298 |
+
"""
|
| 1299 |
+
|
| 1300 |
+
def __init__(self, table: pa.Table, blocks: List[List[TableBlock]]):
|
| 1301 |
+
super().__init__(table)
|
| 1302 |
+
self.blocks = blocks
|
| 1303 |
+
# Check that all the blocks have the right type.
|
| 1304 |
+
# Only InMemoryTable and MemoryMappedTable are allowed.
|
| 1305 |
+
for subtables in blocks:
|
| 1306 |
+
for subtable in subtables:
|
| 1307 |
+
if not isinstance(subtable, TableBlock):
|
| 1308 |
+
raise TypeError(
|
| 1309 |
+
"The blocks of a ConcatenationTable must be InMemoryTable or MemoryMappedTable objects"
|
| 1310 |
+
f", but got {subtable}."
|
| 1311 |
+
)
|
| 1312 |
+
|
| 1313 |
+
def __getstate__(self):
|
| 1314 |
+
return {"blocks": self.blocks}
|
| 1315 |
+
|
| 1316 |
+
def __setstate__(self, state):
|
| 1317 |
+
blocks = state["blocks"]
|
| 1318 |
+
table = self._concat_blocks_horizontally_and_vertically(blocks)
|
| 1319 |
+
ConcatenationTable.__init__(self, table, blocks=blocks)
|
| 1320 |
+
|
| 1321 |
+
@staticmethod
|
| 1322 |
+
def _concat_blocks(blocks: List[Union[TableBlock, pa.Table]], axis: int = 0) -> pa.Table:
|
| 1323 |
+
pa_tables = [table.table if hasattr(table, "table") else table for table in blocks]
|
| 1324 |
+
if axis == 0:
|
| 1325 |
+
# we set promote=True to fill missing columns with null values
|
| 1326 |
+
if config.PYARROW_VERSION.major < 14:
|
| 1327 |
+
return pa.concat_tables(pa_tables, promote=True)
|
| 1328 |
+
else:
|
| 1329 |
+
return pa.concat_tables(pa_tables, promote_options="default")
|
| 1330 |
+
elif axis == 1:
|
| 1331 |
+
for i, table in enumerate(pa_tables):
|
| 1332 |
+
if i == 0:
|
| 1333 |
+
pa_table = table
|
| 1334 |
+
else:
|
| 1335 |
+
for name, col in zip(table.column_names, table.columns):
|
| 1336 |
+
pa_table = pa_table.append_column(name, col)
|
| 1337 |
+
return pa_table
|
| 1338 |
+
else:
|
| 1339 |
+
raise ValueError("'axis' must be either 0 or 1")
|
| 1340 |
+
|
| 1341 |
+
@classmethod
|
| 1342 |
+
def _concat_blocks_horizontally_and_vertically(cls, blocks: List[List[TableBlock]]) -> pa.Table:
|
| 1343 |
+
pa_tables_to_concat_vertically = []
|
| 1344 |
+
for i, tables in enumerate(blocks):
|
| 1345 |
+
if not tables:
|
| 1346 |
+
continue
|
| 1347 |
+
pa_table_horizontally_concatenated = cls._concat_blocks(tables, axis=1)
|
| 1348 |
+
pa_tables_to_concat_vertically.append(pa_table_horizontally_concatenated)
|
| 1349 |
+
return cls._concat_blocks(pa_tables_to_concat_vertically, axis=0)
|
| 1350 |
+
|
| 1351 |
+
@classmethod
|
| 1352 |
+
def _merge_blocks(cls, blocks: TableBlockContainer, axis: Optional[int] = None) -> TableBlockContainer:
|
| 1353 |
+
if axis is not None:
|
| 1354 |
+
merged_blocks = []
|
| 1355 |
+
for is_in_memory, block_group in groupby(blocks, key=lambda x: isinstance(x, InMemoryTable)):
|
| 1356 |
+
if is_in_memory:
|
| 1357 |
+
block_group = [InMemoryTable(cls._concat_blocks(list(block_group), axis=axis))]
|
| 1358 |
+
merged_blocks += list(block_group)
|
| 1359 |
+
else: # both
|
| 1360 |
+
merged_blocks = [cls._merge_blocks(row_block, axis=1) for row_block in blocks]
|
| 1361 |
+
if all(len(row_block) == 1 for row_block in merged_blocks):
|
| 1362 |
+
merged_blocks = cls._merge_blocks(
|
| 1363 |
+
[block for row_block in merged_blocks for block in row_block], axis=0
|
| 1364 |
+
)
|
| 1365 |
+
return merged_blocks
|
| 1366 |
+
|
| 1367 |
+
@classmethod
|
| 1368 |
+
def _consolidate_blocks(cls, blocks: TableBlockContainer) -> TableBlockContainer:
|
| 1369 |
+
if isinstance(blocks, TableBlock):
|
| 1370 |
+
return blocks
|
| 1371 |
+
elif isinstance(blocks[0], TableBlock):
|
| 1372 |
+
return cls._merge_blocks(blocks, axis=0)
|
| 1373 |
+
else:
|
| 1374 |
+
return cls._merge_blocks(blocks)
|
| 1375 |
+
|
| 1376 |
+
@classmethod
|
| 1377 |
+
def from_blocks(cls, blocks: TableBlockContainer) -> "ConcatenationTable":
|
| 1378 |
+
blocks = cls._consolidate_blocks(blocks)
|
| 1379 |
+
if isinstance(blocks, TableBlock):
|
| 1380 |
+
table = blocks
|
| 1381 |
+
return cls(table.table, [[table]])
|
| 1382 |
+
elif isinstance(blocks[0], TableBlock):
|
| 1383 |
+
table = cls._concat_blocks(blocks, axis=0)
|
| 1384 |
+
blocks = [[t] for t in blocks]
|
| 1385 |
+
return cls(table, blocks)
|
| 1386 |
+
else:
|
| 1387 |
+
table = cls._concat_blocks_horizontally_and_vertically(blocks)
|
| 1388 |
+
return cls(table, blocks)
|
| 1389 |
+
|
| 1390 |
+
@classmethod
|
| 1391 |
+
def from_tables(cls, tables: List[Union[pa.Table, Table]], axis: int = 0) -> "ConcatenationTable":
|
| 1392 |
+
"""Create `ConcatenationTable` from list of tables.
|
| 1393 |
+
|
| 1394 |
+
Args:
|
| 1395 |
+
tables (list of `Table` or list of `pyarrow.Table`):
|
| 1396 |
+
List of tables.
|
| 1397 |
+
axis (`{0, 1}`, defaults to `0`, meaning over rows):
|
| 1398 |
+
Axis to concatenate over, where `0` means over rows (vertically) and `1` means over columns
|
| 1399 |
+
(horizontally).
|
| 1400 |
+
|
| 1401 |
+
<Added version="1.6.0"/>
|
| 1402 |
+
"""
|
| 1403 |
+
|
| 1404 |
+
def to_blocks(table: Union[pa.Table, Table]) -> List[List[TableBlock]]:
|
| 1405 |
+
if isinstance(table, pa.Table):
|
| 1406 |
+
return [[InMemoryTable(table)]]
|
| 1407 |
+
elif isinstance(table, ConcatenationTable):
|
| 1408 |
+
return copy.deepcopy(table.blocks)
|
| 1409 |
+
else:
|
| 1410 |
+
return [[table]]
|
| 1411 |
+
|
| 1412 |
+
def _slice_row_block(row_block: List[TableBlock], length: int) -> Tuple[List[TableBlock], List[TableBlock]]:
|
| 1413 |
+
sliced = [table.slice(0, length) for table in row_block]
|
| 1414 |
+
remainder = [table.slice(length, len(row_block[0]) - length) for table in row_block]
|
| 1415 |
+
return sliced, remainder
|
| 1416 |
+
|
| 1417 |
+
def _split_both_like(
|
| 1418 |
+
result: List[List[TableBlock]], blocks: List[List[TableBlock]]
|
| 1419 |
+
) -> Tuple[List[List[TableBlock]], List[List[TableBlock]]]:
|
| 1420 |
+
"""
|
| 1421 |
+
Make sure each row_block contain the same num_rows to be able to concatenate them on axis=1.
|
| 1422 |
+
|
| 1423 |
+
To do so, we modify both blocks sets to have the same row_blocks boundaries.
|
| 1424 |
+
For example, if `result` has 2 row_blocks of 3 rows and `blocks` has 3 row_blocks of 2 rows,
|
| 1425 |
+
we modify both to have 4 row_blocks of size 2, 1, 1 and 2:
|
| 1426 |
+
|
| 1427 |
+
[ x x x | x x x ]
|
| 1428 |
+
+ [ y y | y y | y y ]
|
| 1429 |
+
-----------------------------
|
| 1430 |
+
= [ x x | x | x | x x ]
|
| 1431 |
+
[ y y | y | y | y y ]
|
| 1432 |
+
|
| 1433 |
+
"""
|
| 1434 |
+
result, blocks = list(result), list(blocks)
|
| 1435 |
+
new_result, new_blocks = [], []
|
| 1436 |
+
while result and blocks:
|
| 1437 |
+
# we slice the longest row block to save two row blocks of same length
|
| 1438 |
+
# and we replace the long row block by its remainder if necessary
|
| 1439 |
+
if len(result[0][0]) > len(blocks[0][0]):
|
| 1440 |
+
new_blocks.append(blocks[0])
|
| 1441 |
+
sliced, result[0] = _slice_row_block(result[0], len(blocks.pop(0)[0]))
|
| 1442 |
+
new_result.append(sliced)
|
| 1443 |
+
elif len(result[0][0]) < len(blocks[0][0]):
|
| 1444 |
+
new_result.append(result[0])
|
| 1445 |
+
sliced, blocks[0] = _slice_row_block(blocks[0], len(result.pop(0)[0]))
|
| 1446 |
+
new_blocks.append(sliced)
|
| 1447 |
+
else:
|
| 1448 |
+
new_result.append(result.pop(0))
|
| 1449 |
+
new_blocks.append(blocks.pop(0))
|
| 1450 |
+
if result or blocks:
|
| 1451 |
+
raise ValueError("Failed to concatenate on axis=1 because tables don't have the same number of rows")
|
| 1452 |
+
return new_result, new_blocks
|
| 1453 |
+
|
| 1454 |
+
def _extend_blocks(
|
| 1455 |
+
result: List[List[TableBlock]], blocks: List[List[TableBlock]], axis: int = 0
|
| 1456 |
+
) -> List[List[TableBlock]]:
|
| 1457 |
+
if axis == 0:
|
| 1458 |
+
result.extend(blocks)
|
| 1459 |
+
elif axis == 1:
|
| 1460 |
+
# We make sure each row_block have the same num_rows
|
| 1461 |
+
result, blocks = _split_both_like(result, blocks)
|
| 1462 |
+
for i, row_block in enumerate(blocks):
|
| 1463 |
+
result[i].extend(row_block)
|
| 1464 |
+
return result
|
| 1465 |
+
|
| 1466 |
+
blocks = to_blocks(tables[0])
|
| 1467 |
+
for table in tables[1:]:
|
| 1468 |
+
table_blocks = to_blocks(table)
|
| 1469 |
+
blocks = _extend_blocks(blocks, table_blocks, axis=axis)
|
| 1470 |
+
return cls.from_blocks(blocks)
|
| 1471 |
+
|
| 1472 |
+
@property
|
| 1473 |
+
def _slices(self):
|
| 1474 |
+
offset = 0
|
| 1475 |
+
for tables in self.blocks:
|
| 1476 |
+
length = len(tables[0])
|
| 1477 |
+
yield (offset, length)
|
| 1478 |
+
offset += length
|
| 1479 |
+
|
| 1480 |
+
def slice(self, offset=0, length=None):
|
| 1481 |
+
"""
|
| 1482 |
+
Compute zero-copy slice of this Table.
|
| 1483 |
+
|
| 1484 |
+
Args:
|
| 1485 |
+
offset (`int`, defaults to `0`):
|
| 1486 |
+
Offset from start of table to slice.
|
| 1487 |
+
length (`int`, defaults to `None`):
|
| 1488 |
+
Length of slice (default is until end of table starting from
|
| 1489 |
+
offset).
|
| 1490 |
+
|
| 1491 |
+
Returns:
|
| 1492 |
+
`datasets.table.Table`
|
| 1493 |
+
"""
|
| 1494 |
+
table = self.table.slice(offset, length=length)
|
| 1495 |
+
length = length if length is not None else self.num_rows - offset
|
| 1496 |
+
blocks = []
|
| 1497 |
+
for tables in self.blocks:
|
| 1498 |
+
n_rows = len(tables[0])
|
| 1499 |
+
if length == 0:
|
| 1500 |
+
break
|
| 1501 |
+
elif n_rows <= offset:
|
| 1502 |
+
offset = offset - n_rows
|
| 1503 |
+
elif n_rows <= offset + length:
|
| 1504 |
+
blocks.append([t.slice(offset) for t in tables])
|
| 1505 |
+
length, offset = length + offset - n_rows, 0
|
| 1506 |
+
else:
|
| 1507 |
+
blocks.append([t.slice(offset, length) for t in tables])
|
| 1508 |
+
length, offset = 0, 0
|
| 1509 |
+
return ConcatenationTable(table, blocks)
|
| 1510 |
+
|
| 1511 |
+
def filter(self, mask, *args, **kwargs):
|
| 1512 |
+
"""
|
| 1513 |
+
Select records from a Table. See `pyarrow.compute.filter` for full usage.
|
| 1514 |
+
"""
|
| 1515 |
+
table = self.table.filter(mask, *args, **kwargs)
|
| 1516 |
+
blocks = []
|
| 1517 |
+
for (offset, length), tables in zip(self._slices, self.blocks):
|
| 1518 |
+
submask = mask.slice(offset, length)
|
| 1519 |
+
blocks.append([t.filter(submask, *args, **kwargs) for t in tables])
|
| 1520 |
+
return ConcatenationTable(table, blocks)
|
| 1521 |
+
|
| 1522 |
+
def flatten(self, *args, **kwargs):
|
| 1523 |
+
"""
|
| 1524 |
+
Flatten this Table. Each column with a struct type is flattened
|
| 1525 |
+
into one column per struct field. Other columns are left unchanged.
|
| 1526 |
+
|
| 1527 |
+
Args:
|
| 1528 |
+
memory_pool (`MemoryPool`, defaults to `None`):
|
| 1529 |
+
For memory allocations, if required, otherwise use default pool.
|
| 1530 |
+
|
| 1531 |
+
Returns:
|
| 1532 |
+
`datasets.table.Table`
|
| 1533 |
+
"""
|
| 1534 |
+
table = table_flatten(self.table, *args, **kwargs)
|
| 1535 |
+
blocks = []
|
| 1536 |
+
for tables in self.blocks:
|
| 1537 |
+
blocks.append([t.flatten(*args, **kwargs) for t in tables])
|
| 1538 |
+
return ConcatenationTable(table, blocks)
|
| 1539 |
+
|
| 1540 |
+
def combine_chunks(self, *args, **kwargs):
|
| 1541 |
+
"""
|
| 1542 |
+
Make a new table by combining the chunks this table has.
|
| 1543 |
+
|
| 1544 |
+
All the underlying chunks in the `ChunkedArray` of each column are
|
| 1545 |
+
concatenated into zero or one chunk.
|
| 1546 |
+
|
| 1547 |
+
Args:
|
| 1548 |
+
memory_pool (`MemoryPool`, defaults to `None`):
|
| 1549 |
+
For memory allocations, if required, otherwise use default pool.
|
| 1550 |
+
|
| 1551 |
+
Returns:
|
| 1552 |
+
`datasets.table.Table`
|
| 1553 |
+
"""
|
| 1554 |
+
table = self.table.combine_chunks(*args, **kwargs)
|
| 1555 |
+
blocks = []
|
| 1556 |
+
for tables in self.blocks:
|
| 1557 |
+
blocks.append([t.combine_chunks(*args, **kwargs) for t in tables])
|
| 1558 |
+
return ConcatenationTable(table, blocks)
|
| 1559 |
+
|
| 1560 |
+
def cast(self, target_schema, *args, **kwargs):
|
| 1561 |
+
"""
|
| 1562 |
+
Cast table values to another schema.
|
| 1563 |
+
|
| 1564 |
+
Args:
|
| 1565 |
+
target_schema (`Schema`):
|
| 1566 |
+
Schema to cast to, the names and order of fields must match.
|
| 1567 |
+
safe (`bool`, defaults to `True`):
|
| 1568 |
+
Check for overflows or other unsafe conversions.
|
| 1569 |
+
|
| 1570 |
+
Returns:
|
| 1571 |
+
`datasets.table.Table`
|
| 1572 |
+
"""
|
| 1573 |
+
from .features import Features
|
| 1574 |
+
|
| 1575 |
+
table = table_cast(self.table, target_schema, *args, **kwargs)
|
| 1576 |
+
target_features = Features.from_arrow_schema(target_schema)
|
| 1577 |
+
blocks = []
|
| 1578 |
+
for subtables in self.blocks:
|
| 1579 |
+
new_tables = []
|
| 1580 |
+
fields = list(target_schema)
|
| 1581 |
+
for subtable in subtables:
|
| 1582 |
+
subfields = []
|
| 1583 |
+
for name in subtable.column_names:
|
| 1584 |
+
subfields.append(fields.pop(next(i for i, field in enumerate(fields) if field.name == name)))
|
| 1585 |
+
subfeatures = Features({subfield.name: target_features[subfield.name] for subfield in subfields})
|
| 1586 |
+
subschema = subfeatures.arrow_schema
|
| 1587 |
+
new_tables.append(subtable.cast(subschema, *args, **kwargs))
|
| 1588 |
+
blocks.append(new_tables)
|
| 1589 |
+
return ConcatenationTable(table, blocks)
|
| 1590 |
+
|
| 1591 |
+
def replace_schema_metadata(self, *args, **kwargs):
|
| 1592 |
+
"""
|
| 1593 |
+
EXPERIMENTAL: Create shallow copy of table by replacing schema
|
| 1594 |
+
key-value metadata with the indicated new metadata (which may be `None`,
|
| 1595 |
+
which deletes any existing metadata).
|
| 1596 |
+
|
| 1597 |
+
Args:
|
| 1598 |
+
metadata (`dict`, defaults to `None`):
|
| 1599 |
+
|
| 1600 |
+
Returns:
|
| 1601 |
+
`datasets.table.Table`: shallow_copy
|
| 1602 |
+
"""
|
| 1603 |
+
table = self.table.replace_schema_metadata(*args, **kwargs)
|
| 1604 |
+
blocks = []
|
| 1605 |
+
for tables in self.blocks:
|
| 1606 |
+
blocks.append([t.replace_schema_metadata(*args, **kwargs) for t in tables])
|
| 1607 |
+
return ConcatenationTable(table, self.blocks)
|
| 1608 |
+
|
| 1609 |
+
def add_column(self, *args, **kwargs):
|
| 1610 |
+
"""
|
| 1611 |
+
Add column to Table at position.
|
| 1612 |
+
|
| 1613 |
+
A new table is returned with the column added, the original table
|
| 1614 |
+
object is left unchanged.
|
| 1615 |
+
|
| 1616 |
+
Args:
|
| 1617 |
+
i (`int`):
|
| 1618 |
+
Index to place the column at.
|
| 1619 |
+
field_ (`Union[str, pyarrow.Field]`):
|
| 1620 |
+
If a string is passed then the type is deduced from the column
|
| 1621 |
+
data.
|
| 1622 |
+
column (`Union[pyarrow.Array, List[pyarrow.Array]]`):
|
| 1623 |
+
Column data.
|
| 1624 |
+
|
| 1625 |
+
Returns:
|
| 1626 |
+
`datasets.table.Table`: New table with the passed column added.
|
| 1627 |
+
"""
|
| 1628 |
+
raise NotImplementedError()
|
| 1629 |
+
|
| 1630 |
+
def append_column(self, *args, **kwargs):
|
| 1631 |
+
"""
|
| 1632 |
+
Append column at end of columns.
|
| 1633 |
+
|
| 1634 |
+
Args:
|
| 1635 |
+
field_ (`Union[str, pyarrow.Field]`):
|
| 1636 |
+
If a string is passed then the type is deduced from the column
|
| 1637 |
+
data.
|
| 1638 |
+
column (`Union[pyarrow.Array, List[pyarrow.Array]]`):
|
| 1639 |
+
Column data.
|
| 1640 |
+
|
| 1641 |
+
Returns:
|
| 1642 |
+
`datasets.table.Table`:
|
| 1643 |
+
New table with the passed column added.
|
| 1644 |
+
"""
|
| 1645 |
+
raise NotImplementedError()
|
| 1646 |
+
|
| 1647 |
+
def remove_column(self, i, *args, **kwargs):
|
| 1648 |
+
"""
|
| 1649 |
+
Create new Table with the indicated column removed.
|
| 1650 |
+
|
| 1651 |
+
Args:
|
| 1652 |
+
i (`int`):
|
| 1653 |
+
Index of column to remove.
|
| 1654 |
+
|
| 1655 |
+
Returns:
|
| 1656 |
+
`datasets.table.Table`:
|
| 1657 |
+
New table without the column.
|
| 1658 |
+
"""
|
| 1659 |
+
table = self.table.remove_column(i, *args, **kwargs)
|
| 1660 |
+
name = self.table.column_names[i]
|
| 1661 |
+
blocks = []
|
| 1662 |
+
for tables in self.blocks:
|
| 1663 |
+
blocks.append(
|
| 1664 |
+
[
|
| 1665 |
+
t.remove_column(t.column_names.index(name), *args, **kwargs) if name in t.column_names else t
|
| 1666 |
+
for t in tables
|
| 1667 |
+
]
|
| 1668 |
+
)
|
| 1669 |
+
return ConcatenationTable(table, blocks)
|
| 1670 |
+
|
| 1671 |
+
def set_column(self, *args, **kwargs):
|
| 1672 |
+
"""
|
| 1673 |
+
Replace column in Table at position.
|
| 1674 |
+
|
| 1675 |
+
Args:
|
| 1676 |
+
i (`int`):
|
| 1677 |
+
Index to place the column at.
|
| 1678 |
+
field_ (`Union[str, pyarrow.Field]`):
|
| 1679 |
+
If a string is passed then the type is deduced from the column
|
| 1680 |
+
data.
|
| 1681 |
+
column (`Union[pyarrow.Array, List[pyarrow.Array]]`):
|
| 1682 |
+
Column data.
|
| 1683 |
+
|
| 1684 |
+
Returns:
|
| 1685 |
+
`datasets.table.Table`:
|
| 1686 |
+
New table with the passed column set.
|
| 1687 |
+
"""
|
| 1688 |
+
raise NotImplementedError()
|
| 1689 |
+
|
| 1690 |
+
def rename_columns(self, names, *args, **kwargs):
|
| 1691 |
+
"""
|
| 1692 |
+
Create new table with columns renamed to provided names.
|
| 1693 |
+
"""
|
| 1694 |
+
table = self.table.rename_columns(names, *args, **kwargs)
|
| 1695 |
+
names = dict(zip(self.table.column_names, names))
|
| 1696 |
+
blocks = []
|
| 1697 |
+
for tables in self.blocks:
|
| 1698 |
+
blocks.append(
|
| 1699 |
+
[t.rename_columns([names[name] for name in t.column_names], *args, **kwargs) for t in tables]
|
| 1700 |
+
)
|
| 1701 |
+
return ConcatenationTable(table, blocks)
|
| 1702 |
+
|
| 1703 |
+
def drop(self, columns, *args, **kwargs):
|
| 1704 |
+
"""
|
| 1705 |
+
Drop one or more columns and return a new table.
|
| 1706 |
+
|
| 1707 |
+
Args:
|
| 1708 |
+
columns (`List[str]`):
|
| 1709 |
+
List of field names referencing existing columns.
|
| 1710 |
+
|
| 1711 |
+
Raises:
|
| 1712 |
+
`KeyError` : if any of the passed columns name are not existing.
|
| 1713 |
+
|
| 1714 |
+
Returns:
|
| 1715 |
+
`datasets.table.Table`:
|
| 1716 |
+
New table without the columns.
|
| 1717 |
+
"""
|
| 1718 |
+
table = self.table.drop(columns, *args, **kwargs)
|
| 1719 |
+
blocks = []
|
| 1720 |
+
for tables in self.blocks:
|
| 1721 |
+
blocks.append([t.drop([c for c in columns if c in t.column_names], *args, **kwargs) for t in tables])
|
| 1722 |
+
return ConcatenationTable(table, blocks)
|
| 1723 |
+
|
| 1724 |
+
def select(self, columns, *args, **kwargs):
|
| 1725 |
+
"""
|
| 1726 |
+
Select columns of the table.
|
| 1727 |
+
|
| 1728 |
+
Returns a new table with the specified columns, and metadata preserved.
|
| 1729 |
+
|
| 1730 |
+
Args:
|
| 1731 |
+
columns (:obj:`Union[List[str], List[int]]`):
|
| 1732 |
+
The column names or integer indices to select.
|
| 1733 |
+
|
| 1734 |
+
Returns:
|
| 1735 |
+
:class:`datasets.table.Table`: New table with the specified columns, and metadata preserved.
|
| 1736 |
+
"""
|
| 1737 |
+
table = self.table.select(columns, *args, **kwargs)
|
| 1738 |
+
blocks = []
|
| 1739 |
+
for tables in self.blocks:
|
| 1740 |
+
blocks.append([t.select([c for c in columns if c in t.column_names], *args, **kwargs) for t in tables])
|
| 1741 |
+
return ConcatenationTable(table, blocks)
|
| 1742 |
+
|
| 1743 |
+
|
| 1744 |
+
def concat_tables(tables: List[Table], axis: int = 0) -> Table:
|
| 1745 |
+
"""
|
| 1746 |
+
Concatenate tables.
|
| 1747 |
+
|
| 1748 |
+
Args:
|
| 1749 |
+
tables (list of `Table`):
|
| 1750 |
+
List of tables to be concatenated.
|
| 1751 |
+
axis (`{0, 1}`, defaults to `0`, meaning over rows):
|
| 1752 |
+
Axis to concatenate over, where `0` means over rows (vertically) and `1` means over columns
|
| 1753 |
+
(horizontally).
|
| 1754 |
+
|
| 1755 |
+
<Added version="1.6.0"/>
|
| 1756 |
+
Returns:
|
| 1757 |
+
`datasets.table.Table`:
|
| 1758 |
+
If the number of input tables is > 1, then the returned table is a `datasets.table.ConcatenationTable`.
|
| 1759 |
+
Otherwise if there's only one table, it is returned as is.
|
| 1760 |
+
"""
|
| 1761 |
+
tables = list(tables)
|
| 1762 |
+
if len(tables) == 1:
|
| 1763 |
+
return tables[0]
|
| 1764 |
+
return ConcatenationTable.from_tables(tables, axis=axis)
|
| 1765 |
+
|
| 1766 |
+
|
| 1767 |
+
def list_table_cache_files(table: Table) -> List[str]:
|
| 1768 |
+
"""
|
| 1769 |
+
Get the cache files that are loaded by the table.
|
| 1770 |
+
Cache file are used when parts of the table come from the disk via memory mapping.
|
| 1771 |
+
|
| 1772 |
+
Returns:
|
| 1773 |
+
`List[str]`:
|
| 1774 |
+
A list of paths to the cache files loaded by the table.
|
| 1775 |
+
"""
|
| 1776 |
+
if isinstance(table, ConcatenationTable):
|
| 1777 |
+
cache_files = []
|
| 1778 |
+
for subtables in table.blocks:
|
| 1779 |
+
for subtable in subtables:
|
| 1780 |
+
cache_files += list_table_cache_files(subtable)
|
| 1781 |
+
return cache_files
|
| 1782 |
+
elif isinstance(table, MemoryMappedTable):
|
| 1783 |
+
return [table.path]
|
| 1784 |
+
else:
|
| 1785 |
+
return []
|
| 1786 |
+
|
| 1787 |
+
|
| 1788 |
+
def _wrap_for_chunked_arrays(func):
|
| 1789 |
+
"""Apply the function on each chunk of a `pyarrow.ChunkedArray`, or on the array directly"""
|
| 1790 |
+
|
| 1791 |
+
def wrapper(array, *args, **kwargs):
|
| 1792 |
+
if isinstance(array, pa.ChunkedArray):
|
| 1793 |
+
return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks])
|
| 1794 |
+
else:
|
| 1795 |
+
return func(array, *args, **kwargs)
|
| 1796 |
+
|
| 1797 |
+
return wrapper
|
| 1798 |
+
|
| 1799 |
+
|
| 1800 |
+
def _are_list_values_of_length(array: pa.ListArray, length: int) -> bool:
|
| 1801 |
+
"""Check if all the sub-lists of a `pa.ListArray` have the specified length."""
|
| 1802 |
+
return pc.all(pc.equal(array.value_lengths(), length)).as_py() or array.null_count == len(array)
|
| 1803 |
+
|
| 1804 |
+
|
| 1805 |
+
def _combine_list_array_offsets_with_mask(array: pa.ListArray) -> pa.Array:
|
| 1806 |
+
"""Add the null bitmap to the offsets of a `pa.ListArray`."""
|
| 1807 |
+
offsets = array.offsets
|
| 1808 |
+
if array.null_count > 0:
|
| 1809 |
+
offsets = pa.concat_arrays(
|
| 1810 |
+
[
|
| 1811 |
+
pc.replace_with_mask(offsets[:-1], array.is_null(), pa.nulls(len(array), pa.int32())),
|
| 1812 |
+
offsets[-1:],
|
| 1813 |
+
]
|
| 1814 |
+
)
|
| 1815 |
+
return offsets
|
| 1816 |
+
|
| 1817 |
+
|
| 1818 |
+
def _storage_type(type: pa.DataType) -> pa.DataType:
|
| 1819 |
+
"""Convert a (possibly nested) `pa.ExtensionType` to its storage type."""
|
| 1820 |
+
if isinstance(type, pa.ExtensionType):
|
| 1821 |
+
return _storage_type(type.storage_type)
|
| 1822 |
+
elif isinstance(type, pa.StructType):
|
| 1823 |
+
return pa.struct([pa.field(field.name, _storage_type(field.type)) for field in type])
|
| 1824 |
+
elif isinstance(type, pa.ListType):
|
| 1825 |
+
return pa.list_(_storage_type(type.value_type))
|
| 1826 |
+
elif isinstance(type, pa.FixedSizeListType):
|
| 1827 |
+
return pa.list_(_storage_type(type.value_type), type.list_size)
|
| 1828 |
+
return type
|
| 1829 |
+
|
| 1830 |
+
|
| 1831 |
+
@_wrap_for_chunked_arrays
|
| 1832 |
+
def array_cast(array: pa.Array, pa_type: pa.DataType, allow_number_to_str=True):
|
| 1833 |
+
"""Improved version of `pa.Array.cast`
|
| 1834 |
+
|
| 1835 |
+
It supports casting `pa.StructArray` objects to re-order the fields.
|
| 1836 |
+
It also let you control certain aspects of the casting, e.g. whether
|
| 1837 |
+
to disable numbers (`floats` or `ints`) to strings.
|
| 1838 |
+
|
| 1839 |
+
Args:
|
| 1840 |
+
array (`pa.Array`):
|
| 1841 |
+
PyArrow array to cast
|
| 1842 |
+
pa_type (`pa.DataType`):
|
| 1843 |
+
Target PyArrow type
|
| 1844 |
+
allow_number_to_str (`bool`, defaults to `True`):
|
| 1845 |
+
Whether to allow casting numbers to strings.
|
| 1846 |
+
Defaults to `True`.
|
| 1847 |
+
|
| 1848 |
+
Raises:
|
| 1849 |
+
`pa.ArrowInvalidError`: if the arrow data casting fails
|
| 1850 |
+
`TypeError`: if the target type is not supported according, e.g.
|
| 1851 |
+
|
| 1852 |
+
- if a field is missing
|
| 1853 |
+
- if casting from numbers to strings and `allow_number_to_str` is `False`
|
| 1854 |
+
|
| 1855 |
+
Returns:
|
| 1856 |
+
`List[pyarrow.Array]`: the casted array
|
| 1857 |
+
"""
|
| 1858 |
+
_c = partial(array_cast, allow_number_to_str=allow_number_to_str)
|
| 1859 |
+
if isinstance(array, pa.ExtensionArray):
|
| 1860 |
+
array = array.storage
|
| 1861 |
+
if isinstance(pa_type, pa.ExtensionType):
|
| 1862 |
+
return pa_type.wrap_array(_c(array, pa_type.storage_type))
|
| 1863 |
+
elif array.type == pa_type:
|
| 1864 |
+
return array
|
| 1865 |
+
elif pa.types.is_struct(array.type):
|
| 1866 |
+
if pa.types.is_struct(pa_type) and ({field.name for field in pa_type} == {field.name for field in array.type}):
|
| 1867 |
+
if array.type.num_fields == 0:
|
| 1868 |
+
return array
|
| 1869 |
+
arrays = [_c(array.field(field.name), field.type) for field in pa_type]
|
| 1870 |
+
return pa.StructArray.from_arrays(arrays, fields=list(pa_type), mask=array.is_null())
|
| 1871 |
+
elif pa.types.is_list(array.type):
|
| 1872 |
+
if pa.types.is_fixed_size_list(pa_type):
|
| 1873 |
+
if _are_list_values_of_length(array, pa_type.list_size):
|
| 1874 |
+
if array.null_count > 0:
|
| 1875 |
+
# Ensure each null value in the array translates to [null] * pa_type.list_size in the array's values array
|
| 1876 |
+
array_type = array.type
|
| 1877 |
+
storage_type = _storage_type(array_type)
|
| 1878 |
+
if array_type != storage_type:
|
| 1879 |
+
# Temporarily convert to the storage type to support extension types in the slice operation
|
| 1880 |
+
array = _c(array, storage_type)
|
| 1881 |
+
array = pc.list_slice(array, 0, pa_type.list_size, return_fixed_size_list=True)
|
| 1882 |
+
array = _c(array, array_type)
|
| 1883 |
+
else:
|
| 1884 |
+
array = pc.list_slice(array, 0, pa_type.list_size, return_fixed_size_list=True)
|
| 1885 |
+
array_values = array.values
|
| 1886 |
+
if config.PYARROW_VERSION.major < 15:
|
| 1887 |
+
return pa.Array.from_buffers(
|
| 1888 |
+
pa_type,
|
| 1889 |
+
len(array),
|
| 1890 |
+
[array.is_valid().buffers()[1]],
|
| 1891 |
+
children=[_c(array_values, pa_type.value_type)],
|
| 1892 |
+
)
|
| 1893 |
+
else:
|
| 1894 |
+
return pa.FixedSizeListArray.from_arrays(
|
| 1895 |
+
_c(array_values, pa_type.value_type), pa_type.list_size, mask=array.is_null()
|
| 1896 |
+
)
|
| 1897 |
+
else:
|
| 1898 |
+
array_values = array.values[
|
| 1899 |
+
array.offset * pa_type.length : (array.offset + len(array)) * pa_type.length
|
| 1900 |
+
]
|
| 1901 |
+
return pa.FixedSizeListArray.from_arrays(_c(array_values, pa_type.value_type), pa_type.list_size)
|
| 1902 |
+
elif pa.types.is_list(pa_type):
|
| 1903 |
+
# Merge offsets with the null bitmap to avoid the "Null bitmap with offsets slice not supported" ArrowNotImplementedError
|
| 1904 |
+
array_offsets = _combine_list_array_offsets_with_mask(array)
|
| 1905 |
+
return pa.ListArray.from_arrays(array_offsets, _c(array.values, pa_type.value_type))
|
| 1906 |
+
elif pa.types.is_fixed_size_list(array.type):
|
| 1907 |
+
if pa.types.is_fixed_size_list(pa_type):
|
| 1908 |
+
if pa_type.list_size == array.type.list_size:
|
| 1909 |
+
array_values = array.values[
|
| 1910 |
+
array.offset * array.type.list_size : (array.offset + len(array)) * array.type.list_size
|
| 1911 |
+
]
|
| 1912 |
+
if config.PYARROW_VERSION.major < 15:
|
| 1913 |
+
return pa.Array.from_buffers(
|
| 1914 |
+
pa_type,
|
| 1915 |
+
len(array),
|
| 1916 |
+
[array.is_valid().buffers()[1]],
|
| 1917 |
+
children=[_c(array_values, pa_type.value_type)],
|
| 1918 |
+
)
|
| 1919 |
+
else:
|
| 1920 |
+
return pa.FixedSizeListArray.from_arrays(
|
| 1921 |
+
_c(array_values, pa_type.value_type), pa_type.list_size, mask=array.is_null()
|
| 1922 |
+
)
|
| 1923 |
+
elif pa.types.is_list(pa_type):
|
| 1924 |
+
array_offsets = (np.arange(len(array) + 1) + array.offset) * array.type.list_size
|
| 1925 |
+
return pa.ListArray.from_arrays(array_offsets, _c(array.values, pa_type.value_type), mask=array.is_null())
|
| 1926 |
+
else:
|
| 1927 |
+
if (
|
| 1928 |
+
not allow_number_to_str
|
| 1929 |
+
and pa.types.is_string(pa_type)
|
| 1930 |
+
and (pa.types.is_floating(array.type) or pa.types.is_integer(array.type))
|
| 1931 |
+
):
|
| 1932 |
+
raise TypeError(
|
| 1933 |
+
f"Couldn't cast array of type {array.type} to {pa_type} since allow_number_to_str is set to {allow_number_to_str}"
|
| 1934 |
+
)
|
| 1935 |
+
if pa.types.is_null(pa_type) and not pa.types.is_null(array.type):
|
| 1936 |
+
raise TypeError(f"Couldn't cast array of type {array.type} to {pa_type}")
|
| 1937 |
+
return array.cast(pa_type)
|
| 1938 |
+
raise TypeError(f"Couldn't cast array of type\n{array.type}\nto\n{pa_type}")
|
| 1939 |
+
|
| 1940 |
+
|
| 1941 |
+
@_wrap_for_chunked_arrays
|
| 1942 |
+
def cast_array_to_feature(array: pa.Array, feature: "FeatureType", allow_number_to_str=True):
|
| 1943 |
+
"""Cast an array to the arrow type that corresponds to the requested feature type.
|
| 1944 |
+
For custom features like [`Audio`] or [`Image`], it takes into account the "cast_storage" methods
|
| 1945 |
+
they defined to enable casting from other arrow types.
|
| 1946 |
+
|
| 1947 |
+
Args:
|
| 1948 |
+
array (`pa.Array`):
|
| 1949 |
+
The PyArrow array to cast.
|
| 1950 |
+
feature (`datasets.features.FeatureType`):
|
| 1951 |
+
The target feature type.
|
| 1952 |
+
allow_number_to_str (`bool`, defaults to `True`):
|
| 1953 |
+
Whether to allow casting numbers to strings.
|
| 1954 |
+
Defaults to `True`.
|
| 1955 |
+
|
| 1956 |
+
Raises:
|
| 1957 |
+
`pa.ArrowInvalidError`: if the arrow data casting fails
|
| 1958 |
+
`TypeError`: if the target type is not supported according, e.g.
|
| 1959 |
+
|
| 1960 |
+
- if a field is missing
|
| 1961 |
+
- if casting from numbers to strings and `allow_number_to_str` is `False`
|
| 1962 |
+
|
| 1963 |
+
Returns:
|
| 1964 |
+
array (`pyarrow.Array`): the casted array
|
| 1965 |
+
"""
|
| 1966 |
+
from .features.features import Sequence, get_nested_type
|
| 1967 |
+
|
| 1968 |
+
_c = partial(cast_array_to_feature, allow_number_to_str=allow_number_to_str)
|
| 1969 |
+
|
| 1970 |
+
if isinstance(array, pa.ExtensionArray):
|
| 1971 |
+
array = array.storage
|
| 1972 |
+
if hasattr(feature, "cast_storage"):
|
| 1973 |
+
return feature.cast_storage(array)
|
| 1974 |
+
|
| 1975 |
+
elif pa.types.is_struct(array.type):
|
| 1976 |
+
# feature must be a dict or Sequence(subfeatures_dict)
|
| 1977 |
+
if isinstance(feature, Sequence) and isinstance(feature.feature, dict):
|
| 1978 |
+
feature = {
|
| 1979 |
+
name: Sequence(subfeature, length=feature.length) for name, subfeature in feature.feature.items()
|
| 1980 |
+
}
|
| 1981 |
+
if isinstance(feature, dict) and {field.name for field in array.type} == set(feature):
|
| 1982 |
+
if array.type.num_fields == 0:
|
| 1983 |
+
return array
|
| 1984 |
+
arrays = [_c(array.field(name), subfeature) for name, subfeature in feature.items()]
|
| 1985 |
+
return pa.StructArray.from_arrays(arrays, names=list(feature), mask=array.is_null())
|
| 1986 |
+
elif pa.types.is_list(array.type):
|
| 1987 |
+
# feature must be either [subfeature] or Sequence(subfeature)
|
| 1988 |
+
if isinstance(feature, list):
|
| 1989 |
+
casted_array_values = _c(array.values, feature[0])
|
| 1990 |
+
if casted_array_values.type == array.values.type:
|
| 1991 |
+
return array
|
| 1992 |
+
else:
|
| 1993 |
+
# Merge offsets with the null bitmap to avoid the "Null bitmap with offsets slice not supported" ArrowNotImplementedError
|
| 1994 |
+
array_offsets = _combine_list_array_offsets_with_mask(array)
|
| 1995 |
+
return pa.ListArray.from_arrays(array_offsets, casted_array_values)
|
| 1996 |
+
elif isinstance(feature, Sequence):
|
| 1997 |
+
if feature.length > -1:
|
| 1998 |
+
if _are_list_values_of_length(array, feature.length):
|
| 1999 |
+
if array.null_count > 0:
|
| 2000 |
+
# Ensure each null value in the array translates to [null] * pa_type.list_size in the array's values array
|
| 2001 |
+
array_type = array.type
|
| 2002 |
+
storage_type = _storage_type(array_type)
|
| 2003 |
+
if array_type != storage_type:
|
| 2004 |
+
# Temporarily convert to the storage type to support extension types in the slice operation
|
| 2005 |
+
array = array_cast(array, storage_type, allow_number_to_str=allow_number_to_str)
|
| 2006 |
+
array = pc.list_slice(array, 0, feature.length, return_fixed_size_list=True)
|
| 2007 |
+
array = array_cast(array, array_type, allow_number_to_str=allow_number_to_str)
|
| 2008 |
+
else:
|
| 2009 |
+
array = pc.list_slice(array, 0, feature.length, return_fixed_size_list=True)
|
| 2010 |
+
array_values = array.values
|
| 2011 |
+
casted_array_values = _c(array_values, feature.feature)
|
| 2012 |
+
if config.PYARROW_VERSION.major < 15:
|
| 2013 |
+
return pa.Array.from_buffers(
|
| 2014 |
+
pa.list_(casted_array_values.type, feature.length),
|
| 2015 |
+
len(array),
|
| 2016 |
+
[array.is_valid().buffers()[1]],
|
| 2017 |
+
children=[casted_array_values],
|
| 2018 |
+
)
|
| 2019 |
+
else:
|
| 2020 |
+
return pa.FixedSizeListArray.from_arrays(
|
| 2021 |
+
casted_array_values, feature.length, mask=array.is_null()
|
| 2022 |
+
)
|
| 2023 |
+
else:
|
| 2024 |
+
array_values = array.values[
|
| 2025 |
+
array.offset * feature.length : (array.offset + len(array)) * feature.length
|
| 2026 |
+
]
|
| 2027 |
+
return pa.FixedSizeListArray.from_arrays(_c(array_values, feature.feature), feature.length)
|
| 2028 |
+
else:
|
| 2029 |
+
casted_array_values = _c(array.values, feature.feature)
|
| 2030 |
+
if casted_array_values.type == array.values.type:
|
| 2031 |
+
return array
|
| 2032 |
+
else:
|
| 2033 |
+
# Merge offsets with the null bitmap to avoid the "Null bitmap with offsets slice not supported" ArrowNotImplementedError
|
| 2034 |
+
array_offsets = _combine_list_array_offsets_with_mask(array)
|
| 2035 |
+
return pa.ListArray.from_arrays(array_offsets, casted_array_values)
|
| 2036 |
+
elif pa.types.is_fixed_size_list(array.type):
|
| 2037 |
+
# feature must be either [subfeature] or Sequence(subfeature)
|
| 2038 |
+
if isinstance(feature, list):
|
| 2039 |
+
array_offsets = (np.arange(len(array) + 1) + array.offset) * array.type.list_size
|
| 2040 |
+
return pa.ListArray.from_arrays(array_offsets, _c(array.values, feature[0]), mask=array.is_null())
|
| 2041 |
+
elif isinstance(feature, Sequence):
|
| 2042 |
+
if feature.length > -1:
|
| 2043 |
+
if feature.length == array.type.list_size:
|
| 2044 |
+
array_values = array.values[
|
| 2045 |
+
array.offset * array.type.list_size : (array.offset + len(array)) * array.type.list_size
|
| 2046 |
+
]
|
| 2047 |
+
casted_array_values = _c(array_values, feature.feature)
|
| 2048 |
+
if config.PYARROW_VERSION.major < 15:
|
| 2049 |
+
return pa.Array.from_buffers(
|
| 2050 |
+
pa.list_(casted_array_values.type, feature.length),
|
| 2051 |
+
len(array),
|
| 2052 |
+
[array.is_valid().buffers()[1]],
|
| 2053 |
+
children=[casted_array_values],
|
| 2054 |
+
)
|
| 2055 |
+
else:
|
| 2056 |
+
return pa.FixedSizeListArray.from_arrays(
|
| 2057 |
+
casted_array_values, feature.length, mask=array.is_null()
|
| 2058 |
+
)
|
| 2059 |
+
else:
|
| 2060 |
+
array_offsets = (np.arange(len(array) + 1) + array.offset) * array.type.list_size
|
| 2061 |
+
return pa.ListArray.from_arrays(array_offsets, _c(array.values, feature.feature), mask=array.is_null())
|
| 2062 |
+
if pa.types.is_null(array.type):
|
| 2063 |
+
return array_cast(array, get_nested_type(feature), allow_number_to_str=allow_number_to_str)
|
| 2064 |
+
elif not isinstance(feature, (Sequence, dict, list, tuple)):
|
| 2065 |
+
return array_cast(array, feature(), allow_number_to_str=allow_number_to_str)
|
| 2066 |
+
raise TypeError(f"Couldn't cast array of type\n{array.type}\nto\n{feature}")
|
| 2067 |
+
|
| 2068 |
+
|
| 2069 |
+
@_wrap_for_chunked_arrays
|
| 2070 |
+
def embed_array_storage(array: pa.Array, feature: "FeatureType"):
|
| 2071 |
+
"""Embed data into an arrays's storage.
|
| 2072 |
+
For custom features like Audio or Image, it takes into account the "embed_storage" methods
|
| 2073 |
+
they define to embed external data (e.g. an image file) into an array.
|
| 2074 |
+
|
| 2075 |
+
<Added version="2.4.0"/>
|
| 2076 |
+
|
| 2077 |
+
Args:
|
| 2078 |
+
array (`pa.Array`):
|
| 2079 |
+
The PyArrow array in which to embed data.
|
| 2080 |
+
feature (`datasets.features.FeatureType`):
|
| 2081 |
+
Array features.
|
| 2082 |
+
|
| 2083 |
+
Raises:
|
| 2084 |
+
`TypeError`: if the target type is not supported according, e.g.
|
| 2085 |
+
|
| 2086 |
+
- if a field is missing
|
| 2087 |
+
|
| 2088 |
+
Returns:
|
| 2089 |
+
array (`pyarrow.Array`): the casted array
|
| 2090 |
+
"""
|
| 2091 |
+
from .features import Sequence
|
| 2092 |
+
|
| 2093 |
+
_e = embed_array_storage
|
| 2094 |
+
|
| 2095 |
+
if isinstance(array, pa.ExtensionArray):
|
| 2096 |
+
array = array.storage
|
| 2097 |
+
if hasattr(feature, "embed_storage"):
|
| 2098 |
+
return feature.embed_storage(array)
|
| 2099 |
+
elif pa.types.is_struct(array.type):
|
| 2100 |
+
# feature must be a dict or Sequence(subfeatures_dict)
|
| 2101 |
+
if isinstance(feature, Sequence) and isinstance(feature.feature, dict):
|
| 2102 |
+
feature = {
|
| 2103 |
+
name: Sequence(subfeature, length=feature.length) for name, subfeature in feature.feature.items()
|
| 2104 |
+
}
|
| 2105 |
+
if isinstance(feature, dict):
|
| 2106 |
+
arrays = [_e(array.field(name), subfeature) for name, subfeature in feature.items()]
|
| 2107 |
+
return pa.StructArray.from_arrays(arrays, names=list(feature), mask=array.is_null())
|
| 2108 |
+
elif pa.types.is_list(array.type):
|
| 2109 |
+
# feature must be either [subfeature] or Sequence(subfeature)
|
| 2110 |
+
# Merge offsets with the null bitmap to avoid the "Null bitmap with offsets slice not supported" ArrowNotImplementedError
|
| 2111 |
+
array_offsets = _combine_list_array_offsets_with_mask(array)
|
| 2112 |
+
if isinstance(feature, list):
|
| 2113 |
+
return pa.ListArray.from_arrays(array_offsets, _e(array.values, feature[0]))
|
| 2114 |
+
if isinstance(feature, Sequence) and feature.length == -1:
|
| 2115 |
+
return pa.ListArray.from_arrays(array_offsets, _e(array.values, feature.feature))
|
| 2116 |
+
elif pa.types.is_fixed_size_list(array.type):
|
| 2117 |
+
# feature must be Sequence(subfeature)
|
| 2118 |
+
if isinstance(feature, Sequence) and feature.length > -1:
|
| 2119 |
+
array_values = array.values[
|
| 2120 |
+
array.offset * array.type.list_size : (array.offset + len(array)) * array.type.list_size
|
| 2121 |
+
]
|
| 2122 |
+
embedded_array_values = _e(array_values, feature.feature)
|
| 2123 |
+
if config.PYARROW_VERSION.major < 15:
|
| 2124 |
+
return pa.Array.from_buffers(
|
| 2125 |
+
pa.list_(array_values.type, feature.length),
|
| 2126 |
+
len(array),
|
| 2127 |
+
[array.is_valid().buffers()[1]],
|
| 2128 |
+
children=[embedded_array_values],
|
| 2129 |
+
)
|
| 2130 |
+
else:
|
| 2131 |
+
return pa.FixedSizeListArray.from_arrays(embedded_array_values, feature.length, mask=array.is_null())
|
| 2132 |
+
if not isinstance(feature, (Sequence, dict, list, tuple)):
|
| 2133 |
+
return array
|
| 2134 |
+
raise TypeError(f"Couldn't embed array of type\n{array.type}\nwith\n{feature}")
|
| 2135 |
+
|
| 2136 |
+
|
| 2137 |
+
class CastError(ValueError):
|
| 2138 |
+
"""When it's not possible to cast an Arrow table to a specific schema or set of features"""
|
| 2139 |
+
|
| 2140 |
+
def __init__(self, *args, table_column_names: List[str], requested_column_names: List[str]) -> None:
|
| 2141 |
+
super().__init__(*args)
|
| 2142 |
+
self.table_column_names = table_column_names
|
| 2143 |
+
self.requested_column_names = requested_column_names
|
| 2144 |
+
|
| 2145 |
+
def details(self):
|
| 2146 |
+
new_columns = set(self.table_column_names) - set(self.requested_column_names)
|
| 2147 |
+
missing_columns = set(self.requested_column_names) - set(self.table_column_names)
|
| 2148 |
+
if new_columns and missing_columns:
|
| 2149 |
+
return f"there are {len(new_columns)} new columns ({', '.join(new_columns)}) and {len(missing_columns)} missing columns ({', '.join(missing_columns)})."
|
| 2150 |
+
elif new_columns:
|
| 2151 |
+
return f"there are {len(new_columns)} new columns ({new_columns})"
|
| 2152 |
+
else:
|
| 2153 |
+
return f"there are {len(missing_columns)} missing columns ({missing_columns})"
|
| 2154 |
+
|
| 2155 |
+
|
| 2156 |
+
def cast_table_to_features(table: pa.Table, features: "Features"):
|
| 2157 |
+
"""Cast a table to the arrow schema that corresponds to the requested features.
|
| 2158 |
+
|
| 2159 |
+
Args:
|
| 2160 |
+
table (`pyarrow.Table`):
|
| 2161 |
+
PyArrow table to cast.
|
| 2162 |
+
features ([`Features`]):
|
| 2163 |
+
Target features.
|
| 2164 |
+
|
| 2165 |
+
Returns:
|
| 2166 |
+
table (`pyarrow.Table`): the casted table
|
| 2167 |
+
"""
|
| 2168 |
+
if sorted(table.column_names) != sorted(features):
|
| 2169 |
+
raise CastError(
|
| 2170 |
+
f"Couldn't cast\n{table.schema}\nto\n{features}\nbecause column names don't match",
|
| 2171 |
+
table_column_names=table.column_names,
|
| 2172 |
+
requested_column_names=list(features),
|
| 2173 |
+
)
|
| 2174 |
+
arrays = [cast_array_to_feature(table[name], feature) for name, feature in features.items()]
|
| 2175 |
+
return pa.Table.from_arrays(arrays, schema=features.arrow_schema)
|
| 2176 |
+
|
| 2177 |
+
|
| 2178 |
+
def cast_table_to_schema(table: pa.Table, schema: pa.Schema):
|
| 2179 |
+
"""Cast a table to the arrow schema. Different from `cast_table_to_features`, this method can preserve nullability.
|
| 2180 |
+
|
| 2181 |
+
Args:
|
| 2182 |
+
table (`pa.Table`):
|
| 2183 |
+
PyArrow table to cast.
|
| 2184 |
+
features ([`Features`]):
|
| 2185 |
+
Target features.
|
| 2186 |
+
|
| 2187 |
+
Returns:
|
| 2188 |
+
`pa.Table`: the casted table
|
| 2189 |
+
"""
|
| 2190 |
+
from .features import Features
|
| 2191 |
+
|
| 2192 |
+
features = Features.from_arrow_schema(schema)
|
| 2193 |
+
if sorted(table.column_names) != sorted(features):
|
| 2194 |
+
raise CastError(
|
| 2195 |
+
f"Couldn't cast\n{table.schema}\nto\n{features}\nbecause column names don't match",
|
| 2196 |
+
table_column_names=table.column_names,
|
| 2197 |
+
requested_column_names=list(features),
|
| 2198 |
+
)
|
| 2199 |
+
arrays = [cast_array_to_feature(table[name], feature) for name, feature in features.items()]
|
| 2200 |
+
return pa.Table.from_arrays(arrays, schema=schema)
|
| 2201 |
+
|
| 2202 |
+
|
| 2203 |
+
def embed_table_storage(table: pa.Table):
|
| 2204 |
+
"""Embed external data into a table's storage.
|
| 2205 |
+
|
| 2206 |
+
<Added version="2.4.0"/>
|
| 2207 |
+
|
| 2208 |
+
Args:
|
| 2209 |
+
table (`pyarrow.Table`):
|
| 2210 |
+
PyArrow table in which to embed data.
|
| 2211 |
+
|
| 2212 |
+
Returns:
|
| 2213 |
+
table (`pyarrow.Table`): the table with embedded data
|
| 2214 |
+
"""
|
| 2215 |
+
from .features.features import Features, require_storage_embed
|
| 2216 |
+
|
| 2217 |
+
features = Features.from_arrow_schema(table.schema)
|
| 2218 |
+
arrays = [
|
| 2219 |
+
embed_array_storage(table[name], feature) if require_storage_embed(feature) else table[name]
|
| 2220 |
+
for name, feature in features.items()
|
| 2221 |
+
]
|
| 2222 |
+
return pa.Table.from_arrays(arrays, schema=features.arrow_schema)
|
| 2223 |
+
|
| 2224 |
+
|
| 2225 |
+
def table_cast(table: pa.Table, schema: pa.Schema):
|
| 2226 |
+
"""Improved version of `pa.Table.cast`.
|
| 2227 |
+
|
| 2228 |
+
It supports casting to feature types stored in the schema metadata.
|
| 2229 |
+
|
| 2230 |
+
Args:
|
| 2231 |
+
table (`pyarrow.Table`):
|
| 2232 |
+
PyArrow table to cast.
|
| 2233 |
+
schema (`pyarrow.Schema`):
|
| 2234 |
+
Target PyArrow schema.
|
| 2235 |
+
|
| 2236 |
+
Returns:
|
| 2237 |
+
table (`pyarrow.Table`): the casted table
|
| 2238 |
+
"""
|
| 2239 |
+
if table.schema != schema:
|
| 2240 |
+
return cast_table_to_schema(table, schema)
|
| 2241 |
+
elif table.schema.metadata != schema.metadata:
|
| 2242 |
+
return table.replace_schema_metadata(schema.metadata)
|
| 2243 |
+
else:
|
| 2244 |
+
return table
|
| 2245 |
+
|
| 2246 |
+
|
| 2247 |
+
def table_flatten(table: pa.Table):
|
| 2248 |
+
"""Improved version of `pa.Table.flatten`.
|
| 2249 |
+
|
| 2250 |
+
It behaves as `pa.Table.flatten` in a sense it does 1-step flatten of the columns with a struct type into one column per struct field,
|
| 2251 |
+
but updates the metadata and skips decodable features unless the `decode` attribute of these features is set to False.
|
| 2252 |
+
|
| 2253 |
+
Args:
|
| 2254 |
+
table (`pa.Table`):
|
| 2255 |
+
PyArrow table to flatten.
|
| 2256 |
+
|
| 2257 |
+
Returns:
|
| 2258 |
+
`Table`: the flattened table
|
| 2259 |
+
"""
|
| 2260 |
+
from .features import Features
|
| 2261 |
+
|
| 2262 |
+
features = Features.from_arrow_schema(table.schema)
|
| 2263 |
+
if any(hasattr(subfeature, "flatten") and subfeature.flatten() == subfeature for subfeature in features.values()):
|
| 2264 |
+
flat_arrays = []
|
| 2265 |
+
flat_column_names = []
|
| 2266 |
+
for field in table.schema:
|
| 2267 |
+
array = table.column(field.name)
|
| 2268 |
+
subfeature = features[field.name]
|
| 2269 |
+
if pa.types.is_struct(field.type) and (
|
| 2270 |
+
not hasattr(subfeature, "flatten") or subfeature.flatten() != subfeature
|
| 2271 |
+
):
|
| 2272 |
+
flat_arrays.extend(array.flatten())
|
| 2273 |
+
flat_column_names.extend([f"{field.name}.{subfield.name}" for subfield in field.type])
|
| 2274 |
+
else:
|
| 2275 |
+
flat_arrays.append(array)
|
| 2276 |
+
flat_column_names.append(field.name)
|
| 2277 |
+
flat_table = pa.Table.from_arrays(
|
| 2278 |
+
flat_arrays,
|
| 2279 |
+
names=flat_column_names,
|
| 2280 |
+
)
|
| 2281 |
+
else:
|
| 2282 |
+
flat_table = table.flatten()
|
| 2283 |
+
# Preserve complex types in the metadata
|
| 2284 |
+
flat_features = features.flatten(max_depth=2)
|
| 2285 |
+
flat_features = Features({column_name: flat_features[column_name] for column_name in flat_table.column_names})
|
| 2286 |
+
return flat_table.replace_schema_metadata(flat_features.arrow_schema.metadata)
|
| 2287 |
+
|
| 2288 |
+
|
| 2289 |
+
def table_visitor(table: pa.Table, function: Callable[[pa.Array], None]):
|
| 2290 |
+
"""Visit all arrays in a table and apply a function to them.
|
| 2291 |
+
|
| 2292 |
+
Args:
|
| 2293 |
+
table (`pyarrow.Table`):
|
| 2294 |
+
PyArrow table to visit.
|
| 2295 |
+
function (`Callable[[pa.Array], None]`):
|
| 2296 |
+
Function to apply to each array.
|
| 2297 |
+
"""
|
| 2298 |
+
from .features import Features, Sequence
|
| 2299 |
+
|
| 2300 |
+
features = Features.from_arrow_schema(table.schema)
|
| 2301 |
+
|
| 2302 |
+
def _visit(array, feature):
|
| 2303 |
+
if isinstance(array, pa.ChunkedArray):
|
| 2304 |
+
for chunk in array.chunks:
|
| 2305 |
+
_visit(chunk, feature)
|
| 2306 |
+
else:
|
| 2307 |
+
if isinstance(array, pa.ExtensionArray):
|
| 2308 |
+
array = array.storage
|
| 2309 |
+
function(array, feature)
|
| 2310 |
+
if pa.types.is_struct(array.type) and not hasattr(feature, "cast_storage"):
|
| 2311 |
+
if isinstance(feature, Sequence) and isinstance(feature.feature, dict):
|
| 2312 |
+
feature = {
|
| 2313 |
+
name: Sequence(subfeature, length=feature.length)
|
| 2314 |
+
for name, subfeature in feature.feature.items()
|
| 2315 |
+
}
|
| 2316 |
+
for name, subfeature in feature.items():
|
| 2317 |
+
_visit(array.field(name), subfeature)
|
| 2318 |
+
elif pa.types.is_list(array.type):
|
| 2319 |
+
if isinstance(feature, list):
|
| 2320 |
+
_visit(array.values, feature[0])
|
| 2321 |
+
elif isinstance(feature, Sequence):
|
| 2322 |
+
_visit(array.values, feature.feature)
|
| 2323 |
+
|
| 2324 |
+
for name, feature in features.items():
|
| 2325 |
+
_visit(table[name], feature)
|
| 2326 |
+
|
| 2327 |
+
|
| 2328 |
+
def table_iter(table: Table, batch_size: int, drop_last_batch=False) -> Iterator[pa.Table]:
|
| 2329 |
+
"""Iterate over sub-tables of size `batch_size`.
|
| 2330 |
+
|
| 2331 |
+
Args:
|
| 2332 |
+
table (`pyarrow.Table`):
|
| 2333 |
+
PyArrow table to iterate over.
|
| 2334 |
+
batch_size (`int`):
|
| 2335 |
+
Size of each sub-table to yield.
|
| 2336 |
+
drop_last_batch (`bool`, defaults to `False`):
|
| 2337 |
+
Drop the last batch if it is smaller than `batch_size`.
|
| 2338 |
+
"""
|
| 2339 |
+
chunks_buffer = []
|
| 2340 |
+
chunks_buffer_size = 0
|
| 2341 |
+
for chunk in table.to_reader(max_chunksize=batch_size):
|
| 2342 |
+
if len(chunk) == 0:
|
| 2343 |
+
continue
|
| 2344 |
+
elif chunks_buffer_size + len(chunk) < batch_size:
|
| 2345 |
+
chunks_buffer.append(chunk)
|
| 2346 |
+
chunks_buffer_size += len(chunk)
|
| 2347 |
+
continue
|
| 2348 |
+
elif chunks_buffer_size + len(chunk) == batch_size:
|
| 2349 |
+
chunks_buffer.append(chunk)
|
| 2350 |
+
yield pa.Table.from_batches(chunks_buffer)
|
| 2351 |
+
chunks_buffer = []
|
| 2352 |
+
chunks_buffer_size = 0
|
| 2353 |
+
else:
|
| 2354 |
+
cropped_chunk_length = batch_size - chunks_buffer_size
|
| 2355 |
+
chunks_buffer.append(chunk.slice(0, cropped_chunk_length))
|
| 2356 |
+
yield pa.Table.from_batches(chunks_buffer)
|
| 2357 |
+
chunks_buffer = [chunk.slice(cropped_chunk_length, len(chunk) - cropped_chunk_length)]
|
| 2358 |
+
chunks_buffer_size = len(chunk) - cropped_chunk_length
|
| 2359 |
+
if not drop_last_batch and chunks_buffer:
|
| 2360 |
+
yield pa.Table.from_batches(chunks_buffer)
|
evalkit_tf437/lib/python3.10/site-packages/datasets/utils/_datasets_server.py
ADDED
|
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from typing import Any, Dict, List, Optional, Union
|
| 2 |
+
|
| 3 |
+
from .. import config
|
| 4 |
+
from ..exceptions import DatasetsError
|
| 5 |
+
from .file_utils import (
|
| 6 |
+
get_authentication_headers_for_url,
|
| 7 |
+
http_get,
|
| 8 |
+
)
|
| 9 |
+
from .logging import get_logger
|
| 10 |
+
|
| 11 |
+
|
| 12 |
+
logger = get_logger(__name__)
|
| 13 |
+
|
| 14 |
+
|
| 15 |
+
class DatasetsServerError(DatasetsError):
|
| 16 |
+
"""Dataset-server error.
|
| 17 |
+
|
| 18 |
+
Raised when trying to use the Datasets-server HTTP API and when trying to access:
|
| 19 |
+
- a missing dataset, or
|
| 20 |
+
- a private/gated dataset and the user is not authenticated.
|
| 21 |
+
- unavailable /parquet or /info responses
|
| 22 |
+
"""
|
| 23 |
+
|
| 24 |
+
|
| 25 |
+
def get_exported_parquet_files(dataset: str, revision: str, token: Optional[Union[str, bool]]) -> List[Dict[str, Any]]:
|
| 26 |
+
"""
|
| 27 |
+
Get the dataset exported parquet files
|
| 28 |
+
Docs: https://huggingface.co/docs/datasets-server/parquet
|
| 29 |
+
"""
|
| 30 |
+
datasets_server_parquet_url = config.HF_ENDPOINT.replace("://", "://datasets-server.") + "/parquet?dataset="
|
| 31 |
+
try:
|
| 32 |
+
parquet_data_files_response = http_get(
|
| 33 |
+
url=datasets_server_parquet_url + dataset,
|
| 34 |
+
temp_file=None,
|
| 35 |
+
headers=get_authentication_headers_for_url(config.HF_ENDPOINT + f"datasets/{dataset}", token=token),
|
| 36 |
+
timeout=100.0,
|
| 37 |
+
max_retries=3,
|
| 38 |
+
)
|
| 39 |
+
parquet_data_files_response.raise_for_status()
|
| 40 |
+
if "X-Revision" in parquet_data_files_response.headers:
|
| 41 |
+
if parquet_data_files_response.headers["X-Revision"] == revision or revision is None:
|
| 42 |
+
parquet_data_files_response_json = parquet_data_files_response.json()
|
| 43 |
+
if (
|
| 44 |
+
parquet_data_files_response_json.get("partial") is False
|
| 45 |
+
and not parquet_data_files_response_json.get("pending", True)
|
| 46 |
+
and not parquet_data_files_response_json.get("failed", True)
|
| 47 |
+
and "parquet_files" in parquet_data_files_response_json
|
| 48 |
+
):
|
| 49 |
+
return parquet_data_files_response_json["parquet_files"]
|
| 50 |
+
else:
|
| 51 |
+
logger.debug(f"Parquet export for {dataset} is not completely ready yet.")
|
| 52 |
+
else:
|
| 53 |
+
logger.debug(
|
| 54 |
+
f"Parquet export for {dataset} is available but outdated (revision='{parquet_data_files_response.headers['X-Revision']}')"
|
| 55 |
+
)
|
| 56 |
+
except Exception as e: # noqa catch any exception of the datasets-server and consider the parquet export doesn't exist
|
| 57 |
+
logger.debug(f"No parquet export for {dataset} available ({type(e).__name__}: {e})")
|
| 58 |
+
raise DatasetsServerError("No exported Parquet files available.")
|
| 59 |
+
|
| 60 |
+
|
| 61 |
+
def get_exported_dataset_infos(
|
| 62 |
+
dataset: str, revision: str, token: Optional[Union[str, bool]]
|
| 63 |
+
) -> Dict[str, Dict[str, Any]]:
|
| 64 |
+
"""
|
| 65 |
+
Get the dataset information, can be useful to get e.g. the dataset features.
|
| 66 |
+
Docs: https://huggingface.co/docs/datasets-server/info
|
| 67 |
+
"""
|
| 68 |
+
datasets_server_info_url = config.HF_ENDPOINT.replace("://", "://datasets-server.") + "/info?dataset="
|
| 69 |
+
try:
|
| 70 |
+
info_response = http_get(
|
| 71 |
+
url=datasets_server_info_url + dataset,
|
| 72 |
+
temp_file=None,
|
| 73 |
+
headers=get_authentication_headers_for_url(config.HF_ENDPOINT + f"datasets/{dataset}", token=token),
|
| 74 |
+
timeout=100.0,
|
| 75 |
+
max_retries=3,
|
| 76 |
+
)
|
| 77 |
+
info_response.raise_for_status()
|
| 78 |
+
if "X-Revision" in info_response.headers:
|
| 79 |
+
if info_response.headers["X-Revision"] == revision or revision is None:
|
| 80 |
+
info_response = info_response.json()
|
| 81 |
+
if (
|
| 82 |
+
info_response.get("partial") is False
|
| 83 |
+
and not info_response.get("pending", True)
|
| 84 |
+
and not info_response.get("failed", True)
|
| 85 |
+
and "dataset_info" in info_response
|
| 86 |
+
):
|
| 87 |
+
return info_response["dataset_info"]
|
| 88 |
+
else:
|
| 89 |
+
logger.debug(f"Dataset info for {dataset} is not completely ready yet.")
|
| 90 |
+
else:
|
| 91 |
+
logger.debug(
|
| 92 |
+
f"Dataset info for {dataset} is available but outdated (revision='{info_response.headers['X-Revision']}')"
|
| 93 |
+
)
|
| 94 |
+
except Exception as e: # noqa catch any exception of the datasets-server and consider the dataset info doesn't exist
|
| 95 |
+
logger.debug(f"No dataset info for {dataset} available ({type(e).__name__}: {e})")
|
| 96 |
+
raise DatasetsServerError("No exported dataset infos available.")
|
evalkit_tf437/lib/python3.10/site-packages/datasets/utils/download_manager.py
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
# deprecated, please use datasets.download.download_manager
|
evalkit_tf437/lib/python3.10/site-packages/datasets/utils/logging.py
ADDED
|
@@ -0,0 +1,179 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright 2020 Optuna, Hugging Face
|
| 2 |
+
#
|
| 3 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
| 4 |
+
# you may not use this file except in compliance with the License.
|
| 5 |
+
# You may obtain a copy of the License at
|
| 6 |
+
#
|
| 7 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
| 8 |
+
#
|
| 9 |
+
# Unless required by applicable law or agreed to in writing, software
|
| 10 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
| 11 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| 12 |
+
# See the License for the specific language governing permissions and
|
| 13 |
+
# limitations under the License.
|
| 14 |
+
"""Logging utilities."""
|
| 15 |
+
|
| 16 |
+
import logging
|
| 17 |
+
import os
|
| 18 |
+
from logging import (
|
| 19 |
+
CRITICAL, # NOQA
|
| 20 |
+
DEBUG, # NOQA
|
| 21 |
+
ERROR, # NOQA
|
| 22 |
+
FATAL, # NOQA
|
| 23 |
+
INFO, # NOQA
|
| 24 |
+
NOTSET, # NOQA
|
| 25 |
+
WARN, # NOQA
|
| 26 |
+
WARNING, # NOQA
|
| 27 |
+
)
|
| 28 |
+
from typing import Optional
|
| 29 |
+
|
| 30 |
+
from .tqdm import ( # noqa: F401 # imported for backward compatibility
|
| 31 |
+
disable_progress_bar,
|
| 32 |
+
enable_progress_bar,
|
| 33 |
+
is_progress_bar_enabled,
|
| 34 |
+
tqdm,
|
| 35 |
+
)
|
| 36 |
+
|
| 37 |
+
|
| 38 |
+
log_levels = {
|
| 39 |
+
"debug": logging.DEBUG,
|
| 40 |
+
"info": logging.INFO,
|
| 41 |
+
"warning": logging.WARNING,
|
| 42 |
+
"error": logging.ERROR,
|
| 43 |
+
"critical": logging.CRITICAL,
|
| 44 |
+
}
|
| 45 |
+
|
| 46 |
+
_default_log_level = logging.WARNING
|
| 47 |
+
|
| 48 |
+
|
| 49 |
+
def _get_default_logging_level():
|
| 50 |
+
"""
|
| 51 |
+
If DATASETS_VERBOSITY env var is set to one of the valid choices return that as the new default level.
|
| 52 |
+
If it is not - fall back to ``_default_log_level``
|
| 53 |
+
"""
|
| 54 |
+
env_level_str = os.getenv("DATASETS_VERBOSITY", None)
|
| 55 |
+
if env_level_str:
|
| 56 |
+
if env_level_str in log_levels:
|
| 57 |
+
return log_levels[env_level_str]
|
| 58 |
+
else:
|
| 59 |
+
logging.getLogger().warning(
|
| 60 |
+
f"Unknown option DATASETS_VERBOSITY={env_level_str}, "
|
| 61 |
+
f"has to be one of: { ', '.join(log_levels.keys()) }"
|
| 62 |
+
)
|
| 63 |
+
return _default_log_level
|
| 64 |
+
|
| 65 |
+
|
| 66 |
+
def _get_library_name() -> str:
|
| 67 |
+
return __name__.split(".")[0]
|
| 68 |
+
|
| 69 |
+
|
| 70 |
+
def _get_library_root_logger() -> logging.Logger:
|
| 71 |
+
return logging.getLogger(_get_library_name())
|
| 72 |
+
|
| 73 |
+
|
| 74 |
+
def _configure_library_root_logger() -> None:
|
| 75 |
+
# Apply our default configuration to the library root logger.
|
| 76 |
+
library_root_logger = _get_library_root_logger()
|
| 77 |
+
library_root_logger.addHandler(logging.StreamHandler())
|
| 78 |
+
library_root_logger.setLevel(_get_default_logging_level())
|
| 79 |
+
|
| 80 |
+
|
| 81 |
+
def _reset_library_root_logger() -> None:
|
| 82 |
+
library_root_logger = _get_library_root_logger()
|
| 83 |
+
library_root_logger.setLevel(logging.NOTSET)
|
| 84 |
+
|
| 85 |
+
|
| 86 |
+
def get_logger(name: Optional[str] = None) -> logging.Logger:
|
| 87 |
+
"""Return a logger with the specified name.
|
| 88 |
+
This function can be used in dataset scripts.
|
| 89 |
+
"""
|
| 90 |
+
if name is None:
|
| 91 |
+
name = _get_library_name()
|
| 92 |
+
return logging.getLogger(name)
|
| 93 |
+
|
| 94 |
+
|
| 95 |
+
def get_verbosity() -> int:
|
| 96 |
+
"""Return the current level for the HuggingFace datasets library's root logger.
|
| 97 |
+
Returns:
|
| 98 |
+
Logging level, e.g., `datasets.logging.DEBUG` and `datasets.logging.INFO`.
|
| 99 |
+
|
| 100 |
+
<Tip>
|
| 101 |
+
|
| 102 |
+
HuggingFace datasets library has following logging levels:
|
| 103 |
+
- `datasets.logging.CRITICAL`, `datasets.logging.FATAL`
|
| 104 |
+
- `datasets.logging.ERROR`
|
| 105 |
+
- `datasets.logging.WARNING`, `datasets.logging.WARN`
|
| 106 |
+
- `datasets.logging.INFO`
|
| 107 |
+
- `datasets.logging.DEBUG`
|
| 108 |
+
|
| 109 |
+
</Tip>
|
| 110 |
+
"""
|
| 111 |
+
return _get_library_root_logger().getEffectiveLevel()
|
| 112 |
+
|
| 113 |
+
|
| 114 |
+
def set_verbosity(verbosity: int) -> None:
|
| 115 |
+
"""Set the level for the Hugging Face Datasets library's root logger.
|
| 116 |
+
Args:
|
| 117 |
+
verbosity:
|
| 118 |
+
Logging level, e.g., `datasets.logging.DEBUG` and `datasets.logging.INFO`.
|
| 119 |
+
"""
|
| 120 |
+
_get_library_root_logger().setLevel(verbosity)
|
| 121 |
+
|
| 122 |
+
|
| 123 |
+
def set_verbosity_info():
|
| 124 |
+
"""Set the level for the Hugging Face datasets library's root logger to `INFO`.
|
| 125 |
+
|
| 126 |
+
This will display most of the logging information and tqdm bars.
|
| 127 |
+
|
| 128 |
+
Shortcut to `datasets.logging.set_verbosity(datasets.logging.INFO)`.
|
| 129 |
+
"""
|
| 130 |
+
return set_verbosity(INFO)
|
| 131 |
+
|
| 132 |
+
|
| 133 |
+
def set_verbosity_warning():
|
| 134 |
+
"""Set the level for the Hugging Face datasets library's root logger to `WARNING`.
|
| 135 |
+
|
| 136 |
+
This will display only the warning and errors logging information and tqdm bars.
|
| 137 |
+
|
| 138 |
+
Shortcut to `datasets.logging.set_verbosity(datasets.logging.WARNING)`.
|
| 139 |
+
"""
|
| 140 |
+
return set_verbosity(WARNING)
|
| 141 |
+
|
| 142 |
+
|
| 143 |
+
def set_verbosity_debug():
|
| 144 |
+
"""Set the level for the Hugging Face datasets library's root logger to `DEBUG`.
|
| 145 |
+
|
| 146 |
+
This will display all the logging information and tqdm bars.
|
| 147 |
+
|
| 148 |
+
Shortcut to `datasets.logging.set_verbosity(datasets.logging.DEBUG)`.
|
| 149 |
+
"""
|
| 150 |
+
return set_verbosity(DEBUG)
|
| 151 |
+
|
| 152 |
+
|
| 153 |
+
def set_verbosity_error():
|
| 154 |
+
"""Set the level for the Hugging Face datasets library's root logger to `ERROR`.
|
| 155 |
+
|
| 156 |
+
This will display only the errors logging information and tqdm bars.
|
| 157 |
+
|
| 158 |
+
Shortcut to `datasets.logging.set_verbosity(datasets.logging.ERROR)`.
|
| 159 |
+
"""
|
| 160 |
+
return set_verbosity(ERROR)
|
| 161 |
+
|
| 162 |
+
|
| 163 |
+
def disable_propagation() -> None:
|
| 164 |
+
"""Disable propagation of the library log outputs.
|
| 165 |
+
Note that log propagation is disabled by default.
|
| 166 |
+
"""
|
| 167 |
+
_get_library_root_logger().propagate = False
|
| 168 |
+
|
| 169 |
+
|
| 170 |
+
def enable_propagation() -> None:
|
| 171 |
+
"""Enable propagation of the library log outputs.
|
| 172 |
+
Please disable the Hugging Face datasets library's default handler to prevent double logging if the root logger has
|
| 173 |
+
been configured.
|
| 174 |
+
"""
|
| 175 |
+
_get_library_root_logger().propagate = True
|
| 176 |
+
|
| 177 |
+
|
| 178 |
+
# Configure the library root logger at the module level (singleton-like)
|
| 179 |
+
_configure_library_root_logger()
|
evalkit_tf437/lib/python3.10/site-packages/datasets/utils/metadata.py
ADDED
|
@@ -0,0 +1,320 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import re
|
| 2 |
+
import textwrap
|
| 3 |
+
from collections import Counter
|
| 4 |
+
from itertools import groupby
|
| 5 |
+
from operator import itemgetter
|
| 6 |
+
from pathlib import Path
|
| 7 |
+
from typing import Any, ClassVar, Dict, List, Optional, Tuple, Union
|
| 8 |
+
|
| 9 |
+
import yaml
|
| 10 |
+
from huggingface_hub import DatasetCardData
|
| 11 |
+
|
| 12 |
+
from ..config import METADATA_CONFIGS_FIELD
|
| 13 |
+
from ..info import DatasetInfo, DatasetInfosDict
|
| 14 |
+
from ..naming import _split_re
|
| 15 |
+
from ..utils.logging import get_logger
|
| 16 |
+
from .deprecation_utils import deprecated
|
| 17 |
+
|
| 18 |
+
|
| 19 |
+
logger = get_logger(__name__)
|
| 20 |
+
|
| 21 |
+
|
| 22 |
+
class _NoDuplicateSafeLoader(yaml.SafeLoader):
|
| 23 |
+
def _check_no_duplicates_on_constructed_node(self, node):
|
| 24 |
+
keys = [self.constructed_objects[key_node] for key_node, _ in node.value]
|
| 25 |
+
keys = [tuple(key) if isinstance(key, list) else key for key in keys]
|
| 26 |
+
counter = Counter(keys)
|
| 27 |
+
duplicate_keys = [key for key in counter if counter[key] > 1]
|
| 28 |
+
if duplicate_keys:
|
| 29 |
+
raise TypeError(f"Got duplicate yaml keys: {duplicate_keys}")
|
| 30 |
+
|
| 31 |
+
def construct_mapping(self, node, deep=False):
|
| 32 |
+
mapping = super().construct_mapping(node, deep=deep)
|
| 33 |
+
self._check_no_duplicates_on_constructed_node(node)
|
| 34 |
+
return mapping
|
| 35 |
+
|
| 36 |
+
|
| 37 |
+
def _split_yaml_from_readme(readme_content: str) -> Tuple[Optional[str], str]:
|
| 38 |
+
full_content = list(readme_content.splitlines())
|
| 39 |
+
if full_content and full_content[0] == "---" and "---" in full_content[1:]:
|
| 40 |
+
sep_idx = full_content[1:].index("---") + 1
|
| 41 |
+
yamlblock = "\n".join(full_content[1:sep_idx])
|
| 42 |
+
return yamlblock, "\n".join(full_content[sep_idx + 1 :])
|
| 43 |
+
|
| 44 |
+
return None, "\n".join(full_content)
|
| 45 |
+
|
| 46 |
+
|
| 47 |
+
@deprecated("Use `huggingface_hub.DatasetCardData` instead.")
|
| 48 |
+
class DatasetMetadata(dict):
|
| 49 |
+
# class attributes
|
| 50 |
+
_FIELDS_WITH_DASHES = {"train_eval_index"} # train-eval-index in the YAML metadata
|
| 51 |
+
|
| 52 |
+
@classmethod
|
| 53 |
+
def from_readme(cls, path: Union[Path, str]) -> "DatasetMetadata":
|
| 54 |
+
"""Loads and validates the dataset metadata from its dataset card (README.md)
|
| 55 |
+
|
| 56 |
+
Args:
|
| 57 |
+
path (:obj:`Path`): Path to the dataset card (its README.md file)
|
| 58 |
+
|
| 59 |
+
Returns:
|
| 60 |
+
:class:`DatasetMetadata`: The dataset's metadata
|
| 61 |
+
|
| 62 |
+
Raises:
|
| 63 |
+
:obj:`TypeError`: If the dataset's metadata is invalid
|
| 64 |
+
"""
|
| 65 |
+
with open(path, encoding="utf-8") as readme_file:
|
| 66 |
+
yaml_string, _ = _split_yaml_from_readme(readme_file.read())
|
| 67 |
+
if yaml_string is not None:
|
| 68 |
+
return cls.from_yaml_string(yaml_string)
|
| 69 |
+
else:
|
| 70 |
+
return cls()
|
| 71 |
+
|
| 72 |
+
def to_readme(self, path: Path):
|
| 73 |
+
if path.exists():
|
| 74 |
+
with open(path, encoding="utf-8") as readme_file:
|
| 75 |
+
readme_content = readme_file.read()
|
| 76 |
+
else:
|
| 77 |
+
readme_content = None
|
| 78 |
+
updated_readme_content = self._to_readme(readme_content)
|
| 79 |
+
with open(path, "w", encoding="utf-8") as readme_file:
|
| 80 |
+
readme_file.write(updated_readme_content)
|
| 81 |
+
|
| 82 |
+
def _to_readme(self, readme_content: Optional[str] = None) -> str:
|
| 83 |
+
if readme_content is not None:
|
| 84 |
+
_, content = _split_yaml_from_readme(readme_content)
|
| 85 |
+
full_content = "---\n" + self.to_yaml_string() + "---\n" + content
|
| 86 |
+
else:
|
| 87 |
+
full_content = "---\n" + self.to_yaml_string() + "---\n"
|
| 88 |
+
return full_content
|
| 89 |
+
|
| 90 |
+
@classmethod
|
| 91 |
+
def from_yaml_string(cls, string: str) -> "DatasetMetadata":
|
| 92 |
+
"""Loads and validates the dataset metadata from a YAML string
|
| 93 |
+
|
| 94 |
+
Args:
|
| 95 |
+
string (:obj:`str`): The YAML string
|
| 96 |
+
|
| 97 |
+
Returns:
|
| 98 |
+
:class:`DatasetMetadata`: The dataset's metadata
|
| 99 |
+
|
| 100 |
+
Raises:
|
| 101 |
+
:obj:`TypeError`: If the dataset's metadata is invalid
|
| 102 |
+
"""
|
| 103 |
+
metadata_dict = yaml.load(string, Loader=_NoDuplicateSafeLoader) or {}
|
| 104 |
+
|
| 105 |
+
# Convert the YAML keys to DatasetMetadata fields
|
| 106 |
+
metadata_dict = {
|
| 107 |
+
(key.replace("-", "_") if key.replace("-", "_") in cls._FIELDS_WITH_DASHES else key): value
|
| 108 |
+
for key, value in metadata_dict.items()
|
| 109 |
+
}
|
| 110 |
+
return cls(**metadata_dict)
|
| 111 |
+
|
| 112 |
+
def to_yaml_string(self) -> str:
|
| 113 |
+
return yaml.safe_dump(
|
| 114 |
+
{
|
| 115 |
+
(key.replace("_", "-") if key in self._FIELDS_WITH_DASHES else key): value
|
| 116 |
+
for key, value in self.items()
|
| 117 |
+
},
|
| 118 |
+
sort_keys=False,
|
| 119 |
+
allow_unicode=True,
|
| 120 |
+
encoding="utf-8",
|
| 121 |
+
).decode("utf-8")
|
| 122 |
+
|
| 123 |
+
|
| 124 |
+
class MetadataConfigs(Dict[str, Dict[str, Any]]):
|
| 125 |
+
"""Should be in format {config_name: {**config_params}}."""
|
| 126 |
+
|
| 127 |
+
FIELD_NAME: ClassVar[str] = METADATA_CONFIGS_FIELD
|
| 128 |
+
|
| 129 |
+
@staticmethod
|
| 130 |
+
def _raise_if_data_files_field_not_valid(metadata_config: dict):
|
| 131 |
+
yaml_data_files = metadata_config.get("data_files")
|
| 132 |
+
if yaml_data_files is not None:
|
| 133 |
+
yaml_error_message = textwrap.dedent(
|
| 134 |
+
f"""
|
| 135 |
+
Expected data_files in YAML to be either a string or a list of strings
|
| 136 |
+
or a list of dicts with two keys: 'split' and 'path', but got {yaml_data_files}
|
| 137 |
+
Examples of data_files in YAML:
|
| 138 |
+
|
| 139 |
+
data_files: data.csv
|
| 140 |
+
|
| 141 |
+
data_files: data/*.png
|
| 142 |
+
|
| 143 |
+
data_files:
|
| 144 |
+
- part0/*
|
| 145 |
+
- part1/*
|
| 146 |
+
|
| 147 |
+
data_files:
|
| 148 |
+
- split: train
|
| 149 |
+
path: train/*
|
| 150 |
+
- split: test
|
| 151 |
+
path: test/*
|
| 152 |
+
|
| 153 |
+
data_files:
|
| 154 |
+
- split: train
|
| 155 |
+
path:
|
| 156 |
+
- train/part1/*
|
| 157 |
+
- train/part2/*
|
| 158 |
+
- split: test
|
| 159 |
+
path: test/*
|
| 160 |
+
|
| 161 |
+
PS: some symbols like dashes '-' are not allowed in split names
|
| 162 |
+
"""
|
| 163 |
+
)
|
| 164 |
+
if not isinstance(yaml_data_files, (list, str)):
|
| 165 |
+
raise ValueError(yaml_error_message)
|
| 166 |
+
if isinstance(yaml_data_files, list):
|
| 167 |
+
for yaml_data_files_item in yaml_data_files:
|
| 168 |
+
if (
|
| 169 |
+
not isinstance(yaml_data_files_item, (str, dict))
|
| 170 |
+
or isinstance(yaml_data_files_item, dict)
|
| 171 |
+
and not (
|
| 172 |
+
len(yaml_data_files_item) == 2
|
| 173 |
+
and "split" in yaml_data_files_item
|
| 174 |
+
and re.match(_split_re, yaml_data_files_item["split"])
|
| 175 |
+
and isinstance(yaml_data_files_item.get("path"), (str, list))
|
| 176 |
+
)
|
| 177 |
+
):
|
| 178 |
+
raise ValueError(yaml_error_message)
|
| 179 |
+
|
| 180 |
+
@classmethod
|
| 181 |
+
def _from_exported_parquet_files_and_dataset_infos(
|
| 182 |
+
cls,
|
| 183 |
+
revision: str,
|
| 184 |
+
exported_parquet_files: List[Dict[str, Any]],
|
| 185 |
+
dataset_infos: DatasetInfosDict,
|
| 186 |
+
) -> "MetadataConfigs":
|
| 187 |
+
metadata_configs = {
|
| 188 |
+
config_name: {
|
| 189 |
+
"data_files": [
|
| 190 |
+
{
|
| 191 |
+
"split": split_name,
|
| 192 |
+
"path": [
|
| 193 |
+
parquet_file["url"].replace("refs%2Fconvert%2Fparquet", revision)
|
| 194 |
+
for parquet_file in parquet_files_for_split
|
| 195 |
+
],
|
| 196 |
+
}
|
| 197 |
+
for split_name, parquet_files_for_split in groupby(parquet_files_for_config, itemgetter("split"))
|
| 198 |
+
],
|
| 199 |
+
"version": str(dataset_infos.get(config_name, DatasetInfo()).version or "0.0.0"),
|
| 200 |
+
}
|
| 201 |
+
for config_name, parquet_files_for_config in groupby(exported_parquet_files, itemgetter("config"))
|
| 202 |
+
}
|
| 203 |
+
if dataset_infos:
|
| 204 |
+
# Preserve order of configs and splits
|
| 205 |
+
metadata_configs = {
|
| 206 |
+
config_name: {
|
| 207 |
+
"data_files": [
|
| 208 |
+
data_file
|
| 209 |
+
for split_name in dataset_info.splits
|
| 210 |
+
for data_file in metadata_configs[config_name]["data_files"]
|
| 211 |
+
if data_file["split"] == split_name
|
| 212 |
+
],
|
| 213 |
+
"version": metadata_configs[config_name]["version"],
|
| 214 |
+
}
|
| 215 |
+
for config_name, dataset_info in dataset_infos.items()
|
| 216 |
+
}
|
| 217 |
+
return cls(metadata_configs)
|
| 218 |
+
|
| 219 |
+
@classmethod
|
| 220 |
+
def from_dataset_card_data(cls, dataset_card_data: DatasetCardData) -> "MetadataConfigs":
|
| 221 |
+
if dataset_card_data.get(cls.FIELD_NAME):
|
| 222 |
+
metadata_configs = dataset_card_data[cls.FIELD_NAME]
|
| 223 |
+
if not isinstance(metadata_configs, list):
|
| 224 |
+
raise ValueError(f"Expected {cls.FIELD_NAME} to be a list, but got '{metadata_configs}'")
|
| 225 |
+
for metadata_config in metadata_configs:
|
| 226 |
+
if "config_name" not in metadata_config:
|
| 227 |
+
raise ValueError(
|
| 228 |
+
f"Each config must include `config_name` field with a string name of a config, "
|
| 229 |
+
f"but got {metadata_config}. "
|
| 230 |
+
)
|
| 231 |
+
cls._raise_if_data_files_field_not_valid(metadata_config)
|
| 232 |
+
return cls(
|
| 233 |
+
{
|
| 234 |
+
config["config_name"]: {param: value for param, value in config.items() if param != "config_name"}
|
| 235 |
+
for config in metadata_configs
|
| 236 |
+
}
|
| 237 |
+
)
|
| 238 |
+
return cls()
|
| 239 |
+
|
| 240 |
+
def to_dataset_card_data(self, dataset_card_data: DatasetCardData) -> None:
|
| 241 |
+
if self:
|
| 242 |
+
for metadata_config in self.values():
|
| 243 |
+
self._raise_if_data_files_field_not_valid(metadata_config)
|
| 244 |
+
current_metadata_configs = self.from_dataset_card_data(dataset_card_data)
|
| 245 |
+
total_metadata_configs = dict(sorted({**current_metadata_configs, **self}.items()))
|
| 246 |
+
for config_name, config_metadata in total_metadata_configs.items():
|
| 247 |
+
config_metadata.pop("config_name", None)
|
| 248 |
+
dataset_card_data[self.FIELD_NAME] = [
|
| 249 |
+
{"config_name": config_name, **config_metadata}
|
| 250 |
+
for config_name, config_metadata in total_metadata_configs.items()
|
| 251 |
+
]
|
| 252 |
+
|
| 253 |
+
def get_default_config_name(self) -> Optional[str]:
|
| 254 |
+
default_config_name = None
|
| 255 |
+
for config_name, metadata_config in self.items():
|
| 256 |
+
if len(self) == 1 or config_name == "default" or metadata_config.get("default"):
|
| 257 |
+
if default_config_name is None:
|
| 258 |
+
default_config_name = config_name
|
| 259 |
+
else:
|
| 260 |
+
raise ValueError(
|
| 261 |
+
f"Dataset has several default configs: '{default_config_name}' and '{config_name}'."
|
| 262 |
+
)
|
| 263 |
+
return default_config_name
|
| 264 |
+
|
| 265 |
+
|
| 266 |
+
# DEPRECATED - just here to support old versions of evaluate like 0.2.2
|
| 267 |
+
# To support new tasks on the Hugging Face Hub, please open a PR for this file:
|
| 268 |
+
# https://github.com/huggingface/huggingface.js/blob/main/packages/tasks/src/pipelines.ts
|
| 269 |
+
known_task_ids = {
|
| 270 |
+
"image-classification": [],
|
| 271 |
+
"translation": [],
|
| 272 |
+
"image-segmentation": [],
|
| 273 |
+
"fill-mask": [],
|
| 274 |
+
"automatic-speech-recognition": [],
|
| 275 |
+
"token-classification": [],
|
| 276 |
+
"sentence-similarity": [],
|
| 277 |
+
"audio-classification": [],
|
| 278 |
+
"question-answering": [],
|
| 279 |
+
"summarization": [],
|
| 280 |
+
"zero-shot-classification": [],
|
| 281 |
+
"table-to-text": [],
|
| 282 |
+
"feature-extraction": [],
|
| 283 |
+
"other": [],
|
| 284 |
+
"multiple-choice": [],
|
| 285 |
+
"text-classification": [],
|
| 286 |
+
"text-to-image": [],
|
| 287 |
+
"text2text-generation": [],
|
| 288 |
+
"zero-shot-image-classification": [],
|
| 289 |
+
"tabular-classification": [],
|
| 290 |
+
"tabular-regression": [],
|
| 291 |
+
"image-to-image": [],
|
| 292 |
+
"tabular-to-text": [],
|
| 293 |
+
"unconditional-image-generation": [],
|
| 294 |
+
"text-retrieval": [],
|
| 295 |
+
"text-to-speech": [],
|
| 296 |
+
"object-detection": [],
|
| 297 |
+
"audio-to-audio": [],
|
| 298 |
+
"text-generation": [],
|
| 299 |
+
"conversational": [],
|
| 300 |
+
"table-question-answering": [],
|
| 301 |
+
"visual-question-answering": [],
|
| 302 |
+
"image-to-text": [],
|
| 303 |
+
"reinforcement-learning": [],
|
| 304 |
+
"voice-activity-detection": [],
|
| 305 |
+
"time-series-forecasting": [],
|
| 306 |
+
"document-question-answering": [],
|
| 307 |
+
}
|
| 308 |
+
|
| 309 |
+
|
| 310 |
+
if __name__ == "__main__":
|
| 311 |
+
from argparse import ArgumentParser
|
| 312 |
+
|
| 313 |
+
ap = ArgumentParser(usage="Validate the yaml metadata block of a README.md file.")
|
| 314 |
+
ap.add_argument("readme_filepath")
|
| 315 |
+
args = ap.parse_args()
|
| 316 |
+
|
| 317 |
+
readme_filepath = Path(args.readme_filepath)
|
| 318 |
+
dataset_metadata = DatasetMetadata.from_readme(readme_filepath)
|
| 319 |
+
print(dataset_metadata)
|
| 320 |
+
dataset_metadata.to_readme(readme_filepath)
|
evalkit_tf437/lib/python3.10/site-packages/datasets/utils/track.py
ADDED
|
@@ -0,0 +1,49 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from collections.abc import Iterator
|
| 2 |
+
from typing import Iterable
|
| 3 |
+
|
| 4 |
+
|
| 5 |
+
class tracked_str(str):
|
| 6 |
+
origins = {}
|
| 7 |
+
|
| 8 |
+
def set_origin(self, origin: str):
|
| 9 |
+
if super().__repr__() not in self.origins:
|
| 10 |
+
self.origins[super().__repr__()] = origin
|
| 11 |
+
|
| 12 |
+
def get_origin(self):
|
| 13 |
+
return self.origins.get(super().__repr__(), str(self))
|
| 14 |
+
|
| 15 |
+
def __repr__(self) -> str:
|
| 16 |
+
if super().__repr__() not in self.origins or self.origins[super().__repr__()] == self:
|
| 17 |
+
return super().__repr__()
|
| 18 |
+
else:
|
| 19 |
+
return f"{str(self)} (origin={self.origins[super().__repr__()]})"
|
| 20 |
+
|
| 21 |
+
|
| 22 |
+
class tracked_list(list):
|
| 23 |
+
def __init__(self, *args, **kwargs) -> None:
|
| 24 |
+
super().__init__(*args, **kwargs)
|
| 25 |
+
self.last_item = None
|
| 26 |
+
|
| 27 |
+
def __iter__(self) -> Iterator:
|
| 28 |
+
for x in super().__iter__():
|
| 29 |
+
self.last_item = x
|
| 30 |
+
yield x
|
| 31 |
+
self.last_item = None
|
| 32 |
+
|
| 33 |
+
def __repr__(self) -> str:
|
| 34 |
+
if self.last_item is None:
|
| 35 |
+
return super().__repr__()
|
| 36 |
+
else:
|
| 37 |
+
return f"{self.__class__.__name__}(current={self.last_item})"
|
| 38 |
+
|
| 39 |
+
|
| 40 |
+
class TrackedIterable(Iterable):
|
| 41 |
+
def __init__(self) -> None:
|
| 42 |
+
super().__init__()
|
| 43 |
+
self.last_item = None
|
| 44 |
+
|
| 45 |
+
def __repr__(self) -> str:
|
| 46 |
+
if self.last_item is None:
|
| 47 |
+
super().__repr__()
|
| 48 |
+
else:
|
| 49 |
+
return f"{self.__class__.__name__}(current={self.last_item})"
|
evalkit_tf437/lib/python3.10/site-packages/fontTools/ttLib/tables/__pycache__/otData.cpython-310.pyc
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4a16720bcc7b01f6c5446f960184df7a3d8a9a0ef52a9c0739f1dca9846660c9
|
| 3 |
+
size 107114
|
evalkit_tf437/lib/python3.10/site-packages/latex2mathml-3.77.0.dist-info/INSTALLER
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
pip
|
evalkit_tf437/lib/python3.10/site-packages/latex2mathml-3.77.0.dist-info/LICENSE
ADDED
|
@@ -0,0 +1,21 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
MIT License
|
| 2 |
+
|
| 3 |
+
Copyright (c) 2016 Ronie Martinez
|
| 4 |
+
|
| 5 |
+
Permission is hereby granted, free of charge, to any person obtaining a copy
|
| 6 |
+
of this software and associated documentation files (the "Software"), to deal
|
| 7 |
+
in the Software without restriction, including without limitation the rights
|
| 8 |
+
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
| 9 |
+
copies of the Software, and to permit persons to whom the Software is
|
| 10 |
+
furnished to do so, subject to the following conditions:
|
| 11 |
+
|
| 12 |
+
The above copyright notice and this permission notice shall be included in all
|
| 13 |
+
copies or substantial portions of the Software.
|
| 14 |
+
|
| 15 |
+
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
| 16 |
+
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
| 17 |
+
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
| 18 |
+
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
| 19 |
+
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
| 20 |
+
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
| 21 |
+
SOFTWARE.
|
evalkit_tf437/lib/python3.10/site-packages/latex2mathml-3.77.0.dist-info/METADATA
ADDED
|
@@ -0,0 +1,172 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Metadata-Version: 2.1
|
| 2 |
+
Name: latex2mathml
|
| 3 |
+
Version: 3.77.0
|
| 4 |
+
Summary: Pure Python library for LaTeX to MathML conversion
|
| 5 |
+
Home-page: https://github.com/roniemartinez/latex2mathml
|
| 6 |
+
License: MIT
|
| 7 |
+
Keywords: latex,mathml
|
| 8 |
+
Author: Ronie Martinez
|
| 9 |
+
Author-email: ronmarti18@gmail.com
|
| 10 |
+
Requires-Python: >=3.8.1,<4.0.0
|
| 11 |
+
Classifier: Development Status :: 5 - Production/Stable
|
| 12 |
+
Classifier: License :: OSI Approved :: MIT License
|
| 13 |
+
Classifier: Programming Language :: Python :: 3
|
| 14 |
+
Classifier: Programming Language :: Python :: 3.9
|
| 15 |
+
Classifier: Programming Language :: Python :: 3.10
|
| 16 |
+
Classifier: Programming Language :: Python :: 3.11
|
| 17 |
+
Classifier: Programming Language :: Python :: 3.12
|
| 18 |
+
Classifier: Programming Language :: Python :: 3.8
|
| 19 |
+
Classifier: Programming Language :: Python :: Implementation :: CPython
|
| 20 |
+
Classifier: Topic :: Scientific/Engineering :: Mathematics
|
| 21 |
+
Classifier: Topic :: Software Development :: Libraries :: Python Modules
|
| 22 |
+
Classifier: Topic :: Text Processing :: Markup :: HTML
|
| 23 |
+
Classifier: Topic :: Text Processing :: Markup :: LaTeX
|
| 24 |
+
Project-URL: Donate, https://www.buymeacoffee.com/roniemartinez
|
| 25 |
+
Project-URL: Repository, https://github.com/roniemartinez/latex2mathml
|
| 26 |
+
Description-Content-Type: text/markdown
|
| 27 |
+
|
| 28 |
+
<table>
|
| 29 |
+
<tr>
|
| 30 |
+
<td>License</td>
|
| 31 |
+
<td><img src='https://img.shields.io/pypi/l/latex2mathml.svg?style=for-the-badge' alt="License"></td>
|
| 32 |
+
<td>Version</td>
|
| 33 |
+
<td><img src='https://img.shields.io/pypi/v/latex2mathml.svg?logo=pypi&style=for-the-badge' alt="Version"></td>
|
| 34 |
+
</tr>
|
| 35 |
+
<tr>
|
| 36 |
+
<td>Github Actions</td>
|
| 37 |
+
<td><img src='https://img.shields.io/github/actions/workflow/status/roniemartinez/latex2mathml/python.yml?branch=master&label=actions&logo=github%20actions&style=for-the-badge' alt="Github Actions"></td>
|
| 38 |
+
<td>Coverage</td>
|
| 39 |
+
<td><img src='https://img.shields.io/codecov/c/github/roniemartinez/latex2mathml/master?label=codecov&logo=codecov&style=for-the-badge' alt="CodeCov"></td>
|
| 40 |
+
</tr>
|
| 41 |
+
<tr>
|
| 42 |
+
<td>Supported versions</td>
|
| 43 |
+
<td><img src='https://img.shields.io/pypi/pyversions/latex2mathml.svg?logo=python&style=for-the-badge' alt="Python Versions"></td>
|
| 44 |
+
<td>Wheel</td>
|
| 45 |
+
<td><img src='https://img.shields.io/pypi/wheel/latex2mathml.svg?style=for-the-badge' alt="Wheel"></td>
|
| 46 |
+
</tr>
|
| 47 |
+
<tr>
|
| 48 |
+
<td>Status</td>
|
| 49 |
+
<td><img src='https://img.shields.io/pypi/status/latex2mathml.svg?style=for-the-badge' alt="Status"></td>
|
| 50 |
+
<td>Downloads</td>
|
| 51 |
+
<td><img src='https://img.shields.io/pypi/dm/latex2mathml.svg?style=for-the-badge' alt="Downloads"></td>
|
| 52 |
+
</tr>
|
| 53 |
+
<tr>
|
| 54 |
+
<td>All Contributors</td>
|
| 55 |
+
<td><a href="#contributors-"><img src='https://img.shields.io/github/all-contributors/roniemartinez/latex2mathml?style=for-the-badge' alt="All Contributors"></a></td>
|
| 56 |
+
</tr>
|
| 57 |
+
</table>
|
| 58 |
+
|
| 59 |
+
# latex2mathml
|
| 60 |
+
|
| 61 |
+
Pure Python library for LaTeX to MathML conversion
|
| 62 |
+
|
| 63 |
+
## Installation
|
| 64 |
+
|
| 65 |
+
```bash
|
| 66 |
+
pip install latex2mathml
|
| 67 |
+
```
|
| 68 |
+
|
| 69 |
+
## Usage
|
| 70 |
+
|
| 71 |
+
### Python
|
| 72 |
+
|
| 73 |
+
```python
|
| 74 |
+
import latex2mathml.converter
|
| 75 |
+
|
| 76 |
+
latex_input = "<your_latex_string>"
|
| 77 |
+
mathml_output = latex2mathml.converter.convert(latex_input)
|
| 78 |
+
```
|
| 79 |
+
|
| 80 |
+
### Command-line
|
| 81 |
+
|
| 82 |
+
```shell
|
| 83 |
+
% latex2mathml -h
|
| 84 |
+
usage: latex2mathml [-h] [-V] [-b] [-t TEXT | -f FILE | -s]
|
| 85 |
+
|
| 86 |
+
Pure Python library for LaTeX to MathML conversion
|
| 87 |
+
|
| 88 |
+
options:
|
| 89 |
+
-h, --help show this help message and exit
|
| 90 |
+
-V, --version Show version
|
| 91 |
+
-b, --block Display block
|
| 92 |
+
|
| 93 |
+
required arguments:
|
| 94 |
+
-t TEXT, --text TEXT Text
|
| 95 |
+
-f FILE, --file FILE File
|
| 96 |
+
-s, --stdin Stdin
|
| 97 |
+
```
|
| 98 |
+
|
| 99 |
+
## References
|
| 100 |
+
### LaTeX
|
| 101 |
+
|
| 102 |
+
- https://en.wikibooks.org/wiki/LaTeX/Mathematics
|
| 103 |
+
- http://artofproblemsolving.com/wiki/index.php?title=Main_Page
|
| 104 |
+
- http://milde.users.sourceforge.net/LUCR/Math/
|
| 105 |
+
- https://math-linux.com/latex-26/faq/latex-faq/article/latex-derivatives-limits-sums-products-and-integrals
|
| 106 |
+
- https://www.tutorialspoint.com/tex_commands
|
| 107 |
+
- https://www.giss.nasa.gov/tools/latex/ltx-86.html
|
| 108 |
+
- https://ftp.gwdg.de/pub/ctan/info/l2tabu/english/l2tabuen.pdf
|
| 109 |
+
|
| 110 |
+
### MathML
|
| 111 |
+
|
| 112 |
+
- http://www.xmlmind.com/tutorials/MathML/
|
| 113 |
+
|
| 114 |
+
|
| 115 |
+
## Author
|
| 116 |
+
|
| 117 |
+
- [Ronie Martinez](mailto:ronmarti18@gmail.com)
|
| 118 |
+
|
| 119 |
+
## Contributors β¨
|
| 120 |
+
|
| 121 |
+
Thanks goes to these wonderful people ([emoji key](https://allcontributors.org/docs/en/emoji-key)):
|
| 122 |
+
|
| 123 |
+
<!-- ALL-CONTRIBUTORS-LIST:START - Do not remove or modify this section -->
|
| 124 |
+
<!-- prettier-ignore-start -->
|
| 125 |
+
<!-- markdownlint-disable -->
|
| 126 |
+
<table>
|
| 127 |
+
<tbody>
|
| 128 |
+
<tr>
|
| 129 |
+
<td align="center" valign="top" width="14.28%"><a href="https://ron.sh"><img src="https://avatars.githubusercontent.com/u/2573537?v=4?s=100" width="100px;" alt="Ronie Martinez"/><br /><sub><b>Ronie Martinez</b></sub></a><br /><a href="#maintenance-roniemartinez" title="Maintenance">π§</a> <a href="https://github.com/roniemartinez/latex2mathml/commits?author=roniemartinez" title="Code">π»</a> <a href="#infra-roniemartinez" title="Infrastructure (Hosting, Build-Tools, etc)">π</a></td>
|
| 130 |
+
<td align="center" valign="top" width="14.28%"><a href="https://anwen.cc/"><img src="https://avatars.githubusercontent.com/u/1472850?v=4?s=100" width="100px;" alt="askender"/><br /><sub><b>askender</b></sub></a><br /><a href="https://github.com/roniemartinez/latex2mathml/commits?author=askender" title="Documentation">π</a></td>
|
| 131 |
+
<td align="center" valign="top" width="14.28%"><a href="https://github.com/06180339"><img src="https://avatars.githubusercontent.com/u/25408501?v=4?s=100" width="100px;" alt="06180339"/><br /><sub><b>06180339</b></sub></a><br /><a href="https://github.com/roniemartinez/latex2mathml/commits?author=06180339" title="Code">π»</a></td>
|
| 132 |
+
<td align="center" valign="top" width="14.28%"><a href="https://github.com/chaihahaha"><img src="https://avatars.githubusercontent.com/u/24356676?v=4?s=100" width="100px;" alt="chaihahaha"/><br /><sub><b>chaihahaha</b></sub></a><br /><a href="https://github.com/roniemartinez/latex2mathml/commits?author=chaihahaha" title="Code">π»</a></td>
|
| 133 |
+
<td align="center" valign="top" width="14.28%"><a href="https://github.com/huangradio"><img src="https://avatars.githubusercontent.com/u/63624395?v=4?s=100" width="100px;" alt="HQY"/><br /><sub><b>HQY</b></sub></a><br /><a href="https://github.com/roniemartinez/latex2mathml/issues?q=author%3Ahuangradio" title="Bug reports">π</a></td>
|
| 134 |
+
<td align="center" valign="top" width="14.28%"><a href="https://github.com/Sun-ZhenXing"><img src="https://avatars.githubusercontent.com/u/44517244?v=4?s=100" width="100px;" alt="ιΈζ’¨"/><br /><sub><b>ιΈζ’¨</b></sub></a><br /><a href="https://github.com/roniemartinez/latex2mathml/issues?q=author%3ASun-ZhenXing" title="Bug reports">π</a></td>
|
| 135 |
+
<td align="center" valign="top" width="14.28%"><a href="https://github.com/oliverstefanov"><img src="https://avatars.githubusercontent.com/u/33491656?v=4?s=100" width="100px;" alt="oliverstefanov"/><br /><sub><b>oliverstefanov</b></sub></a><br /><a href="https://github.com/roniemartinez/latex2mathml/issues?q=author%3Aoliverstefanov" title="Bug reports">π</a></td>
|
| 136 |
+
</tr>
|
| 137 |
+
<tr>
|
| 138 |
+
<td align="center" valign="top" width="14.28%"><a href="https://github.com/ghost"><img src="https://avatars.githubusercontent.com/u/10137?v=4?s=100" width="100px;" alt="Deleted user"/><br /><sub><b>Deleted user</b></sub></a><br /><a href="https://github.com/roniemartinez/latex2mathml/issues?q=author%3Aghost" title="Bug reports">π</a></td>
|
| 139 |
+
<td align="center" valign="top" width="14.28%"><a href="https://github.com/cesaryuan"><img src="https://avatars.githubusercontent.com/u/35998162?v=4?s=100" width="100px;" alt="Cesaryuan"/><br /><sub><b>Cesaryuan</b></sub></a><br /><a href="https://github.com/roniemartinez/latex2mathml/issues?q=author%3Acesaryuan" title="Bug reports">π</a></td>
|
| 140 |
+
<td align="center" valign="top" width="14.28%"><a href="https://github.com/tonystank3000"><img src="https://avatars.githubusercontent.com/u/6315974?v=4?s=100" width="100px;" alt="TonyStank"/><br /><sub><b>TonyStank</b></sub></a><br /><a href="https://github.com/roniemartinez/latex2mathml/issues?q=author%3Atonystank3000" title="Bug reports">π</a></td>
|
| 141 |
+
<td align="center" valign="top" width="14.28%"><a href="https://polarwinkel.de"><img src="https://avatars.githubusercontent.com/u/1512713?v=4?s=100" width="100px;" alt="Dirk Winkel"/><br /><sub><b>Dirk Winkel</b></sub></a><br /><a href="https://github.com/roniemartinez/latex2mathml/issues?q=author%3Apolarwinkel" title="Bug reports">π</a></td>
|
| 142 |
+
<td align="center" valign="top" width="14.28%"><a href="https://github.com/sinslu"><img src="https://avatars.githubusercontent.com/u/12248270?v=4?s=100" width="100px;" alt="sinslu"/><br /><sub><b>sinslu</b></sub></a><br /><a href="https://github.com/roniemartinez/latex2mathml/issues?q=author%3Asinslu" title="Bug reports">π</a></td>
|
| 143 |
+
<td align="center" valign="top" width="14.28%"><a href="https://ubavic.rs"><img src="https://avatars.githubusercontent.com/u/53820106?v=4?s=100" width="100px;" alt="Nikola UbaviΔ"/><br /><sub><b>Nikola UbaviΔ</b></sub></a><br /><a href="https://github.com/roniemartinez/latex2mathml/issues?q=author%3Aubavic" title="Bug reports">π</a></td>
|
| 144 |
+
<td align="center" valign="top" width="14.28%"><a href="https://github.com/abhisheksia"><img src="https://avatars.githubusercontent.com/u/68808662?v=4?s=100" width="100px;" alt="abhisheksia"/><br /><sub><b>abhisheksia</b></sub></a><br /><a href="https://github.com/roniemartinez/latex2mathml/issues?q=author%3Aabhisheksia" title="Bug reports">π</a></td>
|
| 145 |
+
</tr>
|
| 146 |
+
<tr>
|
| 147 |
+
<td align="center" valign="top" width="14.28%"><a href="http://denissalem.tuxfamily.org"><img src="https://avatars.githubusercontent.com/u/4476506?v=4?s=100" width="100px;" alt="Denis Salem"/><br /><sub><b>Denis Salem</b></sub></a><br /><a href="https://github.com/roniemartinez/latex2mathml/issues?q=author%3ADenisSalem" title="Bug reports">π</a> <a href="https://github.com/roniemartinez/latex2mathml/commits?author=DenisSalem" title="Code">π»</a></td>
|
| 148 |
+
<td align="center" valign="top" width="14.28%"><a href="https://clontz.org"><img src="https://avatars.githubusercontent.com/u/1559632?v=4?s=100" width="100px;" alt="Steven Clontz"/><br /><sub><b>Steven Clontz</b></sub></a><br /><a href="https://github.com/roniemartinez/latex2mathml/issues?q=author%3AStevenClontz" title="Bug reports">π</a></td>
|
| 149 |
+
<td align="center" valign="top" width="14.28%"><a href="https://github.com/yuwenjun1"><img src="https://avatars.githubusercontent.com/u/43265090?v=4?s=100" width="100px;" alt="η©Ίη½"/><br /><sub><b>η©Ίη½</b></sub></a><br /><a href="https://github.com/roniemartinez/latex2mathml/issues?q=author%3Ayuwenjun1" title="Bug reports">π</a></td>
|
| 150 |
+
<td align="center" valign="top" width="14.28%"><a href="https://github.com/amuramatsu"><img src="https://avatars.githubusercontent.com/u/6500918?v=4?s=100" width="100px;" alt="MURAMATSU Atshshi"/><br /><sub><b>MURAMATSU Atshshi</b></sub></a><br /><a href="https://github.com/roniemartinez/latex2mathml/issues?q=author%3Aamuramatsu" title="Bug reports">π</a></td>
|
| 151 |
+
<td align="center" valign="top" width="14.28%"><a href="https://github.com/leingang"><img src="https://avatars.githubusercontent.com/u/570942?v=4?s=100" width="100px;" alt="leingang"/><br /><sub><b>leingang</b></sub></a><br /><a href="https://github.com/roniemartinez/latex2mathml/issues?q=author%3Aleingang" title="Bug reports">π</a></td>
|
| 152 |
+
<td align="center" valign="top" width="14.28%"><a href="https://github.com/Nigel-Amers"><img src="https://avatars.githubusercontent.com/u/14248498?v=4?s=100" width="100px;" alt="Nigel Amers"/><br /><sub><b>Nigel Amers</b></sub></a><br /><a href="https://github.com/roniemartinez/latex2mathml/issues?q=author%3ANigel-Amers" title="Bug reports">π</a></td>
|
| 153 |
+
<td align="center" valign="top" width="14.28%"><a href="https://github.com/retsyo"><img src="https://avatars.githubusercontent.com/u/7960913?v=4?s=100" width="100px;" alt="retsyo"/><br /><sub><b>retsyo</b></sub></a><br /><a href="https://github.com/roniemartinez/latex2mathml/issues?q=author%3Aretsyo" title="Bug reports">π</a></td>
|
| 154 |
+
</tr>
|
| 155 |
+
<tr>
|
| 156 |
+
<td align="center" valign="top" width="14.28%"><a href="https://github.com/miakramer"><img src="https://avatars.githubusercontent.com/u/16845265?v=4?s=100" width="100px;" alt="Mia Kramer"/><br /><sub><b>Mia Kramer</b></sub></a><br /><a href="https://github.com/roniemartinez/latex2mathml/commits?author=miakramer" title="Code">π»</a> <a href="https://github.com/roniemartinez/latex2mathml/commits?author=miakramer" title="Documentation">π</a></td>
|
| 157 |
+
<td align="center" valign="top" width="14.28%"><a href="http://cnx.gdn"><img src="https://avatars.githubusercontent.com/u/13689192?v=4?s=100" width="100px;" alt="Nguyα»
n Gia Phong"/><br /><sub><b>Nguyα»
n Gia Phong</b></sub></a><br /><a href="https://github.com/roniemartinez/latex2mathml/issues?q=author%3AMcSinyx" title="Bug reports">π</a> <a href="https://github.com/roniemartinez/latex2mathml/commits?author=McSinyx" title="Code">π»</a> <a href="https://github.com/roniemartinez/latex2mathml/commits?author=McSinyx" title="Tests">β οΈ</a></td>
|
| 158 |
+
<td align="center" valign="top" width="14.28%"><a href="https://cdelker.github.io"><img src="https://avatars.githubusercontent.com/u/44102190?v=4?s=100" width="100px;" alt="Collin Delker"/><br /><sub><b>Collin Delker</b></sub></a><br /><a href="https://github.com/roniemartinez/latex2mathml/issues?q=author%3Acdelker" title="Bug reports">π</a></td>
|
| 159 |
+
<td align="center" valign="top" width="14.28%"><a href="https://github.com/jiotv321"><img src="https://avatars.githubusercontent.com/u/118644533?v=4?s=100" width="100px;" alt="jiotv321"/><br /><sub><b>jiotv321</b></sub></a><br /><a href="https://github.com/roniemartinez/latex2mathml/issues?q=author%3Ajiotv321" title="Bug reports">π</a></td>
|
| 160 |
+
<td align="center" valign="top" width="14.28%"><a href="https://github.com/felixonmars"><img src="https://avatars.githubusercontent.com/u/1006477?v=4?s=100" width="100px;" alt="Felix Yan"/><br /><sub><b>Felix Yan</b></sub></a><br /><a href="https://github.com/roniemartinez/latex2mathml/commits?author=felixonmars" title="Code">π»</a></td>
|
| 161 |
+
<td align="center" valign="top" width="14.28%"><a href="https://rob-blackbourn.github.io/blog/"><img src="https://avatars.githubusercontent.com/u/2880305?v=4?s=100" width="100px;" alt="Rob Blackbourn"/><br /><sub><b>Rob Blackbourn</b></sub></a><br /><a href="https://github.com/roniemartinez/latex2mathml/commits?author=rob-blackbourn" title="Code">π»</a></td>
|
| 162 |
+
</tr>
|
| 163 |
+
</tbody>
|
| 164 |
+
</table>
|
| 165 |
+
|
| 166 |
+
<!-- markdownlint-restore -->
|
| 167 |
+
<!-- prettier-ignore-end -->
|
| 168 |
+
|
| 169 |
+
<!-- ALL-CONTRIBUTORS-LIST:END -->
|
| 170 |
+
|
| 171 |
+
This project follows the [all-contributors](https://github.com/all-contributors/all-contributors) specification. Contributions of any kind welcome!
|
| 172 |
+
|
evalkit_tf437/lib/python3.10/site-packages/latex2mathml-3.77.0.dist-info/RECORD
ADDED
|
@@ -0,0 +1,25 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
../../../bin/l2m,sha256=A6cOPrAVRl9g-TTDoMgee4CvtzOWe52sOvGt69idfXc,240
|
| 2 |
+
../../../bin/latex2mathml,sha256=A6cOPrAVRl9g-TTDoMgee4CvtzOWe52sOvGt69idfXc,240
|
| 3 |
+
latex2mathml-3.77.0.dist-info/INSTALLER,sha256=zuuue4knoyJ-UwPPXg8fezS7VCrXJQrAP7zeNuwvFQg,4
|
| 4 |
+
latex2mathml-3.77.0.dist-info/LICENSE,sha256=SsHh2gf2s0PVQZHF4XFoQFNOCFkbP56uyQ1oAM3EdUM,1071
|
| 5 |
+
latex2mathml-3.77.0.dist-info/METADATA,sha256=8I_5snYGZtFTF_ZdrISzE_x4L1PcEc9WFhHJfwkOUhA,14844
|
| 6 |
+
latex2mathml-3.77.0.dist-info/RECORD,,
|
| 7 |
+
latex2mathml-3.77.0.dist-info/REQUESTED,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
| 8 |
+
latex2mathml-3.77.0.dist-info/WHEEL,sha256=FMvqSimYX_P7y0a7UY-_Mc83r5zkBZsCYPm7Lr0Bsq4,88
|
| 9 |
+
latex2mathml-3.77.0.dist-info/entry_points.txt,sha256=8cs8b08QDIhqIeEJaUBPeXlBTEE9a-d8ybiKa5ZGMak,92
|
| 10 |
+
latex2mathml/__init__.py,sha256=xA5GqWgZqbTv0TUeOiwBpOk8D48ibElsZI1OGTEnhuk,79
|
| 11 |
+
latex2mathml/__pycache__/__init__.cpython-310.pyc,,
|
| 12 |
+
latex2mathml/__pycache__/commands.cpython-310.pyc,,
|
| 13 |
+
latex2mathml/__pycache__/converter.cpython-310.pyc,,
|
| 14 |
+
latex2mathml/__pycache__/exceptions.cpython-310.pyc,,
|
| 15 |
+
latex2mathml/__pycache__/symbols_parser.cpython-310.pyc,,
|
| 16 |
+
latex2mathml/__pycache__/tokenizer.cpython-310.pyc,,
|
| 17 |
+
latex2mathml/__pycache__/walker.cpython-310.pyc,,
|
| 18 |
+
latex2mathml/commands.py,sha256=ZyBodQ-PqYfg4-LG6YipiDywzc5iUlYCLA1WTV5La04,13087
|
| 19 |
+
latex2mathml/converter.py,sha256=a3CTyam1xFz2oPkNVxklSXuPny_DmqnALLsDTRLVmeI,23219
|
| 20 |
+
latex2mathml/exceptions.py,sha256=zF1xPclVRSTzvZHPsujMv_bNITnHNW7Jvi4wg0aBQ60,645
|
| 21 |
+
latex2mathml/py.typed,sha256=JZLABTSp2pMHV6dDrYko_mqeQOmCl7_w46Ro2OaynOo,10
|
| 22 |
+
latex2mathml/symbols_parser.py,sha256=xZr_xYVWOj1-3GpdVKd0ly44AAkc2wvSmIUvbHBcJfg,2954
|
| 23 |
+
latex2mathml/tokenizer.py,sha256=FtlIzWm279chdDE6SrIohMDeHIcwT8GCeyNqexxHu94,2360
|
| 24 |
+
latex2mathml/unimathsymbols.txt,sha256=dO-JVO-vAdwfsTExPvbX97dBxPeOMPbW02ALndWlWu4,216334
|
| 25 |
+
latex2mathml/walker.py,sha256=3KZEjg4WbWQUogHmKO5yp0Ziu2n_m1JvyQs3sQc5fPA,19969
|
evalkit_tf437/lib/python3.10/site-packages/latex2mathml-3.77.0.dist-info/REQUESTED
ADDED
|
File without changes
|