| | import importlib |
| | import os |
| | import tempfile |
| | import types |
| | from contextlib import nullcontext as does_not_raise |
| | from multiprocessing import Process |
| | from pathlib import Path |
| | from unittest import TestCase |
| | from unittest.mock import patch |
| |
|
| | import numpy as np |
| | import pyarrow as pa |
| | import pyarrow.parquet as pq |
| | import pytest |
| | from multiprocess.pool import Pool |
| |
|
| | from datasets.arrow_dataset import Dataset |
| | from datasets.arrow_reader import DatasetNotOnHfGcsError |
| | from datasets.arrow_writer import ArrowWriter |
| | from datasets.builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder |
| | from datasets.dataset_dict import DatasetDict, IterableDatasetDict |
| | from datasets.download.download_manager import DownloadMode |
| | from datasets.features import Features, Value |
| | from datasets.info import DatasetInfo, PostProcessedInfo |
| | from datasets.iterable_dataset import IterableDataset |
| | from datasets.splits import Split, SplitDict, SplitGenerator, SplitInfo |
| | from datasets.streaming import xjoin |
| | from datasets.utils.file_utils import is_local_path |
| | from datasets.utils.info_utils import VerificationMode |
| | from datasets.utils.logging import INFO, get_logger |
| |
|
| | from .utils import ( |
| | assert_arrow_memory_doesnt_increase, |
| | assert_arrow_memory_increases, |
| | require_beam, |
| | require_faiss, |
| | set_current_working_directory_to_temp_dir, |
| | ) |
| |
|
| |
|
| | class DummyBuilder(DatasetBuilder): |
| | def _info(self): |
| | return DatasetInfo(features=Features({"text": Value("string")})) |
| |
|
| | def _split_generators(self, dl_manager): |
| | return [SplitGenerator(name=Split.TRAIN)] |
| |
|
| | def _prepare_split(self, split_generator, **kwargs): |
| | fname = f"{self.dataset_name}-{split_generator.name}.arrow" |
| | with ArrowWriter(features=self.info.features, path=os.path.join(self._output_dir, fname)) as writer: |
| | writer.write_batch({"text": ["foo"] * 100}) |
| | num_examples, num_bytes = writer.finalize() |
| | split_generator.split_info.num_examples = num_examples |
| | split_generator.split_info.num_bytes = num_bytes |
| |
|
| |
|
| | class DummyGeneratorBasedBuilder(GeneratorBasedBuilder): |
| | def _info(self): |
| | return DatasetInfo(features=Features({"text": Value("string")})) |
| |
|
| | def _split_generators(self, dl_manager): |
| | return [SplitGenerator(name=Split.TRAIN)] |
| |
|
| | def _generate_examples(self): |
| | for i in range(100): |
| | yield i, {"text": "foo"} |
| |
|
| |
|
| | class DummyArrowBasedBuilder(ArrowBasedBuilder): |
| | def _info(self): |
| | return DatasetInfo(features=Features({"text": Value("string")})) |
| |
|
| | def _split_generators(self, dl_manager): |
| | return [SplitGenerator(name=Split.TRAIN)] |
| |
|
| | def _generate_tables(self): |
| | for i in range(10): |
| | yield i, pa.table({"text": ["foo"] * 10}) |
| |
|
| |
|
| | class DummyBeamBasedBuilder(BeamBasedBuilder): |
| | def _info(self): |
| | return DatasetInfo(features=Features({"text": Value("string")})) |
| |
|
| | def _split_generators(self, dl_manager): |
| | return [SplitGenerator(name=Split.TRAIN)] |
| |
|
| | def _build_pcollection(self, pipeline): |
| | import apache_beam as beam |
| |
|
| | def _process(item): |
| | for i in range(10): |
| | yield f"{i}_{item}", {"text": "foo"} |
| |
|
| | return pipeline | "Initialize" >> beam.Create(range(10)) | "Extract content" >> beam.FlatMap(_process) |
| |
|
| |
|
| | class DummyGeneratorBasedBuilderWithIntegers(GeneratorBasedBuilder): |
| | def _info(self): |
| | return DatasetInfo(features=Features({"id": Value("int8")})) |
| |
|
| | def _split_generators(self, dl_manager): |
| | return [SplitGenerator(name=Split.TRAIN)] |
| |
|
| | def _generate_examples(self): |
| | for i in range(100): |
| | yield i, {"id": i} |
| |
|
| |
|
| | class DummyGeneratorBasedBuilderConfig(BuilderConfig): |
| | def __init__(self, content="foo", times=2, *args, **kwargs): |
| | super().__init__(*args, **kwargs) |
| | self.content = content |
| | self.times = times |
| |
|
| |
|
| | class DummyGeneratorBasedBuilderWithConfig(GeneratorBasedBuilder): |
| | BUILDER_CONFIG_CLASS = DummyGeneratorBasedBuilderConfig |
| |
|
| | def _info(self): |
| | return DatasetInfo(features=Features({"text": Value("string")})) |
| |
|
| | def _split_generators(self, dl_manager): |
| | return [SplitGenerator(name=Split.TRAIN)] |
| |
|
| | def _generate_examples(self): |
| | for i in range(100): |
| | yield i, {"text": self.config.content * self.config.times} |
| |
|
| |
|
| | class DummyBuilderWithMultipleConfigs(DummyBuilder): |
| | BUILDER_CONFIGS = [ |
| | DummyGeneratorBasedBuilderConfig(name="a"), |
| | DummyGeneratorBasedBuilderConfig(name="b"), |
| | ] |
| |
|
| |
|
| | class DummyBuilderWithDefaultConfig(DummyBuilderWithMultipleConfigs): |
| | DEFAULT_CONFIG_NAME = "a" |
| |
|
| |
|
| | class DummyBuilderWithDownload(DummyBuilder): |
| | def __init__(self, *args, rel_path=None, abs_path=None, **kwargs): |
| | super().__init__(*args, **kwargs) |
| | self._rel_path = rel_path |
| | self._abs_path = abs_path |
| |
|
| | def _split_generators(self, dl_manager): |
| | if self._rel_path is not None: |
| | assert os.path.exists(dl_manager.download(self._rel_path)), "dl_manager must support relative paths" |
| | if self._abs_path is not None: |
| | assert os.path.exists(dl_manager.download(self._abs_path)), "dl_manager must support absolute paths" |
| | return [SplitGenerator(name=Split.TRAIN)] |
| |
|
| |
|
| | class DummyBuilderWithManualDownload(DummyBuilderWithMultipleConfigs): |
| | @property |
| | def manual_download_instructions(self): |
| | return "To use the dataset you have to download some stuff manually and pass the data path to data_dir" |
| |
|
| | def _split_generators(self, dl_manager): |
| | if not os.path.exists(self.config.data_dir): |
| | raise FileNotFoundError(f"data_dir {self.config.data_dir} doesn't exist.") |
| | return [SplitGenerator(name=Split.TRAIN)] |
| |
|
| |
|
| | class DummyArrowBasedBuilderWithShards(ArrowBasedBuilder): |
| | def _info(self): |
| | return DatasetInfo(features=Features({"id": Value("int8"), "filepath": Value("string")})) |
| |
|
| | def _split_generators(self, dl_manager): |
| | return [SplitGenerator(name=Split.TRAIN, gen_kwargs={"filepaths": [f"data{i}.txt" for i in range(4)]})] |
| |
|
| | def _generate_tables(self, filepaths): |
| | idx = 0 |
| | for filepath in filepaths: |
| | for i in range(10): |
| | yield idx, pa.table({"id": range(10 * i, 10 * (i + 1)), "filepath": [filepath] * 10}) |
| | idx += 1 |
| |
|
| |
|
| | class DummyGeneratorBasedBuilderWithShards(GeneratorBasedBuilder): |
| | def _info(self): |
| | return DatasetInfo(features=Features({"id": Value("int8"), "filepath": Value("string")})) |
| |
|
| | def _split_generators(self, dl_manager): |
| | return [SplitGenerator(name=Split.TRAIN, gen_kwargs={"filepaths": [f"data{i}.txt" for i in range(4)]})] |
| |
|
| | def _generate_examples(self, filepaths): |
| | idx = 0 |
| | for filepath in filepaths: |
| | for i in range(100): |
| | yield idx, {"id": i, "filepath": filepath} |
| | idx += 1 |
| |
|
| |
|
| | class DummyArrowBasedBuilderWithAmbiguousShards(ArrowBasedBuilder): |
| | def _info(self): |
| | return DatasetInfo(features=Features({"id": Value("int8"), "filepath": Value("string")})) |
| |
|
| | def _split_generators(self, dl_manager): |
| | return [ |
| | SplitGenerator( |
| | name=Split.TRAIN, |
| | gen_kwargs={ |
| | "filepaths": [f"data{i}.txt" for i in range(4)], |
| | "dummy_kwarg_with_different_length": [f"dummy_data{i}.txt" for i in range(3)], |
| | }, |
| | ) |
| | ] |
| |
|
| | def _generate_tables(self, filepaths, dummy_kwarg_with_different_length): |
| | idx = 0 |
| | for filepath in filepaths: |
| | for i in range(10): |
| | yield idx, pa.table({"id": range(10 * i, 10 * (i + 1)), "filepath": [filepath] * 10}) |
| | idx += 1 |
| |
|
| |
|
| | class DummyGeneratorBasedBuilderWithAmbiguousShards(GeneratorBasedBuilder): |
| | def _info(self): |
| | return DatasetInfo(features=Features({"id": Value("int8"), "filepath": Value("string")})) |
| |
|
| | def _split_generators(self, dl_manager): |
| | return [ |
| | SplitGenerator( |
| | name=Split.TRAIN, |
| | gen_kwargs={ |
| | "filepaths": [f"data{i}.txt" for i in range(4)], |
| | "dummy_kwarg_with_different_length": [f"dummy_data{i}.txt" for i in range(3)], |
| | }, |
| | ) |
| | ] |
| |
|
| | def _generate_examples(self, filepaths, dummy_kwarg_with_different_length): |
| | idx = 0 |
| | for filepath in filepaths: |
| | for i in range(100): |
| | yield idx, {"id": i, "filepath": filepath} |
| | idx += 1 |
| |
|
| |
|
| | def _run_concurrent_download_and_prepare(tmp_dir): |
| | builder = DummyBuilder(cache_dir=tmp_dir) |
| | builder.download_and_prepare(try_from_hf_gcs=False, download_mode=DownloadMode.REUSE_DATASET_IF_EXISTS) |
| | return builder |
| |
|
| |
|
| | def check_streaming(builder): |
| | builders_module = importlib.import_module(builder.__module__) |
| | assert builders_module._patched_for_streaming |
| | assert builders_module.os.path.join is xjoin |
| |
|
| |
|
| | class BuilderTest(TestCase): |
| | def test_download_and_prepare(self): |
| | with tempfile.TemporaryDirectory() as tmp_dir: |
| | builder = DummyBuilder(cache_dir=tmp_dir) |
| | builder.download_and_prepare(try_from_hf_gcs=False, download_mode=DownloadMode.FORCE_REDOWNLOAD) |
| | self.assertTrue( |
| | os.path.exists( |
| | os.path.join( |
| | tmp_dir, builder.dataset_name, "default", "0.0.0", f"{builder.dataset_name}-train.arrow" |
| | ) |
| | ) |
| | ) |
| | self.assertDictEqual(builder.info.features, Features({"text": Value("string")})) |
| | self.assertEqual(builder.info.splits["train"].num_examples, 100) |
| | self.assertTrue( |
| | os.path.exists(os.path.join(tmp_dir, builder.dataset_name, "default", "0.0.0", "dataset_info.json")) |
| | ) |
| |
|
| | def test_download_and_prepare_checksum_computation(self): |
| | with tempfile.TemporaryDirectory() as tmp_dir: |
| | builder_no_verification = DummyBuilder(cache_dir=tmp_dir) |
| | builder_no_verification.download_and_prepare( |
| | try_from_hf_gcs=False, download_mode=DownloadMode.FORCE_REDOWNLOAD |
| | ) |
| | self.assertTrue( |
| | all(v["checksum"] is not None for _, v in builder_no_verification.info.download_checksums.items()) |
| | ) |
| | builder_with_verification = DummyBuilder(cache_dir=tmp_dir) |
| | builder_with_verification.download_and_prepare( |
| | try_from_hf_gcs=False, |
| | download_mode=DownloadMode.FORCE_REDOWNLOAD, |
| | verification_mode=VerificationMode.ALL_CHECKS, |
| | ) |
| | self.assertTrue( |
| | all(v["checksum"] is None for _, v in builder_with_verification.info.download_checksums.items()) |
| | ) |
| |
|
| | def test_concurrent_download_and_prepare(self): |
| | with tempfile.TemporaryDirectory() as tmp_dir: |
| | processes = 2 |
| | with Pool(processes=processes) as pool: |
| | jobs = [ |
| | pool.apply_async(_run_concurrent_download_and_prepare, kwds={"tmp_dir": tmp_dir}) |
| | for _ in range(processes) |
| | ] |
| | builders = [job.get() for job in jobs] |
| | for builder in builders: |
| | self.assertTrue( |
| | os.path.exists( |
| | os.path.join( |
| | tmp_dir, |
| | builder.dataset_name, |
| | "default", |
| | "0.0.0", |
| | f"{builder.dataset_name}-train.arrow", |
| | ) |
| | ) |
| | ) |
| | self.assertDictEqual(builder.info.features, Features({"text": Value("string")})) |
| | self.assertEqual(builder.info.splits["train"].num_examples, 100) |
| | self.assertTrue( |
| | os.path.exists( |
| | os.path.join(tmp_dir, builder.dataset_name, "default", "0.0.0", "dataset_info.json") |
| | ) |
| | ) |
| |
|
| | def test_download_and_prepare_with_base_path(self): |
| | with tempfile.TemporaryDirectory() as tmp_dir: |
| | rel_path = "dummy1.data" |
| | abs_path = os.path.join(tmp_dir, "dummy2.data") |
| | |
| | builder = DummyBuilderWithDownload(cache_dir=tmp_dir, rel_path=rel_path) |
| | with self.assertRaises(FileNotFoundError): |
| | builder.download_and_prepare( |
| | try_from_hf_gcs=False, download_mode=DownloadMode.FORCE_REDOWNLOAD, base_path=tmp_dir |
| | ) |
| | |
| | builder = DummyBuilderWithDownload(cache_dir=tmp_dir, abs_path=abs_path) |
| | with self.assertRaises(FileNotFoundError): |
| | builder.download_and_prepare( |
| | try_from_hf_gcs=False, download_mode=DownloadMode.FORCE_REDOWNLOAD, base_path=tmp_dir |
| | ) |
| | |
| | open(os.path.join(tmp_dir, rel_path), "w") |
| | open(abs_path, "w") |
| | builder = DummyBuilderWithDownload(cache_dir=tmp_dir, rel_path=rel_path, abs_path=abs_path) |
| | builder.download_and_prepare( |
| | try_from_hf_gcs=False, download_mode=DownloadMode.FORCE_REDOWNLOAD, base_path=tmp_dir |
| | ) |
| | self.assertTrue( |
| | os.path.exists( |
| | os.path.join( |
| | tmp_dir, |
| | builder.dataset_name, |
| | "default", |
| | "0.0.0", |
| | f"{builder.dataset_name}-train.arrow", |
| | ) |
| | ) |
| | ) |
| |
|
| | def test_as_dataset_with_post_process(self): |
| | def _post_process(self, dataset, resources_paths): |
| | def char_tokenize(example): |
| | return {"tokens": list(example["text"])} |
| |
|
| | return dataset.map(char_tokenize, cache_file_name=resources_paths["tokenized_dataset"]) |
| |
|
| | def _post_processing_resources(self, split): |
| | return {"tokenized_dataset": f"tokenized_dataset-{split}.arrow"} |
| |
|
| | with tempfile.TemporaryDirectory() as tmp_dir: |
| | builder = DummyBuilder(cache_dir=tmp_dir) |
| | builder.info.post_processed = PostProcessedInfo( |
| | features=Features({"text": Value("string"), "tokens": [Value("string")]}) |
| | ) |
| | builder._post_process = types.MethodType(_post_process, builder) |
| | builder._post_processing_resources = types.MethodType(_post_processing_resources, builder) |
| | os.makedirs(builder.cache_dir) |
| |
|
| | builder.info.splits = SplitDict() |
| | builder.info.splits.add(SplitInfo("train", num_examples=10)) |
| | builder.info.splits.add(SplitInfo("test", num_examples=10)) |
| |
|
| | for split in builder.info.splits: |
| | with ArrowWriter( |
| | path=os.path.join(builder.cache_dir, f"{builder.dataset_name}-{split}.arrow"), |
| | features=Features({"text": Value("string")}), |
| | ) as writer: |
| | writer.write_batch({"text": ["foo"] * 10}) |
| | writer.finalize() |
| |
|
| | with ArrowWriter( |
| | path=os.path.join(builder.cache_dir, f"tokenized_dataset-{split}.arrow"), |
| | features=Features({"text": Value("string"), "tokens": [Value("string")]}), |
| | ) as writer: |
| | writer.write_batch({"text": ["foo"] * 10, "tokens": [list("foo")] * 10}) |
| | writer.finalize() |
| |
|
| | dsets = builder.as_dataset() |
| | self.assertIsInstance(dsets, DatasetDict) |
| | self.assertListEqual(list(dsets.keys()), ["train", "test"]) |
| | self.assertEqual(len(dsets["train"]), 10) |
| | self.assertEqual(len(dsets["test"]), 10) |
| | self.assertDictEqual( |
| | dsets["train"].features, Features({"text": Value("string"), "tokens": [Value("string")]}) |
| | ) |
| | self.assertDictEqual( |
| | dsets["test"].features, Features({"text": Value("string"), "tokens": [Value("string")]}) |
| | ) |
| | self.assertListEqual(dsets["train"].column_names, ["text", "tokens"]) |
| | self.assertListEqual(dsets["test"].column_names, ["text", "tokens"]) |
| | del dsets |
| |
|
| | dset = builder.as_dataset("train") |
| | self.assertIsInstance(dset, Dataset) |
| | self.assertEqual(dset.split, "train") |
| | self.assertEqual(len(dset), 10) |
| | self.assertDictEqual(dset.features, Features({"text": Value("string"), "tokens": [Value("string")]})) |
| | self.assertListEqual(dset.column_names, ["text", "tokens"]) |
| | self.assertGreater(builder.info.post_processing_size, 0) |
| | self.assertGreater( |
| | builder.info.post_processed.resources_checksums["train"]["tokenized_dataset"]["num_bytes"], 0 |
| | ) |
| | del dset |
| |
|
| | dset = builder.as_dataset("train+test[:30%]") |
| | self.assertIsInstance(dset, Dataset) |
| | self.assertEqual(dset.split, "train+test[:30%]") |
| | self.assertEqual(len(dset), 13) |
| | self.assertDictEqual(dset.features, Features({"text": Value("string"), "tokens": [Value("string")]})) |
| | self.assertListEqual(dset.column_names, ["text", "tokens"]) |
| | del dset |
| |
|
| | dset = builder.as_dataset("all") |
| | self.assertIsInstance(dset, Dataset) |
| | self.assertEqual(dset.split, "train+test") |
| | self.assertEqual(len(dset), 20) |
| | self.assertDictEqual(dset.features, Features({"text": Value("string"), "tokens": [Value("string")]})) |
| | self.assertListEqual(dset.column_names, ["text", "tokens"]) |
| | del dset |
| |
|
| | def _post_process(self, dataset, resources_paths): |
| | return dataset.select([0, 1], keep_in_memory=True) |
| |
|
| | with tempfile.TemporaryDirectory() as tmp_dir: |
| | builder = DummyBuilder(cache_dir=tmp_dir) |
| | builder._post_process = types.MethodType(_post_process, builder) |
| | os.makedirs(builder.cache_dir) |
| |
|
| | builder.info.splits = SplitDict() |
| | builder.info.splits.add(SplitInfo("train", num_examples=10)) |
| | builder.info.splits.add(SplitInfo("test", num_examples=10)) |
| |
|
| | for split in builder.info.splits: |
| | with ArrowWriter( |
| | path=os.path.join(builder.cache_dir, f"{builder.dataset_name}-{split}.arrow"), |
| | features=Features({"text": Value("string")}), |
| | ) as writer: |
| | writer.write_batch({"text": ["foo"] * 10}) |
| | writer.finalize() |
| |
|
| | with ArrowWriter( |
| | path=os.path.join(builder.cache_dir, f"small_dataset-{split}.arrow"), |
| | features=Features({"text": Value("string")}), |
| | ) as writer: |
| | writer.write_batch({"text": ["foo"] * 2}) |
| | writer.finalize() |
| |
|
| | dsets = builder.as_dataset() |
| | self.assertIsInstance(dsets, DatasetDict) |
| | self.assertListEqual(list(dsets.keys()), ["train", "test"]) |
| | self.assertEqual(len(dsets["train"]), 2) |
| | self.assertEqual(len(dsets["test"]), 2) |
| | self.assertDictEqual(dsets["train"].features, Features({"text": Value("string")})) |
| | self.assertDictEqual(dsets["test"].features, Features({"text": Value("string")})) |
| | self.assertListEqual(dsets["train"].column_names, ["text"]) |
| | self.assertListEqual(dsets["test"].column_names, ["text"]) |
| | del dsets |
| |
|
| | dset = builder.as_dataset("train") |
| | self.assertIsInstance(dset, Dataset) |
| | self.assertEqual(dset.split, "train") |
| | self.assertEqual(len(dset), 2) |
| | self.assertDictEqual(dset.features, Features({"text": Value("string")})) |
| | self.assertListEqual(dset.column_names, ["text"]) |
| | del dset |
| |
|
| | dset = builder.as_dataset("train+test[:30%]") |
| | self.assertIsInstance(dset, Dataset) |
| | self.assertEqual(dset.split, "train+test[:30%]") |
| | self.assertEqual(len(dset), 2) |
| | self.assertDictEqual(dset.features, Features({"text": Value("string")})) |
| | self.assertListEqual(dset.column_names, ["text"]) |
| | del dset |
| |
|
| | @require_faiss |
| | def test_as_dataset_with_post_process_with_index(self): |
| | def _post_process(self, dataset, resources_paths): |
| | if os.path.exists(resources_paths["index"]): |
| | dataset.load_faiss_index("my_index", resources_paths["index"]) |
| | return dataset |
| | else: |
| | dataset.add_faiss_index_from_external_arrays( |
| | external_arrays=np.ones((len(dataset), 8)), string_factory="Flat", index_name="my_index" |
| | ) |
| | dataset.save_faiss_index("my_index", resources_paths["index"]) |
| | return dataset |
| |
|
| | def _post_processing_resources(self, split): |
| | return {"index": f"Flat-{split}.faiss"} |
| |
|
| | with tempfile.TemporaryDirectory() as tmp_dir: |
| | builder = DummyBuilder(cache_dir=tmp_dir) |
| | builder._post_process = types.MethodType(_post_process, builder) |
| | builder._post_processing_resources = types.MethodType(_post_processing_resources, builder) |
| | os.makedirs(builder.cache_dir) |
| |
|
| | builder.info.splits = SplitDict() |
| | builder.info.splits.add(SplitInfo("train", num_examples=10)) |
| | builder.info.splits.add(SplitInfo("test", num_examples=10)) |
| |
|
| | for split in builder.info.splits: |
| | with ArrowWriter( |
| | path=os.path.join(builder.cache_dir, f"{builder.dataset_name}-{split}.arrow"), |
| | features=Features({"text": Value("string")}), |
| | ) as writer: |
| | writer.write_batch({"text": ["foo"] * 10}) |
| | writer.finalize() |
| |
|
| | with ArrowWriter( |
| | path=os.path.join(builder.cache_dir, f"small_dataset-{split}.arrow"), |
| | features=Features({"text": Value("string")}), |
| | ) as writer: |
| | writer.write_batch({"text": ["foo"] * 2}) |
| | writer.finalize() |
| |
|
| | dsets = builder.as_dataset() |
| | self.assertIsInstance(dsets, DatasetDict) |
| | self.assertListEqual(list(dsets.keys()), ["train", "test"]) |
| | self.assertEqual(len(dsets["train"]), 10) |
| | self.assertEqual(len(dsets["test"]), 10) |
| | self.assertDictEqual(dsets["train"].features, Features({"text": Value("string")})) |
| | self.assertDictEqual(dsets["test"].features, Features({"text": Value("string")})) |
| | self.assertListEqual(dsets["train"].column_names, ["text"]) |
| | self.assertListEqual(dsets["test"].column_names, ["text"]) |
| | self.assertListEqual(dsets["train"].list_indexes(), ["my_index"]) |
| | self.assertListEqual(dsets["test"].list_indexes(), ["my_index"]) |
| | self.assertGreater(builder.info.post_processing_size, 0) |
| | self.assertGreater(builder.info.post_processed.resources_checksums["train"]["index"]["num_bytes"], 0) |
| | del dsets |
| |
|
| | dset = builder.as_dataset("train") |
| | self.assertIsInstance(dset, Dataset) |
| | self.assertEqual(dset.split, "train") |
| | self.assertEqual(len(dset), 10) |
| | self.assertDictEqual(dset.features, Features({"text": Value("string")})) |
| | self.assertListEqual(dset.column_names, ["text"]) |
| | self.assertListEqual(dset.list_indexes(), ["my_index"]) |
| | del dset |
| |
|
| | dset = builder.as_dataset("train+test[:30%]") |
| | self.assertIsInstance(dset, Dataset) |
| | self.assertEqual(dset.split, "train+test[:30%]") |
| | self.assertEqual(len(dset), 13) |
| | self.assertDictEqual(dset.features, Features({"text": Value("string")})) |
| | self.assertListEqual(dset.column_names, ["text"]) |
| | self.assertListEqual(dset.list_indexes(), ["my_index"]) |
| | del dset |
| |
|
| | def test_download_and_prepare_with_post_process(self): |
| | def _post_process(self, dataset, resources_paths): |
| | def char_tokenize(example): |
| | return {"tokens": list(example["text"])} |
| |
|
| | return dataset.map(char_tokenize, cache_file_name=resources_paths["tokenized_dataset"]) |
| |
|
| | def _post_processing_resources(self, split): |
| | return {"tokenized_dataset": f"tokenized_dataset-{split}.arrow"} |
| |
|
| | with tempfile.TemporaryDirectory() as tmp_dir: |
| | builder = DummyBuilder(cache_dir=tmp_dir) |
| | builder.info.post_processed = PostProcessedInfo( |
| | features=Features({"text": Value("string"), "tokens": [Value("string")]}) |
| | ) |
| | builder._post_process = types.MethodType(_post_process, builder) |
| | builder._post_processing_resources = types.MethodType(_post_processing_resources, builder) |
| | builder.download_and_prepare(try_from_hf_gcs=False, download_mode=DownloadMode.FORCE_REDOWNLOAD) |
| | self.assertTrue( |
| | os.path.exists( |
| | os.path.join( |
| | tmp_dir, builder.dataset_name, "default", "0.0.0", f"{builder.dataset_name}-train.arrow" |
| | ) |
| | ) |
| | ) |
| | self.assertDictEqual(builder.info.features, Features({"text": Value("string")})) |
| | self.assertDictEqual( |
| | builder.info.post_processed.features, |
| | Features({"text": Value("string"), "tokens": [Value("string")]}), |
| | ) |
| | self.assertEqual(builder.info.splits["train"].num_examples, 100) |
| | self.assertTrue( |
| | os.path.exists(os.path.join(tmp_dir, builder.dataset_name, "default", "0.0.0", "dataset_info.json")) |
| | ) |
| |
|
| | def _post_process(self, dataset, resources_paths): |
| | return dataset.select([0, 1], keep_in_memory=True) |
| |
|
| | with tempfile.TemporaryDirectory() as tmp_dir: |
| | builder = DummyBuilder(cache_dir=tmp_dir) |
| | builder._post_process = types.MethodType(_post_process, builder) |
| | builder.download_and_prepare(try_from_hf_gcs=False, download_mode=DownloadMode.FORCE_REDOWNLOAD) |
| | self.assertTrue( |
| | os.path.exists( |
| | os.path.join( |
| | tmp_dir, builder.dataset_name, "default", "0.0.0", f"{builder.dataset_name}-train.arrow" |
| | ) |
| | ) |
| | ) |
| | self.assertDictEqual(builder.info.features, Features({"text": Value("string")})) |
| | self.assertIsNone(builder.info.post_processed) |
| | self.assertEqual(builder.info.splits["train"].num_examples, 100) |
| | self.assertTrue( |
| | os.path.exists(os.path.join(tmp_dir, builder.dataset_name, "default", "0.0.0", "dataset_info.json")) |
| | ) |
| |
|
| | def _post_process(self, dataset, resources_paths): |
| | if os.path.exists(resources_paths["index"]): |
| | dataset.load_faiss_index("my_index", resources_paths["index"]) |
| | return dataset |
| | else: |
| | dataset = dataset.add_faiss_index_from_external_arrays( |
| | external_arrays=np.ones((len(dataset), 8)), string_factory="Flat", index_name="my_index" |
| | ) |
| | dataset.save_faiss_index("my_index", resources_paths["index"]) |
| | return dataset |
| |
|
| | def _post_processing_resources(self, split): |
| | return {"index": f"Flat-{split}.faiss"} |
| |
|
| | with tempfile.TemporaryDirectory() as tmp_dir: |
| | builder = DummyBuilder(cache_dir=tmp_dir) |
| | builder._post_process = types.MethodType(_post_process, builder) |
| | builder._post_processing_resources = types.MethodType(_post_processing_resources, builder) |
| | builder.download_and_prepare(try_from_hf_gcs=False, download_mode=DownloadMode.FORCE_REDOWNLOAD) |
| | self.assertTrue( |
| | os.path.exists( |
| | os.path.join( |
| | tmp_dir, builder.dataset_name, "default", "0.0.0", f"{builder.dataset_name}-train.arrow" |
| | ) |
| | ) |
| | ) |
| | self.assertDictEqual(builder.info.features, Features({"text": Value("string")})) |
| | self.assertIsNone(builder.info.post_processed) |
| | self.assertEqual(builder.info.splits["train"].num_examples, 100) |
| | self.assertTrue( |
| | os.path.exists(os.path.join(tmp_dir, builder.dataset_name, "default", "0.0.0", "dataset_info.json")) |
| | ) |
| |
|
| | def test_error_download_and_prepare(self): |
| | def _prepare_split(self, split_generator, **kwargs): |
| | raise ValueError() |
| |
|
| | with tempfile.TemporaryDirectory() as tmp_dir: |
| | builder = DummyBuilder(cache_dir=tmp_dir) |
| | builder._prepare_split = types.MethodType(_prepare_split, builder) |
| | self.assertRaises( |
| | ValueError, |
| | builder.download_and_prepare, |
| | try_from_hf_gcs=False, |
| | download_mode=DownloadMode.FORCE_REDOWNLOAD, |
| | ) |
| | self.assertRaises(FileNotFoundError, builder.as_dataset) |
| |
|
| | def test_generator_based_download_and_prepare(self): |
| | with tempfile.TemporaryDirectory() as tmp_dir: |
| | builder = DummyGeneratorBasedBuilder(cache_dir=tmp_dir) |
| | builder.download_and_prepare(try_from_hf_gcs=False, download_mode=DownloadMode.FORCE_REDOWNLOAD) |
| | self.assertTrue( |
| | os.path.exists( |
| | os.path.join( |
| | tmp_dir, |
| | builder.dataset_name, |
| | "default", |
| | "0.0.0", |
| | f"{builder.dataset_name}-train.arrow", |
| | ) |
| | ) |
| | ) |
| | self.assertDictEqual(builder.info.features, Features({"text": Value("string")})) |
| | self.assertEqual(builder.info.splits["train"].num_examples, 100) |
| | self.assertTrue( |
| | os.path.exists(os.path.join(tmp_dir, builder.dataset_name, "default", "0.0.0", "dataset_info.json")) |
| | ) |
| |
|
| | |
| | with tempfile.TemporaryDirectory() as tmp_dir: |
| | builder = DummyGeneratorBasedBuilder(cache_dir=tmp_dir) |
| | with patch("datasets.builder.ArrowWriter", side_effect=ArrowWriter) as mock_arrow_writer: |
| | builder.download_and_prepare( |
| | download_mode=DownloadMode.FORCE_REDOWNLOAD, verification_mode=VerificationMode.NO_CHECKS |
| | ) |
| | mock_arrow_writer.assert_called_once() |
| | args, kwargs = mock_arrow_writer.call_args_list[0] |
| | self.assertFalse(kwargs["check_duplicates"]) |
| |
|
| | mock_arrow_writer.reset_mock() |
| |
|
| | builder.download_and_prepare( |
| | download_mode=DownloadMode.FORCE_REDOWNLOAD, verification_mode=VerificationMode.BASIC_CHECKS |
| | ) |
| | mock_arrow_writer.assert_called_once() |
| | args, kwargs = mock_arrow_writer.call_args_list[0] |
| | self.assertTrue(kwargs["check_duplicates"]) |
| |
|
| | def test_cache_dir_no_args(self): |
| | with tempfile.TemporaryDirectory() as tmp_dir: |
| | builder = DummyGeneratorBasedBuilder(cache_dir=tmp_dir, data_dir=None, data_files=None) |
| | relative_cache_dir_parts = Path(builder._relative_data_dir()).parts |
| | self.assertTupleEqual(relative_cache_dir_parts, (builder.dataset_name, "default", "0.0.0")) |
| |
|
| | def test_cache_dir_for_data_files(self): |
| | with tempfile.TemporaryDirectory() as tmp_dir: |
| | dummy_data1 = os.path.join(tmp_dir, "dummy_data1.txt") |
| | with open(dummy_data1, "w", encoding="utf-8") as f: |
| | f.writelines("foo bar") |
| | dummy_data2 = os.path.join(tmp_dir, "dummy_data2.txt") |
| | with open(dummy_data2, "w", encoding="utf-8") as f: |
| | f.writelines("foo bar\n") |
| |
|
| | builder = DummyGeneratorBasedBuilder(cache_dir=tmp_dir, data_files=dummy_data1) |
| | other_builder = DummyGeneratorBasedBuilder(cache_dir=tmp_dir, data_files=dummy_data1) |
| | self.assertEqual(builder.cache_dir, other_builder.cache_dir) |
| | other_builder = DummyGeneratorBasedBuilder(cache_dir=tmp_dir, data_files=[dummy_data1]) |
| | self.assertEqual(builder.cache_dir, other_builder.cache_dir) |
| | other_builder = DummyGeneratorBasedBuilder(cache_dir=tmp_dir, data_files={"train": dummy_data1}) |
| | self.assertEqual(builder.cache_dir, other_builder.cache_dir) |
| | other_builder = DummyGeneratorBasedBuilder(cache_dir=tmp_dir, data_files={Split.TRAIN: dummy_data1}) |
| | self.assertEqual(builder.cache_dir, other_builder.cache_dir) |
| | other_builder = DummyGeneratorBasedBuilder(cache_dir=tmp_dir, data_files={"train": [dummy_data1]}) |
| | self.assertEqual(builder.cache_dir, other_builder.cache_dir) |
| | other_builder = DummyGeneratorBasedBuilder(cache_dir=tmp_dir, data_files={"test": dummy_data1}) |
| | self.assertNotEqual(builder.cache_dir, other_builder.cache_dir) |
| | other_builder = DummyGeneratorBasedBuilder(cache_dir=tmp_dir, data_files=dummy_data2) |
| | self.assertNotEqual(builder.cache_dir, other_builder.cache_dir) |
| | other_builder = DummyGeneratorBasedBuilder(cache_dir=tmp_dir, data_files=[dummy_data2]) |
| | self.assertNotEqual(builder.cache_dir, other_builder.cache_dir) |
| | other_builder = DummyGeneratorBasedBuilder(cache_dir=tmp_dir, data_files=[dummy_data1, dummy_data2]) |
| | self.assertNotEqual(builder.cache_dir, other_builder.cache_dir) |
| |
|
| | builder = DummyGeneratorBasedBuilder(cache_dir=tmp_dir, data_files=[dummy_data1, dummy_data2]) |
| | other_builder = DummyGeneratorBasedBuilder(cache_dir=tmp_dir, data_files=[dummy_data1, dummy_data2]) |
| | self.assertEqual(builder.cache_dir, other_builder.cache_dir) |
| | other_builder = DummyGeneratorBasedBuilder(cache_dir=tmp_dir, data_files=[dummy_data2, dummy_data1]) |
| | self.assertNotEqual(builder.cache_dir, other_builder.cache_dir) |
| |
|
| | builder = DummyGeneratorBasedBuilder( |
| | cache_dir=tmp_dir, data_files={"train": dummy_data1, "test": dummy_data2} |
| | ) |
| | other_builder = DummyGeneratorBasedBuilder( |
| | cache_dir=tmp_dir, data_files={"train": dummy_data1, "test": dummy_data2} |
| | ) |
| | self.assertEqual(builder.cache_dir, other_builder.cache_dir) |
| | other_builder = DummyGeneratorBasedBuilder( |
| | cache_dir=tmp_dir, data_files={"train": [dummy_data1], "test": dummy_data2} |
| | ) |
| | self.assertEqual(builder.cache_dir, other_builder.cache_dir) |
| | other_builder = DummyGeneratorBasedBuilder( |
| | cache_dir=tmp_dir, data_files={"train": dummy_data1, "validation": dummy_data2} |
| | ) |
| | self.assertNotEqual(builder.cache_dir, other_builder.cache_dir) |
| | other_builder = DummyGeneratorBasedBuilder( |
| | cache_dir=tmp_dir, |
| | data_files={"train": [dummy_data1, dummy_data2], "test": dummy_data2}, |
| | ) |
| | self.assertNotEqual(builder.cache_dir, other_builder.cache_dir) |
| |
|
| | def test_cache_dir_for_features(self): |
| | with tempfile.TemporaryDirectory() as tmp_dir: |
| | f1 = Features({"id": Value("int8")}) |
| | f2 = Features({"id": Value("int32")}) |
| | builder = DummyGeneratorBasedBuilderWithIntegers(cache_dir=tmp_dir, features=f1) |
| | other_builder = DummyGeneratorBasedBuilderWithIntegers(cache_dir=tmp_dir, features=f1) |
| | self.assertEqual(builder.cache_dir, other_builder.cache_dir) |
| | other_builder = DummyGeneratorBasedBuilderWithIntegers(cache_dir=tmp_dir, features=f2) |
| | self.assertNotEqual(builder.cache_dir, other_builder.cache_dir) |
| |
|
| | def test_cache_dir_for_config_kwargs(self): |
| | with tempfile.TemporaryDirectory() as tmp_dir: |
| | |
| | builder = DummyGeneratorBasedBuilderWithConfig(cache_dir=tmp_dir, content="foo", times=2) |
| | other_builder = DummyGeneratorBasedBuilderWithConfig(cache_dir=tmp_dir, times=2, content="foo") |
| | self.assertEqual(builder.cache_dir, other_builder.cache_dir) |
| | self.assertIn("content=foo", builder.cache_dir) |
| | self.assertIn("times=2", builder.cache_dir) |
| | other_builder = DummyGeneratorBasedBuilderWithConfig(cache_dir=tmp_dir, content="bar", times=2) |
| | self.assertNotEqual(builder.cache_dir, other_builder.cache_dir) |
| | other_builder = DummyGeneratorBasedBuilderWithConfig(cache_dir=tmp_dir, content="foo") |
| | self.assertNotEqual(builder.cache_dir, other_builder.cache_dir) |
| |
|
| | with tempfile.TemporaryDirectory() as tmp_dir: |
| | |
| | builder = DummyBuilderWithMultipleConfigs(cache_dir=tmp_dir, config_name="a", content="foo", times=2) |
| | other_builder = DummyBuilderWithMultipleConfigs(cache_dir=tmp_dir, config_name="a", times=2, content="foo") |
| | self.assertEqual(builder.cache_dir, other_builder.cache_dir) |
| | self.assertIn("content=foo", builder.cache_dir) |
| | self.assertIn("times=2", builder.cache_dir) |
| | other_builder = DummyBuilderWithMultipleConfigs(cache_dir=tmp_dir, config_name="a", content="bar", times=2) |
| | self.assertNotEqual(builder.cache_dir, other_builder.cache_dir) |
| | other_builder = DummyBuilderWithMultipleConfigs(cache_dir=tmp_dir, config_name="a", content="foo") |
| | self.assertNotEqual(builder.cache_dir, other_builder.cache_dir) |
| |
|
| | def test_config_names(self): |
| | with tempfile.TemporaryDirectory() as tmp_dir: |
| | with self.assertRaises(ValueError) as error_context: |
| | DummyBuilderWithMultipleConfigs(cache_dir=tmp_dir, data_files=None, data_dir=None) |
| | self.assertIn("Please pick one among the available configs", str(error_context.exception)) |
| |
|
| | builder = DummyBuilderWithMultipleConfigs(cache_dir=tmp_dir, config_name="a") |
| | self.assertEqual(builder.config.name, "a") |
| |
|
| | builder = DummyBuilderWithMultipleConfigs(cache_dir=tmp_dir, config_name="b") |
| | self.assertEqual(builder.config.name, "b") |
| |
|
| | with self.assertRaises(ValueError): |
| | DummyBuilderWithMultipleConfigs(cache_dir=tmp_dir) |
| |
|
| | builder = DummyBuilderWithDefaultConfig(cache_dir=tmp_dir) |
| | self.assertEqual(builder.config.name, "a") |
| |
|
| | def test_cache_dir_for_data_dir(self): |
| | with tempfile.TemporaryDirectory() as tmp_dir, tempfile.TemporaryDirectory() as data_dir: |
| | builder = DummyBuilderWithManualDownload(cache_dir=tmp_dir, config_name="a", data_dir=data_dir) |
| | other_builder = DummyBuilderWithManualDownload(cache_dir=tmp_dir, config_name="a", data_dir=data_dir) |
| | self.assertEqual(builder.cache_dir, other_builder.cache_dir) |
| | other_builder = DummyBuilderWithManualDownload(cache_dir=tmp_dir, config_name="a", data_dir=tmp_dir) |
| | self.assertNotEqual(builder.cache_dir, other_builder.cache_dir) |
| |
|
| |
|
| | def test_arrow_based_download_and_prepare(tmp_path): |
| | builder = DummyArrowBasedBuilder(cache_dir=tmp_path) |
| | builder.download_and_prepare() |
| | assert os.path.exists( |
| | os.path.join( |
| | tmp_path, |
| | builder.dataset_name, |
| | "default", |
| | "0.0.0", |
| | f"{builder.dataset_name}-train.arrow", |
| | ) |
| | ) |
| | assert builder.info.features, Features({"text": Value("string")}) |
| | assert builder.info.splits["train"].num_examples == 100 |
| | assert os.path.exists(os.path.join(tmp_path, builder.dataset_name, "default", "0.0.0", "dataset_info.json")) |
| |
|
| |
|
| | @require_beam |
| | def test_beam_based_download_and_prepare(tmp_path): |
| | builder = DummyBeamBasedBuilder(cache_dir=tmp_path, beam_runner="DirectRunner") |
| | builder.download_and_prepare() |
| | assert os.path.exists( |
| | os.path.join( |
| | tmp_path, |
| | builder.dataset_name, |
| | "default", |
| | "0.0.0", |
| | f"{builder.dataset_name}-train.arrow", |
| | ) |
| | ) |
| | assert builder.info.features, Features({"text": Value("string")}) |
| | assert builder.info.splits["train"].num_examples == 100 |
| | assert os.path.exists(os.path.join(tmp_path, builder.dataset_name, "default", "0.0.0", "dataset_info.json")) |
| |
|
| |
|
| | @require_beam |
| | def test_beam_based_as_dataset(tmp_path): |
| | builder = DummyBeamBasedBuilder(cache_dir=tmp_path, beam_runner="DirectRunner") |
| | builder.download_and_prepare() |
| | dataset = builder.as_dataset() |
| | assert dataset |
| | assert isinstance(dataset["train"], Dataset) |
| | assert len(dataset["train"]) > 0 |
| |
|
| |
|
| | @pytest.mark.parametrize( |
| | "split, expected_dataset_class, expected_dataset_length", |
| | [ |
| | (None, DatasetDict, 10), |
| | ("train", Dataset, 10), |
| | ("train+test[:30%]", Dataset, 13), |
| | ], |
| | ) |
| | @pytest.mark.parametrize("in_memory", [False, True]) |
| | def test_builder_as_dataset(split, expected_dataset_class, expected_dataset_length, in_memory, tmp_path): |
| | cache_dir = str(tmp_path) |
| | builder = DummyBuilder(cache_dir=cache_dir) |
| | os.makedirs(builder.cache_dir) |
| |
|
| | builder.info.splits = SplitDict() |
| | builder.info.splits.add(SplitInfo("train", num_examples=10)) |
| | builder.info.splits.add(SplitInfo("test", num_examples=10)) |
| |
|
| | for info_split in builder.info.splits: |
| | with ArrowWriter( |
| | path=os.path.join(builder.cache_dir, f"{builder.dataset_name}-{info_split}.arrow"), |
| | features=Features({"text": Value("string")}), |
| | ) as writer: |
| | writer.write_batch({"text": ["foo"] * 10}) |
| | writer.finalize() |
| |
|
| | with assert_arrow_memory_increases() if in_memory else assert_arrow_memory_doesnt_increase(): |
| | dataset = builder.as_dataset(split=split, in_memory=in_memory) |
| | assert isinstance(dataset, expected_dataset_class) |
| | if isinstance(dataset, DatasetDict): |
| | assert list(dataset.keys()) == ["train", "test"] |
| | datasets = dataset.values() |
| | expected_splits = ["train", "test"] |
| | elif isinstance(dataset, Dataset): |
| | datasets = [dataset] |
| | expected_splits = [split] |
| | for dataset, expected_split in zip(datasets, expected_splits): |
| | assert dataset.split == expected_split |
| | assert len(dataset) == expected_dataset_length |
| | assert dataset.features == Features({"text": Value("string")}) |
| | dataset.column_names == ["text"] |
| |
|
| |
|
| | @pytest.mark.parametrize("in_memory", [False, True]) |
| | def test_generator_based_builder_as_dataset(in_memory, tmp_path): |
| | cache_dir = tmp_path / "data" |
| | cache_dir.mkdir() |
| | cache_dir = str(cache_dir) |
| | builder = DummyGeneratorBasedBuilder(cache_dir=cache_dir) |
| | builder.download_and_prepare(try_from_hf_gcs=False, download_mode=DownloadMode.FORCE_REDOWNLOAD) |
| | with assert_arrow_memory_increases() if in_memory else assert_arrow_memory_doesnt_increase(): |
| | dataset = builder.as_dataset("train", in_memory=in_memory) |
| | assert dataset.data.to_pydict() == {"text": ["foo"] * 100} |
| |
|
| |
|
| | @pytest.mark.parametrize( |
| | "writer_batch_size, default_writer_batch_size, expected_chunks", [(None, None, 1), (None, 5, 20), (10, None, 10)] |
| | ) |
| | def test_custom_writer_batch_size(tmp_path, writer_batch_size, default_writer_batch_size, expected_chunks): |
| | cache_dir = str(tmp_path) |
| | if default_writer_batch_size: |
| | DummyGeneratorBasedBuilder.DEFAULT_WRITER_BATCH_SIZE = default_writer_batch_size |
| | builder = DummyGeneratorBasedBuilder(cache_dir=cache_dir, writer_batch_size=writer_batch_size) |
| | assert builder._writer_batch_size == (writer_batch_size or default_writer_batch_size) |
| | builder.download_and_prepare(try_from_hf_gcs=False, download_mode=DownloadMode.FORCE_REDOWNLOAD) |
| | dataset = builder.as_dataset("train") |
| | assert len(dataset.data[0].chunks) == expected_chunks |
| |
|
| |
|
| | def test_builder_as_streaming_dataset(tmp_path): |
| | dummy_builder = DummyGeneratorBasedBuilder(cache_dir=str(tmp_path)) |
| | check_streaming(dummy_builder) |
| | dsets = dummy_builder.as_streaming_dataset() |
| | assert isinstance(dsets, IterableDatasetDict) |
| | assert isinstance(dsets["train"], IterableDataset) |
| | assert len(list(dsets["train"])) == 100 |
| | dset = dummy_builder.as_streaming_dataset(split="train") |
| | assert isinstance(dset, IterableDataset) |
| | assert len(list(dset)) == 100 |
| |
|
| |
|
| | @require_beam |
| | def test_beam_based_builder_as_streaming_dataset(tmp_path): |
| | builder = DummyBeamBasedBuilder(cache_dir=tmp_path) |
| | check_streaming(builder) |
| | with pytest.raises(DatasetNotOnHfGcsError): |
| | builder.as_streaming_dataset() |
| |
|
| |
|
| | def _run_test_builder_streaming_works_in_subprocesses(builder): |
| | check_streaming(builder) |
| | dset = builder.as_streaming_dataset(split="train") |
| | assert isinstance(dset, IterableDataset) |
| | assert len(list(dset)) == 100 |
| |
|
| |
|
| | def test_builder_streaming_works_in_subprocess(tmp_path): |
| | dummy_builder = DummyGeneratorBasedBuilder(cache_dir=str(tmp_path)) |
| | p = Process(target=_run_test_builder_streaming_works_in_subprocesses, args=(dummy_builder,)) |
| | p.start() |
| | p.join() |
| |
|
| |
|
| | class DummyBuilderWithVersion(GeneratorBasedBuilder): |
| | VERSION = "2.0.0" |
| |
|
| | def _info(self): |
| | return DatasetInfo(features=Features({"text": Value("string")})) |
| |
|
| | def _split_generators(self, dl_manager): |
| | pass |
| |
|
| | def _generate_examples(self): |
| | pass |
| |
|
| |
|
| | class DummyBuilderWithBuilderConfigs(GeneratorBasedBuilder): |
| | BUILDER_CONFIGS = [BuilderConfig(name="custom", version="2.0.0")] |
| |
|
| | def _info(self): |
| | return DatasetInfo(features=Features({"text": Value("string")})) |
| |
|
| | def _split_generators(self, dl_manager): |
| | pass |
| |
|
| | def _generate_examples(self): |
| | pass |
| |
|
| |
|
| | class CustomBuilderConfig(BuilderConfig): |
| | def __init__(self, date=None, language=None, version="2.0.0", **kwargs): |
| | name = f"{date}.{language}" |
| | super().__init__(name=name, version=version, **kwargs) |
| | self.date = date |
| | self.language = language |
| |
|
| |
|
| | class DummyBuilderWithCustomBuilderConfigs(GeneratorBasedBuilder): |
| | BUILDER_CONFIGS = [CustomBuilderConfig(date="20220501", language="en")] |
| | BUILDER_CONFIG_CLASS = CustomBuilderConfig |
| |
|
| | def _info(self): |
| | return DatasetInfo(features=Features({"text": Value("string")})) |
| |
|
| | def _split_generators(self, dl_manager): |
| | pass |
| |
|
| | def _generate_examples(self): |
| | pass |
| |
|
| |
|
| | @pytest.mark.parametrize( |
| | "builder_class, kwargs", |
| | [ |
| | (DummyBuilderWithVersion, {}), |
| | (DummyBuilderWithBuilderConfigs, {"config_name": "custom"}), |
| | (DummyBuilderWithCustomBuilderConfigs, {"config_name": "20220501.en"}), |
| | (DummyBuilderWithCustomBuilderConfigs, {"date": "20220501", "language": "ca"}), |
| | ], |
| | ) |
| | def test_builder_config_version(builder_class, kwargs, tmp_path): |
| | cache_dir = str(tmp_path) |
| | builder = builder_class(cache_dir=cache_dir, **kwargs) |
| | assert builder.config.version == "2.0.0" |
| |
|
| |
|
| | def test_builder_download_and_prepare_with_absolute_output_dir(tmp_path): |
| | builder = DummyGeneratorBasedBuilder() |
| | output_dir = str(tmp_path) |
| | builder.download_and_prepare(output_dir) |
| | assert builder._output_dir.startswith(tmp_path.resolve().as_posix()) |
| | assert os.path.exists(os.path.join(output_dir, "dataset_info.json")) |
| | assert os.path.exists(os.path.join(output_dir, f"{builder.dataset_name}-train.arrow")) |
| | assert not os.path.exists(os.path.join(output_dir + ".incomplete")) |
| |
|
| |
|
| | def test_builder_download_and_prepare_with_relative_output_dir(): |
| | with set_current_working_directory_to_temp_dir(): |
| | builder = DummyGeneratorBasedBuilder() |
| | output_dir = "test-out" |
| | builder.download_and_prepare(output_dir) |
| | assert Path(builder._output_dir).resolve().as_posix().startswith(Path(output_dir).resolve().as_posix()) |
| | assert os.path.exists(os.path.join(output_dir, "dataset_info.json")) |
| | assert os.path.exists(os.path.join(output_dir, f"{builder.dataset_name}-train.arrow")) |
| | assert not os.path.exists(os.path.join(output_dir + ".incomplete")) |
| |
|
| |
|
| | def test_builder_with_filesystem_download_and_prepare(tmp_path, mockfs): |
| | builder = DummyGeneratorBasedBuilder(cache_dir=tmp_path) |
| | builder.download_and_prepare("mock://my_dataset", storage_options=mockfs.storage_options) |
| | assert builder._output_dir.startswith("mock://my_dataset") |
| | assert is_local_path(builder._cache_downloaded_dir) |
| | assert isinstance(builder._fs, type(mockfs)) |
| | assert builder._fs.storage_options == mockfs.storage_options |
| | assert mockfs.exists("my_dataset/dataset_info.json") |
| | assert mockfs.exists(f"my_dataset/{builder.dataset_name}-train.arrow") |
| | assert not mockfs.exists("my_dataset.incomplete") |
| |
|
| |
|
| | def test_builder_with_filesystem_download_and_prepare_reload(tmp_path, mockfs, caplog): |
| | builder = DummyGeneratorBasedBuilder(cache_dir=tmp_path) |
| | mockfs.makedirs("my_dataset") |
| | DatasetInfo().write_to_directory("mock://my_dataset", storage_options=mockfs.storage_options) |
| | mockfs.touch(f"my_dataset/{builder.dataset_name}-train.arrow") |
| | caplog.clear() |
| | with caplog.at_level(INFO, logger=get_logger().name): |
| | builder.download_and_prepare("mock://my_dataset", storage_options=mockfs.storage_options) |
| | assert "Found cached dataset" in caplog.text |
| |
|
| |
|
| | def test_generator_based_builder_download_and_prepare_as_parquet(tmp_path): |
| | builder = DummyGeneratorBasedBuilder(cache_dir=tmp_path) |
| | builder.download_and_prepare(file_format="parquet") |
| | assert builder.info.splits["train"].num_examples == 100 |
| | parquet_path = os.path.join( |
| | tmp_path, builder.dataset_name, "default", "0.0.0", f"{builder.dataset_name}-train.parquet" |
| | ) |
| | assert os.path.exists(parquet_path) |
| | assert pq.ParquetFile(parquet_path) is not None |
| |
|
| |
|
| | def test_generator_based_builder_download_and_prepare_sharded(tmp_path): |
| | writer_batch_size = 25 |
| | builder = DummyGeneratorBasedBuilder(cache_dir=tmp_path, writer_batch_size=writer_batch_size) |
| | with patch("datasets.config.MAX_SHARD_SIZE", 1): |
| | builder.download_and_prepare(file_format="parquet") |
| | expected_num_shards = 100 // writer_batch_size |
| | assert builder.info.splits["train"].num_examples == 100 |
| | parquet_path = os.path.join( |
| | tmp_path, |
| | builder.dataset_name, |
| | "default", |
| | "0.0.0", |
| | f"{builder.dataset_name}-train-00000-of-{expected_num_shards:05d}.parquet", |
| | ) |
| | assert os.path.exists(parquet_path) |
| | parquet_files = [ |
| | pq.ParquetFile(parquet_path) |
| | for parquet_path in Path(tmp_path).rglob( |
| | f"{builder.dataset_name}-train-*-of-{expected_num_shards:05d}.parquet" |
| | ) |
| | ] |
| | assert len(parquet_files) == expected_num_shards |
| | assert sum(parquet_file.metadata.num_rows for parquet_file in parquet_files) == 100 |
| |
|
| |
|
| | def test_generator_based_builder_download_and_prepare_with_max_shard_size(tmp_path): |
| | writer_batch_size = 25 |
| | builder = DummyGeneratorBasedBuilder(cache_dir=tmp_path, writer_batch_size=writer_batch_size) |
| | builder.download_and_prepare(file_format="parquet", max_shard_size=1) |
| | expected_num_shards = 100 // writer_batch_size |
| | assert builder.info.splits["train"].num_examples == 100 |
| | parquet_path = os.path.join( |
| | tmp_path, |
| | builder.dataset_name, |
| | "default", |
| | "0.0.0", |
| | f"{builder.dataset_name}-train-00000-of-{expected_num_shards:05d}.parquet", |
| | ) |
| | assert os.path.exists(parquet_path) |
| | parquet_files = [ |
| | pq.ParquetFile(parquet_path) |
| | for parquet_path in Path(tmp_path).rglob( |
| | f"{builder.dataset_name}-train-*-of-{expected_num_shards:05d}.parquet" |
| | ) |
| | ] |
| | assert len(parquet_files) == expected_num_shards |
| | assert sum(parquet_file.metadata.num_rows for parquet_file in parquet_files) == 100 |
| |
|
| |
|
| | def test_generator_based_builder_download_and_prepare_with_num_proc(tmp_path): |
| | builder = DummyGeneratorBasedBuilderWithShards(cache_dir=tmp_path) |
| | builder.download_and_prepare(num_proc=2) |
| | expected_num_shards = 2 |
| | assert builder.info.splits["train"].num_examples == 400 |
| | assert builder.info.splits["train"].shard_lengths == [200, 200] |
| | arrow_path = os.path.join( |
| | tmp_path, |
| | builder.dataset_name, |
| | "default", |
| | "0.0.0", |
| | f"{builder.dataset_name}-train-00000-of-{expected_num_shards:05d}.arrow", |
| | ) |
| | assert os.path.exists(arrow_path) |
| | ds = builder.as_dataset("train") |
| | assert len(ds) == 400 |
| | assert ds.to_dict() == { |
| | "id": [i for _ in range(4) for i in range(100)], |
| | "filepath": [f"data{i}.txt" for i in range(4) for _ in range(100)], |
| | } |
| |
|
| |
|
| | @pytest.mark.parametrize( |
| | "num_proc, expectation", [(None, does_not_raise()), (1, does_not_raise()), (2, pytest.raises(RuntimeError))] |
| | ) |
| | def test_generator_based_builder_download_and_prepare_with_ambiguous_shards(num_proc, expectation, tmp_path): |
| | builder = DummyGeneratorBasedBuilderWithAmbiguousShards(cache_dir=tmp_path) |
| | with expectation: |
| | builder.download_and_prepare(num_proc=num_proc) |
| |
|
| |
|
| | def test_arrow_based_builder_download_and_prepare_as_parquet(tmp_path): |
| | builder = DummyArrowBasedBuilder(cache_dir=tmp_path) |
| | builder.download_and_prepare(file_format="parquet") |
| | assert builder.info.splits["train"].num_examples == 100 |
| | parquet_path = os.path.join( |
| | tmp_path, builder.dataset_name, "default", "0.0.0", f"{builder.dataset_name}-train.parquet" |
| | ) |
| | assert os.path.exists(parquet_path) |
| | assert pq.ParquetFile(parquet_path) is not None |
| |
|
| |
|
| | def test_arrow_based_builder_download_and_prepare_sharded(tmp_path): |
| | builder = DummyArrowBasedBuilder(cache_dir=tmp_path) |
| | with patch("datasets.config.MAX_SHARD_SIZE", 1): |
| | builder.download_and_prepare(file_format="parquet") |
| | expected_num_shards = 10 |
| | assert builder.info.splits["train"].num_examples == 100 |
| | parquet_path = os.path.join( |
| | tmp_path, |
| | builder.dataset_name, |
| | "default", |
| | "0.0.0", |
| | f"{builder.dataset_name}-train-00000-of-{expected_num_shards:05d}.parquet", |
| | ) |
| | assert os.path.exists(parquet_path) |
| | parquet_files = [ |
| | pq.ParquetFile(parquet_path) |
| | for parquet_path in Path(tmp_path).rglob( |
| | f"{builder.dataset_name}-train-*-of-{expected_num_shards:05d}.parquet" |
| | ) |
| | ] |
| | assert len(parquet_files) == expected_num_shards |
| | assert sum(parquet_file.metadata.num_rows for parquet_file in parquet_files) == 100 |
| |
|
| |
|
| | def test_arrow_based_builder_download_and_prepare_with_max_shard_size(tmp_path): |
| | builder = DummyArrowBasedBuilder(cache_dir=tmp_path) |
| | builder.download_and_prepare(file_format="parquet", max_shard_size=1) |
| | expected_num_shards = 10 |
| | assert builder.info.splits["train"].num_examples == 100 |
| | parquet_path = os.path.join( |
| | tmp_path, |
| | builder.dataset_name, |
| | "default", |
| | "0.0.0", |
| | f"{builder.dataset_name}-train-00000-of-{expected_num_shards:05d}.parquet", |
| | ) |
| | assert os.path.exists(parquet_path) |
| | parquet_files = [ |
| | pq.ParquetFile(parquet_path) |
| | for parquet_path in Path(tmp_path).rglob( |
| | f"{builder.dataset_name}-train-*-of-{expected_num_shards:05d}.parquet" |
| | ) |
| | ] |
| | assert len(parquet_files) == expected_num_shards |
| | assert sum(parquet_file.metadata.num_rows for parquet_file in parquet_files) == 100 |
| |
|
| |
|
| | def test_arrow_based_builder_download_and_prepare_with_num_proc(tmp_path): |
| | builder = DummyArrowBasedBuilderWithShards(cache_dir=tmp_path) |
| | builder.download_and_prepare(num_proc=2) |
| | expected_num_shards = 2 |
| | assert builder.info.splits["train"].num_examples == 400 |
| | assert builder.info.splits["train"].shard_lengths == [200, 200] |
| | arrow_path = os.path.join( |
| | tmp_path, |
| | builder.dataset_name, |
| | "default", |
| | "0.0.0", |
| | f"{builder.dataset_name}-train-00000-of-{expected_num_shards:05d}.arrow", |
| | ) |
| | assert os.path.exists(arrow_path) |
| | ds = builder.as_dataset("train") |
| | assert len(ds) == 400 |
| | assert ds.to_dict() == { |
| | "id": [i for _ in range(4) for i in range(100)], |
| | "filepath": [f"data{i}.txt" for i in range(4) for _ in range(100)], |
| | } |
| |
|
| |
|
| | @pytest.mark.parametrize( |
| | "num_proc, expectation", [(None, does_not_raise()), (1, does_not_raise()), (2, pytest.raises(RuntimeError))] |
| | ) |
| | def test_arrow_based_builder_download_and_prepare_with_ambiguous_shards(num_proc, expectation, tmp_path): |
| | builder = DummyArrowBasedBuilderWithAmbiguousShards(cache_dir=tmp_path) |
| | with expectation: |
| | builder.download_and_prepare(num_proc=num_proc) |
| |
|
| |
|
| | @require_beam |
| | def test_beam_based_builder_download_and_prepare_as_parquet(tmp_path): |
| | builder = DummyBeamBasedBuilder(cache_dir=tmp_path, beam_runner="DirectRunner") |
| | builder.download_and_prepare(file_format="parquet") |
| | assert builder.info.splits["train"].num_examples == 100 |
| | parquet_path = os.path.join( |
| | tmp_path, builder.dataset_name, "default", "0.0.0", f"{builder.dataset_name}-train.parquet" |
| | ) |
| | assert os.path.exists(parquet_path) |
| | assert pq.ParquetFile(parquet_path) is not None |
| |
|