| | import os |
| | import tempfile |
| | from pathlib import Path |
| | from unittest import TestCase |
| |
|
| | import pyarrow as pa |
| | import pytest |
| |
|
| | from datasets.arrow_dataset import Dataset |
| | from datasets.arrow_reader import ArrowReader, BaseReader, FileInstructions, ReadInstruction, make_file_instructions |
| | from datasets.info import DatasetInfo |
| | from datasets.splits import NamedSplit, Split, SplitDict, SplitInfo |
| |
|
| | from .utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases |
| |
|
| |
|
| | class ReaderTest(BaseReader): |
| | """ |
| | Build a Dataset object out of Instruction instance(s). |
| | This reader is made for testing. It mocks file reads. |
| | """ |
| |
|
| | def _get_table_from_filename(self, filename_skip_take, in_memory=False): |
| | """Returns a Dataset instance from given (filename, skip, take).""" |
| | filename, skip, take = ( |
| | filename_skip_take["filename"], |
| | filename_skip_take["skip"] if "skip" in filename_skip_take else None, |
| | filename_skip_take["take"] if "take" in filename_skip_take else None, |
| | ) |
| | open(os.path.join(filename), "wb").close() |
| | pa_table = pa.Table.from_pydict({"filename": [Path(filename).name] * 100}) |
| | if take == -1: |
| | take = len(pa_table) - skip |
| | if skip is not None and take is not None: |
| | pa_table = pa_table.slice(skip, take) |
| | return pa_table |
| |
|
| |
|
| | class BaseReaderTest(TestCase): |
| | def test_read(self): |
| | name = "my_name" |
| | train_info = SplitInfo(name="train", num_examples=100) |
| | test_info = SplitInfo(name="test", num_examples=100) |
| | split_infos = [train_info, test_info] |
| | split_dict = SplitDict() |
| | split_dict.add(train_info) |
| | split_dict.add(test_info) |
| | info = DatasetInfo(splits=split_dict) |
| |
|
| | with tempfile.TemporaryDirectory() as tmp_dir: |
| | reader = ReaderTest(tmp_dir, info) |
| |
|
| | instructions = "test[:33%]" |
| | dset = Dataset(**reader.read(name, instructions, split_infos)) |
| | self.assertEqual(dset["filename"][0], f"{name}-test") |
| | self.assertEqual(dset.num_rows, 33) |
| | self.assertEqual(dset.num_columns, 1) |
| |
|
| | instructions1 = ["train", "test[:33%]"] |
| | instructions2 = [Split.TRAIN, ReadInstruction.from_spec("test[:33%]")] |
| | for instructions in [instructions1, instructions2]: |
| | datasets_kwargs = [reader.read(name, instr, split_infos) for instr in instructions] |
| | train_dset, test_dset = (Dataset(**dataset_kwargs) for dataset_kwargs in datasets_kwargs) |
| | self.assertEqual(train_dset["filename"][0], f"{name}-train") |
| | self.assertEqual(train_dset.num_rows, 100) |
| | self.assertEqual(train_dset.num_columns, 1) |
| | self.assertIsInstance(train_dset.split, NamedSplit) |
| | self.assertEqual(str(train_dset.split), "train") |
| | self.assertEqual(test_dset["filename"][0], f"{name}-test") |
| | self.assertEqual(test_dset.num_rows, 33) |
| | self.assertEqual(test_dset.num_columns, 1) |
| | self.assertIsInstance(test_dset.split, NamedSplit) |
| | self.assertEqual(str(test_dset.split), "test[:33%]") |
| | del train_dset, test_dset |
| |
|
| | def test_read_sharded(self): |
| | name = "my_name" |
| | train_info = SplitInfo(name="train", num_examples=1000, shard_lengths=[100] * 10) |
| | split_infos = [train_info] |
| | split_dict = SplitDict() |
| | split_dict.add(train_info) |
| | info = DatasetInfo(splits=split_dict) |
| |
|
| | with tempfile.TemporaryDirectory() as tmp_dir: |
| | reader = ReaderTest(tmp_dir, info) |
| |
|
| | instructions = "train[:33%]" |
| | dset = Dataset(**reader.read(name, instructions, split_infos)) |
| | self.assertEqual(dset["filename"][0], f"{name}-train-00000-of-00010") |
| | self.assertEqual(dset["filename"][-1], f"{name}-train-00003-of-00010") |
| | self.assertEqual(dset.num_rows, 330) |
| | self.assertEqual(dset.num_columns, 1) |
| |
|
| | def test_read_files(self): |
| | train_info = SplitInfo(name="train", num_examples=100) |
| | test_info = SplitInfo(name="test", num_examples=100) |
| | split_dict = SplitDict() |
| | split_dict.add(train_info) |
| | split_dict.add(test_info) |
| | info = DatasetInfo(splits=split_dict) |
| |
|
| | with tempfile.TemporaryDirectory() as tmp_dir: |
| | reader = ReaderTest(tmp_dir, info) |
| |
|
| | files = [ |
| | {"filename": os.path.join(tmp_dir, "train")}, |
| | {"filename": os.path.join(tmp_dir, "test"), "skip": 10, "take": 10}, |
| | ] |
| | dset = Dataset(**reader.read_files(files, original_instructions="train+test[10:20]")) |
| | self.assertEqual(dset.num_rows, 110) |
| | self.assertEqual(dset.num_columns, 1) |
| | del dset |
| |
|
| |
|
| | @pytest.mark.parametrize("in_memory", [False, True]) |
| | def test_read_table(in_memory, dataset, arrow_file): |
| | filename = arrow_file |
| | with assert_arrow_memory_increases() if in_memory else assert_arrow_memory_doesnt_increase(): |
| | table = ArrowReader.read_table(filename, in_memory=in_memory) |
| | assert table.shape == dataset.data.shape |
| | assert set(table.column_names) == set(dataset.data.column_names) |
| | assert dict(table.to_pydict()) == dict(dataset.data.to_pydict()) |
| |
|
| |
|
| | @pytest.mark.parametrize("in_memory", [False, True]) |
| | def test_read_files(in_memory, dataset, arrow_file): |
| | filename = arrow_file |
| | reader = ArrowReader("", None) |
| | with assert_arrow_memory_increases() if in_memory else assert_arrow_memory_doesnt_increase(): |
| | dataset_kwargs = reader.read_files([{"filename": filename}], in_memory=in_memory) |
| | assert dataset_kwargs.keys() == {"arrow_table", "info", "split"} |
| | table = dataset_kwargs["arrow_table"] |
| | assert table.shape == dataset.data.shape |
| | assert set(table.column_names) == set(dataset.data.column_names) |
| | assert dict(table.to_pydict()) == dict(dataset.data.to_pydict()) |
| |
|
| |
|
| | def test_read_instruction_spec(): |
| | assert ReadInstruction("train", to=10, unit="abs").to_spec() == "train[:10]" |
| | assert ReadInstruction("train", from_=-80, to=10, unit="%").to_spec() == "train[-80%:10%]" |
| |
|
| | spec_train_test = "train+test" |
| | assert ReadInstruction.from_spec(spec_train_test).to_spec() == spec_train_test |
| |
|
| | spec_train_abs = "train[2:10]" |
| | assert ReadInstruction.from_spec(spec_train_abs).to_spec() == spec_train_abs |
| |
|
| | spec_train_pct = "train[15%:-20%]" |
| | assert ReadInstruction.from_spec(spec_train_pct).to_spec() == spec_train_pct |
| |
|
| | spec_train_pct_rounding = "train[:10%](closest)" |
| | assert ReadInstruction.from_spec(spec_train_pct_rounding).to_spec() == "train[:10%]" |
| |
|
| | spec_train_pct_rounding = "train[:10%](pct1_dropremainder)" |
| | assert ReadInstruction.from_spec(spec_train_pct_rounding).to_spec() == spec_train_pct_rounding |
| |
|
| | spec_train_test_pct_rounding = "train[:10%](pct1_dropremainder)+test[-10%:](pct1_dropremainder)" |
| | assert ReadInstruction.from_spec(spec_train_test_pct_rounding).to_spec() == spec_train_test_pct_rounding |
| |
|
| |
|
| | def test_make_file_instructions(): |
| | name = "dummy" |
| | split_infos = [SplitInfo(name="train", num_examples=100)] |
| | instruction = "train[:33%]" |
| | filetype_suffix = "arrow" |
| | prefix_path = "prefix" |
| |
|
| | file_instructions = make_file_instructions(name, split_infos, instruction, filetype_suffix, prefix_path) |
| | assert isinstance(file_instructions, FileInstructions) |
| | assert file_instructions.num_examples == 33 |
| | assert file_instructions.file_instructions == [ |
| | {"filename": os.path.join(prefix_path, f"{name}-train.arrow"), "skip": 0, "take": 33} |
| | ] |
| |
|
| | split_infos = [SplitInfo(name="train", num_examples=100, shard_lengths=[10] * 10)] |
| | file_instructions = make_file_instructions(name, split_infos, instruction, filetype_suffix, prefix_path) |
| | assert isinstance(file_instructions, FileInstructions) |
| | assert file_instructions.num_examples == 33 |
| | assert file_instructions.file_instructions == [ |
| | {"filename": os.path.join(prefix_path, f"{name}-train-00000-of-00010.arrow"), "skip": 0, "take": -1}, |
| | {"filename": os.path.join(prefix_path, f"{name}-train-00001-of-00010.arrow"), "skip": 0, "take": -1}, |
| | {"filename": os.path.join(prefix_path, f"{name}-train-00002-of-00010.arrow"), "skip": 0, "take": -1}, |
| | {"filename": os.path.join(prefix_path, f"{name}-train-00003-of-00010.arrow"), "skip": 0, "take": 3}, |
| | ] |
| |
|
| |
|
| | @pytest.mark.parametrize("name, expected_exception", [(None, TypeError), ("", ValueError)]) |
| | def test_make_file_instructions_raises(name, expected_exception): |
| | split_infos = [SplitInfo(name="train", num_examples=100)] |
| | instruction = "train" |
| | filetype_suffix = "arrow" |
| | prefix_path = "prefix_path" |
| | with pytest.raises(expected_exception): |
| | _ = make_file_instructions(name, split_infos, instruction, filetype_suffix, prefix_path) |
| |
|