| import csv |
| import os |
|
|
| import pytest |
|
|
| from datasets import Dataset, DatasetDict, Features, NamedSplit, Value |
| from datasets.io.csv import CsvDatasetReader, CsvDatasetWriter |
|
|
| from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases |
|
|
|
|
| def _check_csv_dataset(dataset, expected_features): |
| assert isinstance(dataset, Dataset) |
| assert dataset.num_rows == 4 |
| assert dataset.num_columns == 3 |
| assert dataset.column_names == ["col_1", "col_2", "col_3"] |
| for feature, expected_dtype in expected_features.items(): |
| assert dataset.features[feature].dtype == expected_dtype |
|
|
|
|
| @pytest.mark.parametrize("keep_in_memory", [False, True]) |
| def test_dataset_from_csv_keep_in_memory(keep_in_memory, csv_path, tmp_path): |
| cache_dir = tmp_path / "cache" |
| expected_features = {"col_1": "int64", "col_2": "int64", "col_3": "float64"} |
| with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): |
| dataset = CsvDatasetReader(csv_path, cache_dir=cache_dir, keep_in_memory=keep_in_memory).read() |
| _check_csv_dataset(dataset, expected_features) |
|
|
|
|
| @pytest.mark.parametrize( |
| "features", |
| [ |
| None, |
| {"col_1": "string", "col_2": "int64", "col_3": "float64"}, |
| {"col_1": "string", "col_2": "string", "col_3": "string"}, |
| {"col_1": "int32", "col_2": "int32", "col_3": "int32"}, |
| {"col_1": "float32", "col_2": "float32", "col_3": "float32"}, |
| ], |
| ) |
| def test_dataset_from_csv_features(features, csv_path, tmp_path): |
| cache_dir = tmp_path / "cache" |
| |
| default_expected_features = {"col_1": "int64", "col_2": "int64", "col_3": "float64"} |
| expected_features = features.copy() if features else default_expected_features |
| features = ( |
| Features({feature: Value(dtype) for feature, dtype in features.items()}) if features is not None else None |
| ) |
| dataset = CsvDatasetReader(csv_path, features=features, cache_dir=cache_dir).read() |
| _check_csv_dataset(dataset, expected_features) |
|
|
|
|
| @pytest.mark.parametrize("split", [None, NamedSplit("train"), "train", "test"]) |
| def test_dataset_from_csv_split(split, csv_path, tmp_path): |
| cache_dir = tmp_path / "cache" |
| expected_features = {"col_1": "int64", "col_2": "int64", "col_3": "float64"} |
| dataset = CsvDatasetReader(csv_path, cache_dir=cache_dir, split=split).read() |
| _check_csv_dataset(dataset, expected_features) |
| assert dataset.split == split if split else "train" |
|
|
|
|
| @pytest.mark.parametrize("path_type", [str, list]) |
| def test_dataset_from_csv_path_type(path_type, csv_path, tmp_path): |
| if issubclass(path_type, str): |
| path = csv_path |
| elif issubclass(path_type, list): |
| path = [csv_path] |
| cache_dir = tmp_path / "cache" |
| expected_features = {"col_1": "int64", "col_2": "int64", "col_3": "float64"} |
| dataset = CsvDatasetReader(path, cache_dir=cache_dir).read() |
| _check_csv_dataset(dataset, expected_features) |
|
|
|
|
| def _check_csv_datasetdict(dataset_dict, expected_features, splits=("train",)): |
| assert isinstance(dataset_dict, DatasetDict) |
| for split in splits: |
| dataset = dataset_dict[split] |
| assert dataset.num_rows == 4 |
| assert dataset.num_columns == 3 |
| assert dataset.column_names == ["col_1", "col_2", "col_3"] |
| for feature, expected_dtype in expected_features.items(): |
| assert dataset.features[feature].dtype == expected_dtype |
|
|
|
|
| @pytest.mark.parametrize("keep_in_memory", [False, True]) |
| def test_csv_datasetdict_reader_keep_in_memory(keep_in_memory, csv_path, tmp_path): |
| cache_dir = tmp_path / "cache" |
| expected_features = {"col_1": "int64", "col_2": "int64", "col_3": "float64"} |
| with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): |
| dataset = CsvDatasetReader({"train": csv_path}, cache_dir=cache_dir, keep_in_memory=keep_in_memory).read() |
| _check_csv_datasetdict(dataset, expected_features) |
|
|
|
|
| @pytest.mark.parametrize( |
| "features", |
| [ |
| None, |
| {"col_1": "string", "col_2": "int64", "col_3": "float64"}, |
| {"col_1": "string", "col_2": "string", "col_3": "string"}, |
| {"col_1": "int32", "col_2": "int32", "col_3": "int32"}, |
| {"col_1": "float32", "col_2": "float32", "col_3": "float32"}, |
| ], |
| ) |
| def test_csv_datasetdict_reader_features(features, csv_path, tmp_path): |
| cache_dir = tmp_path / "cache" |
| |
| default_expected_features = {"col_1": "int64", "col_2": "int64", "col_3": "float64"} |
| expected_features = features.copy() if features else default_expected_features |
| features = ( |
| Features({feature: Value(dtype) for feature, dtype in features.items()}) if features is not None else None |
| ) |
| dataset = CsvDatasetReader({"train": csv_path}, features=features, cache_dir=cache_dir).read() |
| _check_csv_datasetdict(dataset, expected_features) |
|
|
|
|
| @pytest.mark.parametrize("split", [None, NamedSplit("train"), "train", "test"]) |
| def test_csv_datasetdict_reader_split(split, csv_path, tmp_path): |
| if split: |
| path = {split: csv_path} |
| else: |
| path = {"train": csv_path, "test": csv_path} |
| cache_dir = tmp_path / "cache" |
| expected_features = {"col_1": "int64", "col_2": "int64", "col_3": "float64"} |
| dataset = CsvDatasetReader(path, cache_dir=cache_dir).read() |
| _check_csv_datasetdict(dataset, expected_features, splits=list(path.keys())) |
| assert all(dataset[split].split == split for split in path.keys()) |
|
|
|
|
| def iter_csv_file(csv_path): |
| with open(csv_path, encoding="utf-8") as csvfile: |
| yield from csv.reader(csvfile) |
|
|
|
|
| def test_dataset_to_csv(csv_path, tmp_path): |
| cache_dir = tmp_path / "cache" |
| output_csv = os.path.join(cache_dir, "tmp.csv") |
| dataset = CsvDatasetReader({"train": csv_path}, cache_dir=cache_dir).read() |
| CsvDatasetWriter(dataset["train"], output_csv, num_proc=1).write() |
|
|
| original_csv = iter_csv_file(csv_path) |
| expected_csv = iter_csv_file(output_csv) |
|
|
| for row1, row2 in zip(original_csv, expected_csv): |
| assert row1 == row2 |
|
|
|
|
| def test_dataset_to_csv_multiproc(csv_path, tmp_path): |
| cache_dir = tmp_path / "cache" |
| output_csv = os.path.join(cache_dir, "tmp.csv") |
| dataset = CsvDatasetReader({"train": csv_path}, cache_dir=cache_dir).read() |
| CsvDatasetWriter(dataset["train"], output_csv, num_proc=2).write() |
|
|
| original_csv = iter_csv_file(csv_path) |
| expected_csv = iter_csv_file(output_csv) |
|
|
| for row1, row2 in zip(original_csv, expected_csv): |
| assert row1 == row2 |
|
|
|
|
| def test_dataset_to_csv_invalidproc(csv_path, tmp_path): |
| cache_dir = tmp_path / "cache" |
| output_csv = os.path.join(cache_dir, "tmp.csv") |
| dataset = CsvDatasetReader({"train": csv_path}, cache_dir=cache_dir).read() |
| with pytest.raises(ValueError): |
| CsvDatasetWriter(dataset["train"], output_csv, num_proc=0) |
|
|