| import datetime |
| from unittest import TestCase |
| from unittest.mock import patch |
|
|
| import numpy as np |
| import pandas as pd |
| import pyarrow as pa |
| import pytest |
|
|
| from datasets import Array2D |
| from datasets.arrow_dataset import Dataset |
| from datasets.features import Audio, ClassLabel, Features, Image, Sequence, Value |
| from datasets.features.features import ( |
| _arrow_to_datasets_dtype, |
| _cast_to_python_objects, |
| cast_to_python_objects, |
| encode_nested_example, |
| generate_from_dict, |
| string_to_arrow, |
| ) |
| from datasets.features.translation import Translation, TranslationVariableLanguages |
| from datasets.info import DatasetInfo |
| from datasets.utils.py_utils import asdict |
|
|
| from ..utils import require_jax, require_tf, require_torch |
|
|
|
|
| class FeaturesTest(TestCase): |
| def test_from_arrow_schema_simple(self): |
| data = {"a": [{"b": {"c": "text"}}] * 10, "foo": [1] * 10} |
| original_features = Features({"a": {"b": {"c": Value("string")}}, "foo": Value("int64")}) |
| dset = Dataset.from_dict(data, features=original_features) |
| new_features = dset.features |
| new_dset = Dataset.from_dict(data, features=new_features) |
| self.assertEqual(original_features.type, new_features.type) |
| self.assertDictEqual(dset[0], new_dset[0]) |
| self.assertDictEqual(dset[:], new_dset[:]) |
|
|
| def test_from_arrow_schema_with_sequence(self): |
| data = {"a": [{"b": {"c": ["text"]}}] * 10, "foo": [1] * 10} |
| original_features = Features({"a": {"b": Sequence({"c": Value("string")})}, "foo": Value("int64")}) |
| dset = Dataset.from_dict(data, features=original_features) |
| new_features = dset.features |
| new_dset = Dataset.from_dict(data, features=new_features) |
| self.assertEqual(original_features.type, new_features.type) |
| self.assertDictEqual(dset[0], new_dset[0]) |
| self.assertDictEqual(dset[:], new_dset[:]) |
|
|
| def test_string_to_arrow_bijection_for_primitive_types(self): |
| supported_pyarrow_datatypes = [ |
| pa.time32("s"), |
| pa.time64("us"), |
| pa.timestamp("s"), |
| pa.timestamp("ns", tz="America/New_York"), |
| pa.date32(), |
| pa.date64(), |
| pa.duration("s"), |
| pa.decimal128(10, 2), |
| pa.decimal256(40, -3), |
| pa.string(), |
| pa.int32(), |
| pa.float64(), |
| pa.array([datetime.time(1, 1, 1)]).type, |
| ] |
| for dt in supported_pyarrow_datatypes: |
| self.assertEqual(dt, string_to_arrow(_arrow_to_datasets_dtype(dt))) |
|
|
| unsupported_pyarrow_datatypes = [pa.list_(pa.float64())] |
| for dt in unsupported_pyarrow_datatypes: |
| with self.assertRaises(ValueError): |
| string_to_arrow(_arrow_to_datasets_dtype(dt)) |
|
|
| supported_datasets_dtypes = [ |
| "time32[s]", |
| "timestamp[ns]", |
| "timestamp[ns, tz=+07:30]", |
| "duration[us]", |
| "decimal128(30, -4)", |
| "int32", |
| "float64", |
| ] |
| for sdt in supported_datasets_dtypes: |
| self.assertEqual(sdt, _arrow_to_datasets_dtype(string_to_arrow(sdt))) |
|
|
| unsupported_datasets_dtypes = [ |
| "time32[ns]", |
| "timestamp[blob]", |
| "timestamp[[ns]]", |
| "timestamp[ns, tz=[ns]]", |
| "duration[[us]]", |
| "decimal20(30, -4)", |
| "int", |
| ] |
| for sdt in unsupported_datasets_dtypes: |
| with self.assertRaises(ValueError): |
| string_to_arrow(sdt) |
|
|
| def test_feature_named_type(self): |
| """reference: issue #1110""" |
| features = Features({"_type": Value("string")}) |
| ds_info = DatasetInfo(features=features) |
| reloaded_features = Features.from_dict(asdict(ds_info)["features"]) |
| assert features == reloaded_features |
|
|
| def test_feature_named_self_as_kwarg(self): |
| """reference: issue #5641""" |
| features = Features(self=Value("string")) |
| ds_info = DatasetInfo(features=features) |
| reloaded_features = Features.from_dict(asdict(ds_info)["features"]) |
| assert features == reloaded_features |
|
|
| def test_class_label_feature_with_no_labels(self): |
| """reference: issue #4681""" |
| features = Features({"label": ClassLabel(names=[])}) |
| ds_info = DatasetInfo(features=features) |
| reloaded_features = Features.from_dict(asdict(ds_info)["features"]) |
| assert features == reloaded_features |
|
|
| def test_reorder_fields_as(self): |
| features = Features( |
| { |
| "id": Value("string"), |
| "document": { |
| "title": Value("string"), |
| "url": Value("string"), |
| "html": Value("string"), |
| "tokens": Sequence({"token": Value("string"), "is_html": Value("bool")}), |
| }, |
| "question": { |
| "text": Value("string"), |
| "tokens": Sequence(Value("string")), |
| }, |
| "annotations": Sequence( |
| { |
| "id": Value("string"), |
| "long_answer": { |
| "start_token": Value("int64"), |
| "end_token": Value("int64"), |
| "start_byte": Value("int64"), |
| "end_byte": Value("int64"), |
| }, |
| "short_answers": Sequence( |
| { |
| "start_token": Value("int64"), |
| "end_token": Value("int64"), |
| "start_byte": Value("int64"), |
| "end_byte": Value("int64"), |
| "text": Value("string"), |
| } |
| ), |
| "yes_no_answer": ClassLabel(names=["NO", "YES"]), |
| } |
| ), |
| } |
| ) |
|
|
| other = Features( |
| { |
| "id": Value("string"), |
| "document": { |
| "tokens": Sequence({"token": Value("string"), "is_html": Value("bool")}), |
| "title": Value("string"), |
| "url": Value("string"), |
| "html": Value("string"), |
| }, |
| "question": { |
| "text": Value("string"), |
| "tokens": [Value("string")], |
| }, |
| "annotations": { |
| "yes_no_answer": [ClassLabel(names=["NO", "YES"])], |
| "id": [Value("string")], |
| "long_answer": [ |
| { |
| "end_byte": Value("int64"), |
| "start_token": Value("int64"), |
| "end_token": Value("int64"), |
| "start_byte": Value("int64"), |
| } |
| ], |
| "short_answers": [ |
| Sequence( |
| { |
| "text": Value("string"), |
| "start_token": Value("int64"), |
| "end_token": Value("int64"), |
| "start_byte": Value("int64"), |
| "end_byte": Value("int64"), |
| } |
| ) |
| ], |
| }, |
| } |
| ) |
|
|
| expected = Features( |
| { |
| "id": Value("string"), |
| "document": { |
| "tokens": Sequence({"token": Value("string"), "is_html": Value("bool")}), |
| "title": Value("string"), |
| "url": Value("string"), |
| "html": Value("string"), |
| }, |
| "question": { |
| "text": Value("string"), |
| "tokens": Sequence(Value("string")), |
| }, |
| "annotations": Sequence( |
| { |
| "yes_no_answer": ClassLabel(names=["NO", "YES"]), |
| "id": Value("string"), |
| "long_answer": { |
| "end_byte": Value("int64"), |
| "start_token": Value("int64"), |
| "end_token": Value("int64"), |
| "start_byte": Value("int64"), |
| }, |
| "short_answers": Sequence( |
| { |
| "text": Value("string"), |
| "start_token": Value("int64"), |
| "end_token": Value("int64"), |
| "start_byte": Value("int64"), |
| "end_byte": Value("int64"), |
| } |
| ), |
| } |
| ), |
| } |
| ) |
|
|
| reordered_features = features.reorder_fields_as(other) |
| self.assertDictEqual(reordered_features, expected) |
| self.assertEqual(reordered_features.type, other.type) |
| self.assertEqual(reordered_features.type, expected.type) |
| self.assertNotEqual(reordered_features.type, features.type) |
|
|
| def test_flatten(self): |
| features = Features({"foo": {"bar1": Value("int32"), "bar2": {"foobar": Value("string")}}}) |
| _features = features.copy() |
| flattened_features = features.flatten() |
| assert flattened_features == {"foo.bar1": Value("int32"), "foo.bar2.foobar": Value("string")} |
| assert features == _features, "calling flatten shouldn't alter the current features" |
|
|
| def test_flatten_with_sequence(self): |
| features = Features({"foo": Sequence({"bar": {"my_value": Value("int32")}})}) |
| _features = features.copy() |
| flattened_features = features.flatten() |
| assert flattened_features == {"foo.bar": [{"my_value": Value("int32")}]} |
| assert features == _features, "calling flatten shouldn't alter the current features" |
|
|
| def test_features_dicts_are_synced(self): |
| def assert_features_dicts_are_synced(features: Features): |
| assert ( |
| hasattr(features, "_column_requires_decoding") |
| and features.keys() == features._column_requires_decoding.keys() |
| ) |
|
|
| features = Features({"foo": Sequence({"bar": {"my_value": Value("int32")}})}) |
| assert_features_dicts_are_synced(features) |
| features["barfoo"] = Image() |
| assert_features_dicts_are_synced(features) |
| del features["barfoo"] |
| assert_features_dicts_are_synced(features) |
| features.update({"foobar": Value("string")}) |
| assert_features_dicts_are_synced(features) |
| features.pop("foobar") |
| assert_features_dicts_are_synced(features) |
| features.popitem() |
| assert_features_dicts_are_synced(features) |
| features.setdefault("xyz", Value("bool")) |
| assert_features_dicts_are_synced(features) |
| features.clear() |
| assert_features_dicts_are_synced(features) |
|
|
|
|
| def test_classlabel_init(tmp_path_factory): |
| names = ["negative", "positive"] |
| names_file = str(tmp_path_factory.mktemp("features") / "labels.txt") |
| with open(names_file, "w", encoding="utf-8") as f: |
| f.write("\n".join(names)) |
| classlabel = ClassLabel(names=names) |
| assert classlabel.names == names and classlabel.num_classes == len(names) |
| classlabel = ClassLabel(names_file=names_file) |
| assert classlabel.names == names and classlabel.num_classes == len(names) |
| classlabel = ClassLabel(num_classes=len(names), names=names) |
| assert classlabel.names == names and classlabel.num_classes == len(names) |
| classlabel = ClassLabel(num_classes=len(names)) |
| assert classlabel.names == [str(i) for i in range(len(names))] and classlabel.num_classes == len(names) |
| with pytest.raises(ValueError): |
| classlabel = ClassLabel(num_classes=len(names) + 1, names=names) |
| with pytest.raises(ValueError): |
| classlabel = ClassLabel(names=names, names_file=names_file) |
| with pytest.raises(ValueError): |
| classlabel = ClassLabel() |
| with pytest.raises(TypeError): |
| classlabel = ClassLabel(names=np.array(names)) |
|
|
|
|
| def test_classlabel_str2int(): |
| names = ["negative", "positive"] |
| classlabel = ClassLabel(names=names) |
| for label in names: |
| assert classlabel.str2int(label) == names.index(label) |
| with pytest.raises(ValueError): |
| classlabel.str2int("__bad_label_name__") |
| with pytest.raises(ValueError): |
| classlabel.str2int(1) |
| with pytest.raises(ValueError): |
| classlabel.str2int(None) |
|
|
|
|
| def test_classlabel_int2str(): |
| names = ["negative", "positive"] |
| classlabel = ClassLabel(names=names) |
| for i in range(len(names)): |
| assert classlabel.int2str(i) == names[i] |
| with pytest.raises(ValueError): |
| classlabel.int2str(len(names)) |
| with pytest.raises(ValueError): |
| classlabel.int2str(-1) |
| with pytest.raises(ValueError): |
| classlabel.int2str(None) |
|
|
|
|
| def test_classlabel_cast_storage(): |
| names = ["negative", "positive"] |
| classlabel = ClassLabel(names=names) |
| |
| arr = pa.array([0, 1, -1, -100], type=pa.int64()) |
| result = classlabel.cast_storage(arr) |
| assert result.type == pa.int64() |
| assert result.to_pylist() == [0, 1, -1, -100] |
| arr = pa.array([0, 1, -1, -100], type=pa.int32()) |
| result = classlabel.cast_storage(arr) |
| assert result.type == pa.int64() |
| assert result.to_pylist() == [0, 1, -1, -100] |
| arr = pa.array([3]) |
| with pytest.raises(ValueError): |
| classlabel.cast_storage(arr) |
| |
| arr = pa.array(["negative", "positive"]) |
| result = classlabel.cast_storage(arr) |
| assert result.type == pa.int64() |
| assert result.to_pylist() == [0, 1] |
| arr = pa.array(["__label_that_doesnt_exist__"]) |
| with pytest.raises(ValueError): |
| classlabel.cast_storage(arr) |
| |
| arr = pa.array([None]) |
| result = classlabel.cast_storage(arr) |
| assert result.type == pa.int64() |
| assert result.to_pylist() == [None] |
| |
| arr = pa.array([], pa.int64()) |
| result = classlabel.cast_storage(arr) |
| assert result.type == pa.int64() |
| assert result.to_pylist() == [] |
| arr = pa.array([], pa.string()) |
| result = classlabel.cast_storage(arr) |
| assert result.type == pa.int64() |
| assert result.to_pylist() == [] |
|
|
|
|
| @pytest.mark.parametrize("class_label_arg", ["names", "names_file"]) |
| def test_class_label_to_and_from_dict(class_label_arg, tmp_path_factory): |
| names = ["negative", "positive"] |
| names_file = str(tmp_path_factory.mktemp("features") / "labels.txt") |
| with open(names_file, "w", encoding="utf-8") as f: |
| f.write("\n".join(names)) |
| if class_label_arg == "names": |
| class_label = ClassLabel(names=names) |
| elif class_label_arg == "names_file": |
| class_label = ClassLabel(names_file=names_file) |
| generated_class_label = generate_from_dict(asdict(class_label)) |
| assert generated_class_label == class_label |
|
|
|
|
| @pytest.mark.parametrize("inner_type", [Value("int32"), {"subcolumn": Value("int32")}]) |
| def test_encode_nested_example_sequence_with_none(inner_type): |
| schema = Sequence(inner_type) |
| obj = None |
| result = encode_nested_example(schema, obj) |
| assert result is None |
|
|
|
|
| def test_encode_batch_with_example_with_empty_first_elem(): |
| features = Features( |
| { |
| "x": Sequence(Sequence(ClassLabel(names=["a", "b"]))), |
| } |
| ) |
| encoded_batch = features.encode_batch( |
| { |
| "x": [ |
| [["a"], ["b"]], |
| [[], ["b"]], |
| ] |
| } |
| ) |
| assert encoded_batch == {"x": [[[0], [1]], [[], [1]]]} |
|
|
|
|
| @pytest.mark.parametrize( |
| "feature", |
| [ |
| Value("int32"), |
| ClassLabel(num_classes=2), |
| Translation(languages=["en", "fr"]), |
| TranslationVariableLanguages(languages=["en", "fr"]), |
| ], |
| ) |
| def test_dataset_feature_with_none(feature): |
| data = {"col": [None]} |
| features = Features({"col": feature}) |
| dset = Dataset.from_dict(data, features=features) |
| item = dset[0] |
| assert item.keys() == {"col"} |
| assert item["col"] is None |
| batch = dset[:1] |
| assert len(batch) == 1 |
| assert batch.keys() == {"col"} |
| assert isinstance(batch["col"], list) and all(item is None for item in batch["col"]) |
| column = dset["col"] |
| assert len(column) == 1 |
| assert isinstance(column, list) and all(item is None for item in column) |
|
|
| |
|
|
| data = {"col": [[None]]} |
| features = Features({"col": Sequence(feature)}) |
| dset = Dataset.from_dict(data, features=features) |
| item = dset[0] |
| assert item.keys() == {"col"} |
| assert all(i is None for i in item["col"]) |
|
|
| data = {"nested": [{"col": None}]} |
| features = Features({"nested": {"col": feature}}) |
| dset = Dataset.from_dict(data, features=features) |
| item = dset[0] |
| assert item.keys() == {"nested"} |
| assert item["nested"].keys() == {"col"} |
| assert item["nested"]["col"] is None |
|
|
|
|
| def iternumpy(key1, value1, value2): |
| if value1.dtype != value2.dtype: |
| raise AssertionError( |
| f"dtype of '{key1}' key for casted object: {value1.dtype} and expected object: {value2.dtype} not matching" |
| ) |
|
|
|
|
| def dict_diff(d1: dict, d2: dict): |
| np.testing.assert_equal(d1, d2) |
|
|
| for (k1, v1), (k2, v2) in zip(d1.items(), d2.items()): |
| if isinstance(v1, dict): |
| dict_diff(v1, v2) |
| elif isinstance(v1, np.ndarray): |
| iternumpy(k1, v1, v2) |
| elif isinstance(v1, list): |
| for element1, element2 in zip(v1, v2): |
| if isinstance(element1, dict): |
| dict_diff(element1, element2) |
| elif isinstance(element1, np.ndarray): |
| iternumpy(k1, element1, element2) |
|
|
|
|
| class CastToPythonObjectsTest(TestCase): |
| def test_cast_to_python_objects_list(self): |
| obj = {"col_1": [{"vec": [1, 2, 3], "txt": "foo"}] * 3, "col_2": [[1, 2], [3, 4], [5, 6]]} |
| expected_obj = {"col_1": [{"vec": [1, 2, 3], "txt": "foo"}] * 3, "col_2": [[1, 2], [3, 4], [5, 6]]} |
| casted_obj = cast_to_python_objects(obj) |
| self.assertDictEqual(casted_obj, expected_obj) |
|
|
| def test_cast_to_python_objects_tuple(self): |
| obj = {"col_1": [{"vec": (1, 2, 3), "txt": "foo"}] * 3, "col_2": [(1, 2), (3, 4), (5, 6)]} |
| expected_obj = {"col_1": [{"vec": (1, 2, 3), "txt": "foo"}] * 3, "col_2": [(1, 2), (3, 4), (5, 6)]} |
| casted_obj = cast_to_python_objects(obj) |
| self.assertDictEqual(casted_obj, expected_obj) |
|
|
| def test_cast_to_python_or_numpy(self): |
| obj = {"col_1": [{"vec": np.arange(1, 4), "txt": "foo"}] * 3, "col_2": np.arange(1, 7).reshape(3, 2)} |
| expected_obj = { |
| "col_1": [{"vec": np.array([1, 2, 3]), "txt": "foo"}] * 3, |
| "col_2": np.array([[1, 2], [3, 4], [5, 6]]), |
| } |
| casted_obj = cast_to_python_objects(obj) |
| dict_diff(casted_obj, expected_obj) |
|
|
| def test_cast_to_python_objects_series(self): |
| obj = { |
| "col_1": pd.Series([{"vec": [1, 2, 3], "txt": "foo"}] * 3), |
| "col_2": pd.Series([[1, 2], [3, 4], [5, 6]]), |
| } |
| expected_obj = {"col_1": [{"vec": [1, 2, 3], "txt": "foo"}] * 3, "col_2": [[1, 2], [3, 4], [5, 6]]} |
| casted_obj = cast_to_python_objects(obj) |
| self.assertDictEqual(casted_obj, expected_obj) |
|
|
| def test_cast_to_python_objects_dataframe(self): |
| obj = pd.DataFrame({"col_1": [{"vec": [1, 2, 3], "txt": "foo"}] * 3, "col_2": [[1, 2], [3, 4], [5, 6]]}) |
| expected_obj = {"col_1": [{"vec": [1, 2, 3], "txt": "foo"}] * 3, "col_2": [[1, 2], [3, 4], [5, 6]]} |
| casted_obj = cast_to_python_objects(obj) |
| self.assertDictEqual(casted_obj, expected_obj) |
|
|
| def test_cast_to_python_objects_pandas_timestamp(self): |
| obj = pd.Timestamp(2020, 1, 1) |
| expected_obj = obj.to_pydatetime() |
| casted_obj = cast_to_python_objects(obj) |
| self.assertEqual(casted_obj, expected_obj) |
| casted_obj = cast_to_python_objects(pd.Series([obj])) |
| self.assertListEqual(casted_obj, [expected_obj]) |
| casted_obj = cast_to_python_objects(pd.DataFrame({"a": [obj]})) |
| self.assertDictEqual(casted_obj, {"a": [expected_obj]}) |
|
|
| def test_cast_to_python_objects_pandas_timedelta(self): |
| obj = pd.Timedelta(seconds=1) |
| expected_obj = obj.to_pytimedelta() |
| casted_obj = cast_to_python_objects(obj) |
| self.assertEqual(casted_obj, expected_obj) |
| casted_obj = cast_to_python_objects(pd.Series([obj])) |
| self.assertListEqual(casted_obj, [expected_obj]) |
| casted_obj = cast_to_python_objects(pd.DataFrame({"a": [obj]})) |
| self.assertDictEqual(casted_obj, {"a": [expected_obj]}) |
|
|
| @require_torch |
| def test_cast_to_python_objects_torch(self): |
| import torch |
|
|
| obj = { |
| "col_1": [{"vec": torch.tensor(np.arange(1, 4)), "txt": "foo"}] * 3, |
| "col_2": torch.tensor(np.arange(1, 7).reshape(3, 2)), |
| } |
| expected_obj = { |
| "col_1": [{"vec": np.array([1, 2, 3]), "txt": "foo"}] * 3, |
| "col_2": np.array([[1, 2], [3, 4], [5, 6]]), |
| } |
| casted_obj = cast_to_python_objects(obj) |
| dict_diff(casted_obj, expected_obj) |
|
|
| @require_tf |
| def test_cast_to_python_objects_tf(self): |
| import tensorflow as tf |
|
|
| obj = { |
| "col_1": [{"vec": tf.constant(np.arange(1, 4)), "txt": "foo"}] * 3, |
| "col_2": tf.constant(np.arange(1, 7).reshape(3, 2)), |
| } |
| expected_obj = { |
| "col_1": [{"vec": np.array([1, 2, 3]), "txt": "foo"}] * 3, |
| "col_2": np.array([[1, 2], [3, 4], [5, 6]]), |
| } |
| casted_obj = cast_to_python_objects(obj) |
| dict_diff(casted_obj, expected_obj) |
|
|
| @require_jax |
| def test_cast_to_python_objects_jax(self): |
| import jax.numpy as jnp |
|
|
| obj = { |
| "col_1": [{"vec": jnp.array(np.arange(1, 4)), "txt": "foo"}] * 3, |
| "col_2": jnp.array(np.arange(1, 7).reshape(3, 2)), |
| } |
| assert obj["col_2"].dtype == jnp.int32 |
| expected_obj = { |
| "col_1": [{"vec": np.array([1, 2, 3], dtype=np.int32), "txt": "foo"}] * 3, |
| "col_2": np.array([[1, 2], [3, 4], [5, 6]], dtype=np.int32), |
| } |
| casted_obj = cast_to_python_objects(obj) |
| dict_diff(casted_obj, expected_obj) |
|
|
| @patch("datasets.features.features._cast_to_python_objects", side_effect=_cast_to_python_objects) |
| def test_dont_iterate_over_each_element_in_a_list(self, mocked_cast): |
| obj = {"col_1": [[1, 2], [3, 4], [5, 6]]} |
| cast_to_python_objects(obj) |
| self.assertEqual(mocked_cast.call_count, 4) |
|
|
|
|
| SIMPLE_FEATURES = [ |
| Features(), |
| Features({"a": Value("int32")}), |
| Features({"a": Value("int32", id="my feature")}), |
| Features({"a": Value("int32"), "b": Value("float64"), "c": Value("string")}), |
| ] |
|
|
| CUSTOM_FEATURES = [ |
| Features({"label": ClassLabel(names=["negative", "positive"])}), |
| Features({"array": Array2D(dtype="float32", shape=(4, 4))}), |
| Features({"image": Image()}), |
| Features({"audio": Audio()}), |
| Features({"image": Image(decode=False)}), |
| Features({"audio": Audio(decode=False)}), |
| Features({"translation": Translation(["en", "fr"])}), |
| Features({"translation": TranslationVariableLanguages(["en", "fr"])}), |
| ] |
|
|
| NESTED_FEATURES = [ |
| Features({"foo": {}}), |
| Features({"foo": {"bar": Value("int32")}}), |
| Features({"foo": {"bar1": Value("int32"), "bar2": Value("float64")}}), |
| Features({"foo": Sequence(Value("int32"))}), |
| Features({"foo": Sequence({})}), |
| Features({"foo": Sequence({"bar": Value("int32")})}), |
| Features({"foo": [Value("int32")]}), |
| Features({"foo": [{"bar": Value("int32")}]}), |
| ] |
|
|
| NESTED_CUSTOM_FEATURES = [ |
| Features({"foo": {"bar": ClassLabel(names=["negative", "positive"])}}), |
| Features({"foo": Sequence(ClassLabel(names=["negative", "positive"]))}), |
| Features({"foo": Sequence({"bar": ClassLabel(names=["negative", "positive"])})}), |
| Features({"foo": [ClassLabel(names=["negative", "positive"])]}), |
| Features({"foo": [{"bar": ClassLabel(names=["negative", "positive"])}]}), |
| ] |
|
|
|
|
| @pytest.mark.parametrize("features", SIMPLE_FEATURES + CUSTOM_FEATURES + NESTED_FEATURES + NESTED_CUSTOM_FEATURES) |
| def test_features_to_dict(features: Features): |
| features_dict = features.to_dict() |
| assert isinstance(features_dict, dict) |
| reloaded = Features.from_dict(features_dict) |
| assert features == reloaded |
|
|
|
|
| @pytest.mark.parametrize("features", SIMPLE_FEATURES + CUSTOM_FEATURES + NESTED_FEATURES + NESTED_CUSTOM_FEATURES) |
| def test_features_to_yaml_list(features: Features): |
| features_yaml_list = features._to_yaml_list() |
| assert isinstance(features_yaml_list, list) |
| reloaded = Features._from_yaml_list(features_yaml_list) |
| assert features == reloaded |
|
|
|
|
| @pytest.mark.parametrize("features", SIMPLE_FEATURES + CUSTOM_FEATURES + NESTED_FEATURES + NESTED_CUSTOM_FEATURES) |
| def test_features_to_arrow_schema(features: Features): |
| arrow_schema = features.arrow_schema |
| assert isinstance(arrow_schema, pa.Schema) |
| reloaded = Features.from_arrow_schema(arrow_schema) |
| assert features == reloaded |
|
|