File size: 1,428 Bytes
3ce159a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
492c23f
 
 
 
 
 
 
3b2b0ce
 
 
 
3ce159a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3b2b0ce
492c23f
3ce159a
3b2b0ce
492c23f
 
 
 
 
3b2b0ce
3ce159a
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
import os

import datasets as ds
import pytest


@pytest.fixture
def dataset_path() -> str:
    return "COCOA.py"


@pytest.fixture
def data_dir() -> str:
    is_ci = bool(os.environ.get("CI", False))
    if is_ci:
        raise NotImplementedError
    else:
        return "annotations.tar.gz"


@pytest.mark.parametrize(
    argnames="use_auth_token",
    argvalues=(
        True,
        False,
    ),
)
@pytest.mark.parametrize(
    argnames="decode_rle,",
    argvalues=(False, True),
)
@pytest.mark.parametrize(
    argnames=(
        "dataset_name",
        "expected_num_train",
        "expected_num_validation",
        "expected_num_test",
    ),
    argvalues=(
        ("COCO", 2500, 1323, 1250),
        ("BSDS", 200, 100, 200),
    ),
)
def test_load_dataset(
    dataset_path: str,
    dataset_name: str,
    data_dir: str,
    expected_num_train: int,
    expected_num_validation: int,
    expected_num_test: int,
    decode_rle: bool,
    use_auth_token: bool,
):
    dataset = ds.load_dataset(
        path=dataset_path,
        name=dataset_name,
        data_dir=data_dir if not use_auth_token else None,
        decode_rle=decode_rle,
        token=use_auth_token,
    )

    assert dataset["train"].num_rows == expected_num_train  # type: ignore
    assert dataset["validation"].num_rows == expected_num_validation  # type: ignore
    assert dataset["test"].num_rows == expected_num_test  # type: ignore