| | import os |
| | import sys |
| | from pathlib import Path |
| |
|
| | import pytest |
| |
|
| | from datasets import Dataset, IterableDataset |
| | from datasets.distributed import split_dataset_by_node |
| |
|
| | from .utils import execute_subprocess_async, get_torch_dist_unique_port, require_torch |
| |
|
| |
|
| | def test_split_dataset_by_node_map_style(): |
| | full_ds = Dataset.from_dict({"i": range(17)}) |
| | full_size = len(full_ds) |
| | world_size = 3 |
| | datasets_per_rank = [ |
| | split_dataset_by_node(full_ds, rank=rank, world_size=world_size) for rank in range(world_size) |
| | ] |
| | assert sum(len(ds) for ds in datasets_per_rank) == full_size |
| | assert len({tuple(x.values()) for ds in datasets_per_rank for x in ds}) == full_size |
| |
|
| |
|
| | def test_split_dataset_by_node_iterable(): |
| | def gen(): |
| | return ({"i": i} for i in range(17)) |
| |
|
| | world_size = 3 |
| | full_ds = IterableDataset.from_generator(gen) |
| | full_size = len(list(full_ds)) |
| | datasets_per_rank = [ |
| | split_dataset_by_node(full_ds, rank=rank, world_size=world_size) for rank in range(world_size) |
| | ] |
| | assert sum(len(list(ds)) for ds in datasets_per_rank) == full_size |
| | assert len({tuple(x.values()) for ds in datasets_per_rank for x in ds}) == full_size |
| |
|
| |
|
| | @pytest.mark.parametrize("shards_per_node", [1, 2, 3]) |
| | def test_split_dataset_by_node_iterable_sharded(shards_per_node): |
| | def gen(shards): |
| | for shard in shards: |
| | yield from ({"i": i, "shard": shard} for i in range(17)) |
| |
|
| | world_size = 3 |
| | num_shards = shards_per_node * world_size |
| | gen_kwargs = {"shards": [f"shard_{shard_idx}.txt" for shard_idx in range(num_shards)]} |
| | full_ds = IterableDataset.from_generator(gen, gen_kwargs=gen_kwargs) |
| | full_size = len(list(full_ds)) |
| | assert full_ds.n_shards == world_size * shards_per_node |
| | datasets_per_rank = [ |
| | split_dataset_by_node(full_ds, rank=rank, world_size=world_size) for rank in range(world_size) |
| | ] |
| | assert [ds.n_shards for ds in datasets_per_rank] == [shards_per_node] * world_size |
| | assert sum(len(list(ds)) for ds in datasets_per_rank) == full_size |
| | assert len({tuple(x.values()) for ds in datasets_per_rank for x in ds}) == full_size |
| |
|
| |
|
| | def test_distributed_shuffle_iterable(): |
| | def gen(): |
| | return ({"i": i} for i in range(17)) |
| |
|
| | world_size = 2 |
| | full_ds = IterableDataset.from_generator(gen) |
| | full_size = len(list(full_ds)) |
| |
|
| | ds_rank0 = split_dataset_by_node(full_ds, rank=0, world_size=world_size).shuffle(seed=42) |
| | assert len(list(ds_rank0)) == 1 + full_size // world_size |
| | with pytest.raises(RuntimeError): |
| | split_dataset_by_node(full_ds, rank=0, world_size=world_size).shuffle() |
| |
|
| | ds_rank0 = split_dataset_by_node(full_ds.shuffle(seed=42), rank=0, world_size=world_size) |
| | assert len(list(ds_rank0)) == 1 + full_size // world_size |
| | with pytest.raises(RuntimeError): |
| | split_dataset_by_node(full_ds.shuffle(), rank=0, world_size=world_size) |
| |
|
| |
|
| | @pytest.mark.parametrize("streaming", [False, True]) |
| | @require_torch |
| | @pytest.mark.skipif(os.name == "nt", reason="execute_subprocess_async doesn't support windows") |
| | @pytest.mark.integration |
| | def test_torch_distributed_run(streaming): |
| | nproc_per_node = 2 |
| | master_port = get_torch_dist_unique_port() |
| | test_script = Path(__file__).resolve().parent / "distributed_scripts" / "run_torch_distributed.py" |
| | distributed_args = f""" |
| | -m torch.distributed.run |
| | --nproc_per_node={nproc_per_node} |
| | --master_port={master_port} |
| | {test_script} |
| | """.split() |
| | args = f""" |
| | --streaming={streaming} |
| | """.split() |
| | cmd = [sys.executable] + distributed_args + args |
| | execute_subprocess_async(cmd, env=os.environ.copy()) |
| |
|
| |
|
| | @pytest.mark.parametrize( |
| | "nproc_per_node, num_workers", |
| | [ |
| | (2, 2), |
| | (3, 2), |
| | ], |
| | ) |
| | @require_torch |
| | @pytest.mark.skipif(os.name == "nt", reason="execute_subprocess_async doesn't support windows") |
| | @pytest.mark.integration |
| | def test_torch_distributed_run_streaming_with_num_workers(nproc_per_node, num_workers): |
| | streaming = True |
| | master_port = get_torch_dist_unique_port() |
| | test_script = Path(__file__).resolve().parent / "distributed_scripts" / "run_torch_distributed.py" |
| | distributed_args = f""" |
| | -m torch.distributed.run |
| | --nproc_per_node={nproc_per_node} |
| | --master_port={master_port} |
| | {test_script} |
| | """.split() |
| | args = f""" |
| | --streaming={streaming} |
| | --num_workers={num_workers} |
| | """.split() |
| | cmd = [sys.executable] + distributed_args + args |
| | execute_subprocess_async(cmd, env=os.environ.copy()) |
| |
|