seed
stringlengths
53
1.87k
seed_api
stringlengths
22
51
index
int64
0
259
from torchdata.datapipes.iter import IterDataPipe, Mapper, Filter path=path, image=EncodedImage.from_file(buffer), ) def _filter_split(self, data: Tuple[str, Any], *, split: str) -> bool: return pathlib.Path(data[0]).parent.parent.name == split def _make_datapipe( ...
torchdata.datapipes.iter.Mapper
0
from torchdata.datapipes.iter import FileOpener, HttpReader, IterableWrapper @_wrap_split_argument(('train', 'valid', 'test')) def WikiText2(root: str, split: Union[Tuple[str], str]): if not is_module_available("torchdata"): raise ModuleNotFoundError("Package `torchdata` not found. Please install following...
torchdata.datapipes.iter.FileLister
1
from torchdata.datapipes.utils import StreamWrapper def _get_response_from_http(url: str, *, timeout: Optional[float]) -> Tuple[str, StreamWrapper]: try: with requests.Session() as session: if timeout is None: r = session.get(url, stream=True) else: ...
torchdata.datapipes.iter.ParagraphAggregator
2
from torchdata.datapipes.iter import FileOpener, HttpReader, IterableWrapper filepath_fn=lambda x: os.path.join(root, os.path.basename(x)), hash_dict={os.path.join(root, os.path.basename(URL[split])): MD5[split]}, hash_type="md5", ) cache_dp = HttpReader(cache_dp).end_caching(mode="wb",...
torchdata.datapipes.iter.FileLister
3
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader for _, gz_stream in gz_decompress_dp: ...
torchdata.datapipes.iter.Demultiplexer
4
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader def test_rar_archive_loader(self): self._...
torchdata.datapipes.iter.IterableWrapper
5
from torchdata.dataloader2 import DataLoader2 expected_batch = 0 for batch in iter(data_loader): self.assertEqual(batch, expected_batch) expected_batch += 1 def test_dataloader2_shutdown(self) -> None: test_data_pipe = IterableWrapper(range(3)) data_loader ...
torchdata.datapipes.iter.Mapper
6
from torchdata.dataloader2.graph import find_dps, remove_dp, replace_dp, traverse class TestGraph(expecttest.TestCase): def _get_datapipes(self) -> Tuple[IterDataPipe, IterDataPipe, IterDataPipe]: src_dp = IterableWrapper(range(20)) m1 = src_dp.map(_x_and_x_plus_5) ub = m1.unbatch() ...
torchdata.datapipes.iter.HttpReader
7
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader # Functional Test: work with .xz files ...
torchdata.datapipes.utils.StreamWrapper
8
from torchdata.datapipes.iter import FileOpener, HttpReader, IterableWrapper DATASET_NAME = "AG_NEWS" @_add_docstring_header(num_lines=NUM_LINES, num_classes=4) @_create_dataset_directory(dataset_name=DATASET_NAME) @_wrap_split_argument(("train", "test")) def AG_NEWS(root: str, split: Union[Tuple[str], str]): if...
torchdata.datapipes.iter.FileOpener
9
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader # Nested RAR in TAR datapipe1 = Iter...
torchdata.datapipes.iter.HttpReader
10
from torchdata.datapipes.iter import S3FileLister, S3FileLoader from torchrec.datasets.utils import ( LoadFiles, ReadLinesFromCSV) from torch.utils.data import IterDataPipe from torchrec.datasets.criteo import _default_row_mapper s3_prefixes = ['s3://criteo-dataset/day_0'] dp_s3_urls = S3FileLister(s3_prefixe...
torchdata.datapipes.iter.Mapper
11
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader # Functional Test: work without file type as...
torchdata.datapipes.iter.HttpReader
12
from torchdata.datapipes.iter import FileOpener, HttpReader, IterableWrapper """ if not is_module_available("torchdata"): raise ModuleNotFoundError( "Package `torchdata` not found. Please install following instructions at `https://github.com/pytorch/data`" ) url_dp = IterableWr...
torchdata.datapipes.iter.Decompressor
13
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader def fill_hash_dict(): for path in...
torchdata.datapipes.iter.FileOpener
14
from torchdata.datapipes.iter import IterDataPipe, LineReader, IterKeyZipper, Mapper, Filter, Demultiplexer, TarArchiveLoader, Enumerator config = self.info.make_config(split="val") resources = self.resources(config) devkit_dp = resources[1].load(root) meta_dp = Filter(devkit_dp, path_...
torchdata.datapipes.iter.FileOpener
15
from torchdata.datapipes.iter import IterDataPipe, LineReader, IterKeyZipper, Mapper, Filter, Demultiplexer, TarArchiveLoader, Enumerator dp = IterKeyZipper( label_dp, images_dp, key_fn=getitem(0), ref_key_fn=path_accessor(self._val_test_imag...
torchdata.datapipes.iter.S3FileLoader
16
from torchdata.dataloader2.graph import find_dps, remove_dp, replace_dp, traverse def initialize(self, datapipe: IterDataPipe) -> IterDataPipe: graph = traverse(datapipe, only_datapipe=True) dps = find_dps(graph, Mapper)
torchdata.datapipes.iter.FileOpener
17
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader # Functional Test: Compression Type throws e...
torchdata.datapipes.iter.IterableWrapper
18
from torchdata.datapipes.iter import IterDataPipe, Mapper, Shuffler, Filter, IterKeyZipper, Demultiplexer, LineReader, CSVParser splits_dp = LineReader(splits_dp, decode=True, return_path=False) splits_dp = Shuffler(splits_dp, buffer_size=INFINITE_BUFFER_SIZE)
torchdata.datapipes.iter.IterableWrapper
19
from torchdata.datapipes.iter import IterDataPipe, LineReader, IterKeyZipper, Mapper, Filter, Demultiplexer, TarArchiveLoader, Enumerator dp = Mapper(dp, self._prepare_train_data if config.split == "train" else self._prepare_test_data) else: # config.split == "val": images_dp, devkit_d...
torchdata.datapipes.iter.Mapper
20
from torchdata.datapipes.iter import IterDataPipe, Mapper, Filter, Demultiplexer, IterKeyZipper, LineReader def _datapipe(self, resource_dps: List[IterDataPipe]) -> IterDataPipe[Dict[str, Any]]: archive_dp = resource_dps[0] split_dp, images_dp, anns_dp = Demultiplexer( archive_dp, ...
torchdata.datapipes.iter.Shuffler
21
from torchdata.datapipes.iter import BucketBatcher, Cycler, Header, IndexAdder, InMemoryCacheHolder, IterableWrapper, IterDataPipe, IterKeyZipper, LineReader, MapKeyZipper, ParagraphAggregator, Rows2Columnar, SampleMultiplexer def test_header_iterdatapipe(self) -> None: # Functional Test: ensure the limit...
torchdata.dataloader2.graph.replace_dp
22
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader xz_decompress_dp = Decompressor(xz_load_dp) ...
torchdata.datapipes.iter.IterableWrapper
23
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader self._write_test_xz_files() datapipe1...
torchdata.datapipes.functional_datapipe
24
from torchdata.datapipes.iter import FileLister, HttpReader, IterDataPipe # Yes, we had to scan files twice. Alternativelly it is possible to use # `fork` DataPipe, but it will require buffer equal to the size of all # full file names # TODO(125): Make sure that `fork` complains when bu...
torchdata.datapipes.iter.Enumerator
25
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader # Functional Test: work with .gz files ...
torchdata.datapipes.iter.Mapper
26
from torchdata.datapipes.iter import FileLister, HttpReader, IterDataPipe category = cat_to_dp[self.parse_category_fn(data)] yield (data, category) def MyImageFolder(root=IMAGES_ROOT, transform=None): if not USE_FORK_DATAPIPE: # Yes, we had to scan files twice. Alternative...
torchdata.datapipes.iter.Mapper
27
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader myzip.write(self.temp_files[2], arcname=o...
torchdata.datapipes.iter.FileLister
28
from torchdata.datapipes.iter import IterDataPipe, FileLister, FileOpener, Mapper, Shuffler, Filter ) -> Tuple[IterDataPipe, List[str]]: root = pathlib.Path(root).expanduser().resolve() categories = sorted(entry.name for entry in os.scandir(root) if entry.is_dir()) masks: Union[List[str], str] = [f"*.{ext}...
torchdata.datapipes.iter.FileOpener
29
from torchdata.datapipes.utils import StreamWrapper self.source_datapipe: IterDataPipe[str] = source_datapipe self.pathmgr = g_pathmgr self.mode: str = mode def __iter__(self) -> Iterator[Tuple[str, StreamWrapper]]: for file_uri in self.source_datapipe: with self.pathm...
torchdata.datapipes.iter.LineReader
30
from torchdata.datapipes.iter import FileOpener, IterableWrapper :returns: DataPipe that yields rows from QQP dataset (label (int), question1 (str), question2 (str)) :rtype: (int, str, str) """ if not is_module_available("torchdata"): raise ModuleNotFoundError( "Package `torchdata` ...
torchdata.datapipes.iter.FileLister
31
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader with open(p) as f: self.a...
torchdata.datapipes.iter.Header
32
from torchdata.datapipes.iter import BucketBatcher, Cycler, Header, IndexAdder, InMemoryCacheHolder, IterableWrapper, IterDataPipe, IterKeyZipper, LineReader, MapKeyZipper, ParagraphAggregator, Rows2Columnar, SampleMultiplexer enum_dp = source_dp.enumerate(starting_index=10) self.assertEqual([(10, "a")...
torchdata.datapipes.iter.FileLister
33
from torchdata.datapipes.iter import FileOpener, GDriveReader, IterableWrapper cache_compressed_dp = GDriveReader(cache_compressed_dp).end_caching(mode="wb", same_filepath_fn=True) cache_decompressed_dp = cache_compressed_dp.on_disk_cache( filepath_fn=lambda x: os.path.join(root, _EXTRACTED_FILES[spli...
torchdata.datapipes.iter.IoPathFileOpener
34
from torchdata.datapipes.iter import BucketBatcher, Cycler, Header, IndexAdder, InMemoryCacheHolder, IterableWrapper, IterDataPipe, IterKeyZipper, LineReader, MapKeyZipper, ParagraphAggregator, Rows2Columnar, SampleMultiplexer ("file2", "Line2,1\r\n"), ("file2", "Line2,2\r\n"), ("fi...
torchdata.datapipes.iter.IterableWrapper
35
from torchdata.datapipes.iter import HttpReader, IterableWrapper # Reset Test: http_reader_dp has been read, but we reset when calling check_hash() check_cache_dp = http_reader_dp.check_hash({file_url: expected_MD5_hash}, "md5", rewind=False) it = iter(check_cache_dp) path, stream = nex...
torchdata.datapipes.iter.FileLister
36
from torchdata.datapipes.iter import FileOpener, IterableWrapper, HttpReader raise ModuleNotFoundError("Package `torchdata` not found. Please install following instructions at `https://github.com/pytorch/data`") url_dp = IterableWrapper([URL])
torchdata.datapipes.iter.FileLister
37
from torchdata.datapipes.iter import IterDataPipe, Mapper, Demultiplexer, Filter, IterKeyZipper, LineReader def _generate_categories(self, root: pathlib.Path) -> Tuple[str, ...]: dp = self.resources(self.default_config)[0].load(pathlib.Path(root) / self.name) dp = Filter(dp, path_comparator("name"...
torchdata.datapipes.iter.FileOpener
38
from torchdata.datapipes.iter import FileOpener, HttpReader, IterableWrapper hash_dict={os.path.join(root, split + ".csv"): MD5[split]}, hash_type="md5" ) cache_dp = HttpReader(cache_dp) cache_dp = cache_dp.end_caching(mode="wb", same_filepath_fn=True) cache_dp = FileOpener(cache_dp, mo...
torchdata.datapipes.iter.FileOpener
39
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader tar.add(self.temp_files[1]) t...
torchdata.datapipes.iter.CSVDictParser
40
from torchdata.datapipes import functional_datapipe
torchdata.datapipes.iter.LineReader
41
from torchdata.dataloader2.graph import find_dps, remove_dp, replace_dp, traverse exp_g1 = { dp: { m2: {c1: {dm: {ub: {new_dp1: {m1: {src_dp: {}}}}}}}, c2: {dm: {ub: {new_dp1: {m1: {src_dp: {}}}}}}, } } self.assertEqual(graph, exp_g1) ...
torchdata.datapipes.iter.FileLister
42
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader self.assertEqual(f.read(), gz_stream....
torchdata.datapipes.iter.Mapper
43
from torchdata.datapipes.iter import IterableWrapper ), (iterdp.SampleMultiplexer, {IterableWrapper([0] * 10): 0.5, IterableWrapper([1] * 10): 0.5}, (), {}),
torchdata.datapipes.iter.FileOpener
44
from torchdata.datapipes.iter import BucketBatcher, Cycler, Header, IndexAdder, InMemoryCacheHolder, IterableWrapper, IterDataPipe, IterKeyZipper, LineReader, MapKeyZipper, ParagraphAggregator, Rows2Columnar, SampleMultiplexer # Reset Test source_dp = IterableWrapper([{i: i} for i in "abcdefg"]) ...
torchdata.dataloader2.graph.remove_dp
45
from torchdata.datapipes.iter import IterDataPipe, Mapper, Filter, CSVDictParser, Zipper, Demultiplexer resource_dps[0], 2, self._classify_train_archive, drop_none=True, buffer_size=INFINITE_BUFFER_SIZE ) else: images_dp, ann_dp = resource_dps images_dp = Fil...
torchdata.datapipes.iter.FileOpener
46
from torchdata.datapipes.iter import FileOpener, IterableWrapper if not is_module_available("torchdata"): raise ModuleNotFoundError( "Package `torchdata` not found. Please install following instructions at `https://github.com/pytorch/data`" ) url_dp = IterableWrapper([URL]) cac...
torchdata.datapipes.iter.Mapper
47
from torchdata.datapipes.iter import S3FileLister, S3FileLoader s3_urls = list(filter(is_final_day, s3_urls)) rank = ( dist.get_rank() if stage == "val" else dist.get_rank() + dist.get_world_size() ) world_size = dist.get_world_size() * 2 s3_urls...
torchdata.datapipes.iter.Mapper
48
from torchdata.datapipes.iter import BucketBatcher, Cycler, Header, IndexAdder, InMemoryCacheHolder, IterableWrapper, IterDataPipe, IterKeyZipper, LineReader, MapKeyZipper, ParagraphAggregator, Rows2Columnar, SampleMultiplexer self.assertEqual(expected_res, list(result_dp)) # Functional Test: ensure t...
torchdata.dataloader2.graph.find_dps
49
from torchdata.datapipes.iter import FileOpener, IterableWrapper - train: 87599 - dev: 10570 Args: root: Directory where the datasets are saved. Default: os.path.expanduser('~/.torchtext/cache') split: split or splits to be returned. Can be a string or tuple of strings. Default: (`...
torchdata.datapipes.iter.TFRecordLoader
50
from torchdata.dataloader2 import DataLoader2, MultiProcessingReadingService, ReadingServiceInterface mp.set_start_method("fork") rs = MultiProcessingReadingService(2, persistent_workers=True) dl = DataLoader2(dp, reading_service=rs) d1 = list(dl) d2 = list(dl) self.as...
torchdata.datapipes.iter.Mapper
51
from torchdata.datapipes.iter import FileOpener, HttpReader, IterableWrapper cache_decompressed_dp = cache_compressed_dp.on_disk_cache(filepath_fn=lambda x: os.path.join(root, _EXTRACTED_FILES[split])) # Extract zip and filter the appropriate split file cache_decompressed_dp = FileOpener(cache_decompressed...
torchdata.datapipes.iter.IoPathFileOpener
52
from torchdata.datapipes.iter import FileOpener, HttpReader, IterableWrapper @_wrap_split_argument(("train", "test")) def AG_NEWS(root: str, split: Union[Tuple[str], str]): if not is_module_available("torchdata"): raise ModuleNotFoundError("Package `torchdata` not found. Please install following instructio...
torchdata.datapipes.iter.Decompressor
53
from torchdata.datapipes.iter import IterDataPipe, Mapper, Filter, IterKeyZipper, Demultiplexer, JsonParser, UnBatcher archive_dp = resource_dps[0] images_dp, scenes_dp = Demultiplexer(
torchdata.datapipes.functional_datapipe
54
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader with self.assertRaisesRegex(TypeError, "insta...
torchdata.datapipes.iter.Mapper
55
from torchdata.datapipes.iter import IterDataPipe, LineReader, IterKeyZipper, Mapper, Filter, Demultiplexer, TarArchiveLoader, Enumerator dp = hint_shuffling(dp) dp = hint_sharding(dp) dp = Mapper(dp, self._prepare_train_data if self._split == "train" else self._prepare_test_data) ...
torchdata.datapipes.iter.S3FileLister
56
from torchdata.datapipes.iter import IterDataPipe, Mapper, UnBatcher self, resource_dps: List[IterDataPipe], *, config: DatasetConfig, decoder: Optional[Callable[[io.IOBase], torch.Tensor]], ) -> IterDataPipe[Dict[str, Any]]: dp = resource_dps[0] dp = Mapper(...
torchdata.datapipes.iter.IoPathFileLister
57
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader def _write_test_xz_files(self): for path...
torchdata.datapipes.iter.Demultiplexer
58
from torchdata.datapipes.iter import IterDataPipe, Mapper, Filter, IterKeyZipper, Demultiplexer, JsonParser, UnBatcher scenes_dp = JsonParser(scenes_dp) scenes_dp = Mapper(scenes_dp, getitem(1, "scenes")) scenes_dp = UnBatcher(scenes_dp)
torchdata.datapipes.iter.HttpReader
59
from torchdata.datapipes.iter import IterDataPipe, Filter, Mapper def _make_datapipe( self, resource_dps: List[IterDataPipe], *, config: DatasetConfig, decoder: Optional[Callable[[io.IOBase], torch.Tensor]], ) -> IterDataPipe[Dict[str, Any]]: dp = resource_dps[0]...
torchdata.datapipes.iter.FileOpener
60
from torchdata.datapipes.iter import BucketBatcher, Cycler, Header, IndexAdder, InMemoryCacheHolder, IterableWrapper, IterDataPipe, IterKeyZipper, LineReader, MapKeyZipper, ParagraphAggregator, Rows2Columnar, SampleMultiplexer # Functional Test: aggregate lines correctly with different joiner para_agg...
torchdata.dataloader2.graph.remove_dp
61
from torchdata.datapipes.iter import IterableWrapper (iterdp.BatchMapper, IterableWrapper([(0, 0), (0, 0), (0, 0), (0, 0)]), (_fake_batch_fn, 2, 1), {}), (iterdp.BucketBatcher, IterableWrapper([0, 0, 0, 0, 0, 0, 0]), (5,), {}),
torchdata.datapipes.iter.HashChecker
62
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader hash_func = hashlib.sha256() ...
torchdata.datapipes.iter.Shuffler
63
from torchdata.datapipes.iter import BucketBatcher, Cycler, Header, IndexAdder, InMemoryCacheHolder, IterableWrapper, IterDataPipe, IterKeyZipper, LineReader, MapKeyZipper, ParagraphAggregator, Rows2Columnar, SampleMultiplexer self.assertEqual(30, len(header_dp)) self.assertEqual(len(wa), 1) ...
torchdata.datapipes.iter.XzFileLoader
64
from torchdata.dataloader2.graph import find_dps, remove_dp, replace_dp, traverse m1, ub, dm, c1, c2, m2, dp, ) = self._get_datapipes() graph = remove_dp(graph, m1) exp_g1 = {dp: {m2: {c1: {dm: {ub: {src_dp: {}...
torchdata.dataloader2.graph.remove_dp
65
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader lister_dp = FileLister(self.temp_dir.name, "*...
torchdata.datapipes.iter.Cycler
66
from torchdata.datapipes.iter import BucketBatcher, Cycler, Header, IndexAdder, InMemoryCacheHolder, IterableWrapper, IterDataPipe, IterKeyZipper, LineReader, MapKeyZipper, ParagraphAggregator, Rows2Columnar, SampleMultiplexer # Functional Test: raises error when the elements of source_dp is not of type Dict ...
torchdata.datapipes.iter.FileOpener
67
from torchdata.datapipes.iter import IterDataPipe, FileLister, FileOpener, Mapper, Shuffler, Filter def from_image_folder( root: Union[str, pathlib.Path], *, decoder: Optional[Callable[[io.IOBase], torch.Tensor]] = pil, valid_extensions: Collection[str] = ("jpg", "jpeg", "png", "ppm", "bmp", "pgm", "...
torchdata.datapipes.iter.Filter
68
from torchdata.datapipes import functional_datapipe def __iter__(self) -> Iterator[str]: if self.pathmgr.isfile(self.root): yield self.root else: for file_name in self.pathmgr.ls(self.root): yield os.path.join(self.root, file_name) @functional_datapipe("loa...
torchdata.datapipes.iter.Mapper
69
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader tar_gz_load_dp = FileOpener(tar_gz_file_dp, m...
torchdata.datapipes.iter.Mapper
70
from torchdata.mpii import MPII_Joint_Names, MpiiData def run_gui(preds, subset, model=None): mpii_data = MpiiData('/datasets/mpii')
torchdata.datapipes.iter.ZipArchiveReader
71
from torchdata.datapipes.iter import IterDataPipe, LineReader, IterKeyZipper, Mapper, Filter, Demultiplexer, TarArchiveLoader, Enumerator dict(zip(("label", "wnid"), label_data if label_data else (None, None))), path=path, image=EncodedImage.from_file(buffer), ) def _ma...
torchdata.datapipes.iter.CSVDictParser
72
from torchdata.datapipes.iter import FileLister, FileOpener, FSSpecFileLister, FSSpecFileOpener, FSSpecSaver, IterableWrapper, TFRecordLoader } @torch.no_grad() def test_tfrecord_loader_example_iterdatapipe(self): filename = f"{self.temp_dir}/example.tfrecord" datapipe1 = IterableW...
torchdata.datapipes.iter.RarArchiveLoader
73
from torchdata.datapipes.iter import IterDataPipe, LineReader, IterKeyZipper, Mapper, Filter, Demultiplexer, TarArchiveLoader, Enumerator key_fn=getitem(0), ref_key_fn=self._val_test_image_key, buffer_size=INFINITE_BUFFER_SIZE, ) dp = Mapper(dp, s...
torchdata.datapipes.iter.HttpReader
74
from torchdata.datapipes.iter import IterDataPipe, Mapper, Shuffler, Filter, IterKeyZipper, Demultiplexer, LineReader, CSVParser ) -> IterDataPipe[Dict[str, Any]]: archive_dp = resource_dps[0] splits_dp, joint_categories_dp, images_dp = Demultiplexer( archive_dp, 3, self._classify_arch...
torchdata.datapipes.iter.IterableWrapper
75
from torchdata.datapipes.iter import BucketBatcher, Cycler, Header, IndexAdder, InMemoryCacheHolder, IterableWrapper, IterDataPipe, IterKeyZipper, LineReader, MapKeyZipper, ParagraphAggregator, Rows2Columnar, SampleMultiplexer self.assertEqual(2, len(result_dp)) def test_sample_multiplexer_iterdatapipe(se...
torchdata.dataloader2.graph.replace_dp
76
from torchdata.datapipes.iter import IterDataPipe, FileLister, FileOpener, Mapper, Shuffler, Filter dp = FileLister(str(root), recursive=recursive, masks=masks) dp: IterDataPipe = Filter(dp, functools.partial(_is_not_top_level_file, root=root)) dp = hint_sharding(dp) dp = Shuffler(dp, buffer_size=INFIN...
torchdata.datapipes.iter.IterableWrapper
77
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader with self.assertRaisesRegex(TypeError, "insta...
torchdata.datapipes.iter.IterableWrapper
78
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader tar.add(self.temp_files[0]) t...
torchdata.datapipes.iter.RarArchiveLoader
79
from torchdata.datapipes.iter import IterableWrapper ), (), {}, ), (iterdp.Cycler, None, (2,), {}), (iterdp.DataFrameMaker, IterableWrapper([(i,) for i in range(3)]), (), {"dtype": DTYPE}), (iterdp.Decompressor, None, (), {...
torchdata.datapipes.iter.HashChecker
80
from torchdata.datapipes.iter import HttpReader @_create_dataset_directory(dataset_name=DATASET_NAME) @_wrap_split_argument(("train", "test")) def AG_NEWS(root, split): """Demonstrating streaming use case This might be useful when we do not want to cache or download the data. The limitation is that we do n...
torchdata.datapipes.iter.FileOpener
81
from torchdata.datapipes.iter import Mapper train_dp = train_img_dp.zip(train_height_dp).zip(train_label_dp) val_dp = val_img_dp.zip(val_height_dp).zip(val_label_dp) test_dp = test_img_dp.zip(test_height_dp).zip(test_label_dp) '''tfs = transforms.Compose(transforms.Resize((256,...
torchdata.datapipes.iter.GDriveReader
82
from torchdata.datapipes.iter import FileLister, FileOpener, FSSpecFileLister, FSSpecFileOpener, FSSpecSaver, IterableWrapper, TFRecordLoader # Functional Test: raises error if missing spec feature with self.assertRaises(RuntimeError): tfrecord_parser = datapipe2.load_from_tfrecord( ...
torchdata.datapipes.iter.IterableWrapper
83
from torchdata.datapipes.iter import FileOpener, HttpReader, IterableWrapper _EXTRACTED_FILES = { 'train': os.path.join('wikitext-2', 'wiki.train.tokens'), 'test': os.path.join('wikitext-2', 'wiki.test.tokens'), 'valid': os.path.join('wikitext-2', 'wiki.valid.tokens'), } @_add_docstring_header(num_lines=...
torchdata.datapipes.iter.JsonParser
84
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader ("2.json", {"__complex__": True, "real": ...
torchdata.datapipes.iter.IterableWrapper
85
from torchdata.datapipes import functional_datapipe from torchdata.datapipes import functional_datapipe from torchdata.datapipes.iter import IterDataPipe try: # TODO: Create dependency on TorchArrow? import pyarrow.parquet as parquet import torcharrow except ImportError: torcharrow = None parquet = N...
torchdata.datapipes.iter.IterableWrapper
86
from torchdata.datapipes.iter import BucketBatcher, Cycler, Header, IndexAdder, InMemoryCacheHolder, IterableWrapper, IterDataPipe, IterKeyZipper, LineReader, MapKeyZipper, ParagraphAggregator, Rows2Columnar, SampleMultiplexer # Functional Test: using sort_key, without in_batch_shuffle batch_dp = sourc...
torchdata.dataloader2.graph.traverse
87
from torchdata.dataloader2.graph import find_dps, remove_dp, replace_dp, traverse self.assertEqual(traverse(dp, only_datapipe=True), exp_g2) def test_remove_dps(self) -> None: # pyre-fixme[23]: Unable to unpack 3 values, 2 were expected. graph, ( src_dp, m1, ...
torchdata.datapipes.iter.FileLister
88
from torchdata.datapipes.iter import IterDataPipe, Mapper, Filter, Demultiplexer, IterKeyZipper, LineReader ) return Mapper(dp, self._prepare_sample)
torchdata.datapipes.iter.Demultiplexer
89
from torchdata.datapipes.iter import S3FileLister, S3FileLoader ) -> DataLoader: s3_urls = S3FileLister([args.s3_criteo_prefix])
torchdata.datapipes.iter.Mapper
90
from torchdata.datapipes.iter import BucketBatcher, Cycler, Header, IndexAdder, InMemoryCacheHolder, IterableWrapper, IterDataPipe, IterKeyZipper, LineReader, MapKeyZipper, ParagraphAggregator, Rows2Columnar, SampleMultiplexer # Functional Test: Cache DP should just return the data without changing the values ...
torchdata.datapipes.iter.S3FileLister
91
from torchdata.datapipes.iter import IoPathFileLister, IoPathFileOpener, IterDataPipe, ShardingFilter, Shuffler def _make_sharded_datapipe(root: str, dataset_size: int) -> IterDataPipe[Dict[str, Any]]: dp = IoPathFileLister(root=root) dp = SharderDataPipe(dp) dp = dp.shuffle(buffer_size=INFINITE_BUFFER_S...
torchdata.datapipes.iter.IterKeyZipper
92
from torchdata.datapipes.iter import FileLister, FileOpener, FSSpecFileLister, FSSpecFileOpener, FSSpecSaver, IterableWrapper, TFRecordLoader @torch.no_grad() def test_tfrecord_loader_example_iterdatapipe(self): filename = f"{self.temp_dir}/example.tfrecord" datapipe1 = IterableWrapper([filena...
torchdata.datapipes.iter.CSVParser
93
from torchdata.datapipes.iter import IterDataPipe, LineReader, IterKeyZipper, Mapper, Filter, Demultiplexer, TarArchiveLoader, Enumerator dict(zip(("label", "wnid"), label_data if label_data else (None, None))), path=path, image=EncodedImage.from_file(buffer), ) def _da...
torchdata.maps.Flatten
94
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader with self.assertRaisesRegex(TypeError, "has n...
torchdata.datapipes.iter.ShardingFilter
95
from torchdata.datapipes.iter import IterDataPipe, Mapper, CSVDictParser return dict( image=Image(torch.tensor([int(idx) for idx in data["pixels"].split()], dtype=torch.uint8).reshape(48, 48)), label=Label(int(label_id), categories=self.categories) if label_id is not None else None, ...
torchdata.datapipes.iter.FileOpener
96
from torchdata.datapipes.iter import IterableWrapper ] ), (), {}, ), ( iterdp.LineReader, IterableWrapper( [("file1", StringIO("Line1\nLine2")), ("file2", StringIO("Line2,1\r\...
torchdata.datapipes.iter.FileOpener
97
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader list(rar_loader_dp_2) # Nested R...
torchdata.datapipes.iter.Mapper
98
from torchdata.datapipes.iter import BucketBatcher, Cycler, Header, IndexAdder, InMemoryCacheHolder, IterableWrapper, IterDataPipe, IterKeyZipper, LineReader, MapKeyZipper, ParagraphAggregator, Rows2Columnar, SampleMultiplexer # Reset Test: result_dp = Rows2Columnar(source_dp, column_names_list) ...
torchdata.datapipes.iter.FileLister
99