seed stringlengths 53 1.87k | seed_api stringlengths 22 51 | index int64 0 259 |
|---|---|---|
from torchdata.datapipes.iter import IterDataPipe, Mapper, Filter
path=path,
image=EncodedImage.from_file(buffer),
)
def _filter_split(self, data: Tuple[str, Any], *, split: str) -> bool:
return pathlib.Path(data[0]).parent.parent.name == split
def _make_datapipe(
... | torchdata.datapipes.iter.Mapper | 0 |
from torchdata.datapipes.iter import FileOpener, HttpReader, IterableWrapper
@_wrap_split_argument(('train', 'valid', 'test'))
def WikiText2(root: str, split: Union[Tuple[str], str]):
if not is_module_available("torchdata"):
raise ModuleNotFoundError("Package `torchdata` not found. Please install following... | torchdata.datapipes.iter.FileLister | 1 |
from torchdata.datapipes.utils import StreamWrapper
def _get_response_from_http(url: str, *, timeout: Optional[float]) -> Tuple[str, StreamWrapper]:
try:
with requests.Session() as session:
if timeout is None:
r = session.get(url, stream=True)
else:
... | torchdata.datapipes.iter.ParagraphAggregator | 2 |
from torchdata.datapipes.iter import FileOpener, HttpReader, IterableWrapper
filepath_fn=lambda x: os.path.join(root, os.path.basename(x)),
hash_dict={os.path.join(root, os.path.basename(URL[split])): MD5[split]},
hash_type="md5",
)
cache_dp = HttpReader(cache_dp).end_caching(mode="wb",... | torchdata.datapipes.iter.FileLister | 3 |
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader
for _, gz_stream in gz_decompress_dp:
... | torchdata.datapipes.iter.Demultiplexer | 4 |
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader
def test_rar_archive_loader(self):
self._... | torchdata.datapipes.iter.IterableWrapper | 5 |
from torchdata.dataloader2 import DataLoader2
expected_batch = 0
for batch in iter(data_loader):
self.assertEqual(batch, expected_batch)
expected_batch += 1
def test_dataloader2_shutdown(self) -> None:
test_data_pipe = IterableWrapper(range(3))
data_loader ... | torchdata.datapipes.iter.Mapper | 6 |
from torchdata.dataloader2.graph import find_dps, remove_dp, replace_dp, traverse
class TestGraph(expecttest.TestCase):
def _get_datapipes(self) -> Tuple[IterDataPipe, IterDataPipe, IterDataPipe]:
src_dp = IterableWrapper(range(20))
m1 = src_dp.map(_x_and_x_plus_5)
ub = m1.unbatch()
... | torchdata.datapipes.iter.HttpReader | 7 |
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader
# Functional Test: work with .xz files
... | torchdata.datapipes.utils.StreamWrapper | 8 |
from torchdata.datapipes.iter import FileOpener, HttpReader, IterableWrapper
DATASET_NAME = "AG_NEWS"
@_add_docstring_header(num_lines=NUM_LINES, num_classes=4)
@_create_dataset_directory(dataset_name=DATASET_NAME)
@_wrap_split_argument(("train", "test"))
def AG_NEWS(root: str, split: Union[Tuple[str], str]):
if... | torchdata.datapipes.iter.FileOpener | 9 |
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader
# Nested RAR in TAR
datapipe1 = Iter... | torchdata.datapipes.iter.HttpReader | 10 |
from torchdata.datapipes.iter import S3FileLister, S3FileLoader
from torchrec.datasets.utils import (
LoadFiles,
ReadLinesFromCSV)
from torch.utils.data import IterDataPipe
from torchrec.datasets.criteo import _default_row_mapper
s3_prefixes = ['s3://criteo-dataset/day_0']
dp_s3_urls = S3FileLister(s3_prefixe... | torchdata.datapipes.iter.Mapper | 11 |
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader
# Functional Test: work without file type as... | torchdata.datapipes.iter.HttpReader | 12 |
from torchdata.datapipes.iter import FileOpener, HttpReader, IterableWrapper
"""
if not is_module_available("torchdata"):
raise ModuleNotFoundError(
"Package `torchdata` not found. Please install following instructions at `https://github.com/pytorch/data`"
)
url_dp = IterableWr... | torchdata.datapipes.iter.Decompressor | 13 |
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader
def fill_hash_dict():
for path in... | torchdata.datapipes.iter.FileOpener | 14 |
from torchdata.datapipes.iter import IterDataPipe, LineReader, IterKeyZipper, Mapper, Filter, Demultiplexer, TarArchiveLoader, Enumerator
config = self.info.make_config(split="val")
resources = self.resources(config)
devkit_dp = resources[1].load(root)
meta_dp = Filter(devkit_dp, path_... | torchdata.datapipes.iter.FileOpener | 15 |
from torchdata.datapipes.iter import IterDataPipe, LineReader, IterKeyZipper, Mapper, Filter, Demultiplexer, TarArchiveLoader, Enumerator
dp = IterKeyZipper(
label_dp,
images_dp,
key_fn=getitem(0),
ref_key_fn=path_accessor(self._val_test_imag... | torchdata.datapipes.iter.S3FileLoader | 16 |
from torchdata.dataloader2.graph import find_dps, remove_dp, replace_dp, traverse
def initialize(self, datapipe: IterDataPipe) -> IterDataPipe:
graph = traverse(datapipe, only_datapipe=True)
dps = find_dps(graph, Mapper)
| torchdata.datapipes.iter.FileOpener | 17 |
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader
# Functional Test: Compression Type throws e... | torchdata.datapipes.iter.IterableWrapper | 18 |
from torchdata.datapipes.iter import IterDataPipe, Mapper, Shuffler, Filter, IterKeyZipper, Demultiplexer, LineReader, CSVParser
splits_dp = LineReader(splits_dp, decode=True, return_path=False)
splits_dp = Shuffler(splits_dp, buffer_size=INFINITE_BUFFER_SIZE)
| torchdata.datapipes.iter.IterableWrapper | 19 |
from torchdata.datapipes.iter import IterDataPipe, LineReader, IterKeyZipper, Mapper, Filter, Demultiplexer, TarArchiveLoader, Enumerator
dp = Mapper(dp, self._prepare_train_data if config.split == "train" else self._prepare_test_data)
else: # config.split == "val":
images_dp, devkit_d... | torchdata.datapipes.iter.Mapper | 20 |
from torchdata.datapipes.iter import IterDataPipe, Mapper, Filter, Demultiplexer, IterKeyZipper, LineReader
def _datapipe(self, resource_dps: List[IterDataPipe]) -> IterDataPipe[Dict[str, Any]]:
archive_dp = resource_dps[0]
split_dp, images_dp, anns_dp = Demultiplexer(
archive_dp,
... | torchdata.datapipes.iter.Shuffler | 21 |
from torchdata.datapipes.iter import BucketBatcher, Cycler, Header, IndexAdder, InMemoryCacheHolder, IterableWrapper, IterDataPipe, IterKeyZipper, LineReader, MapKeyZipper, ParagraphAggregator, Rows2Columnar, SampleMultiplexer
def test_header_iterdatapipe(self) -> None:
# Functional Test: ensure the limit... | torchdata.dataloader2.graph.replace_dp | 22 |
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader
xz_decompress_dp = Decompressor(xz_load_dp)
... | torchdata.datapipes.iter.IterableWrapper | 23 |
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader
self._write_test_xz_files()
datapipe1... | torchdata.datapipes.functional_datapipe | 24 |
from torchdata.datapipes.iter import FileLister, HttpReader, IterDataPipe
# Yes, we had to scan files twice. Alternativelly it is possible to use
# `fork` DataPipe, but it will require buffer equal to the size of all
# full file names
# TODO(125): Make sure that `fork` complains when bu... | torchdata.datapipes.iter.Enumerator | 25 |
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader
# Functional Test: work with .gz files
... | torchdata.datapipes.iter.Mapper | 26 |
from torchdata.datapipes.iter import FileLister, HttpReader, IterDataPipe
category = cat_to_dp[self.parse_category_fn(data)]
yield (data, category)
def MyImageFolder(root=IMAGES_ROOT, transform=None):
if not USE_FORK_DATAPIPE:
# Yes, we had to scan files twice. Alternative... | torchdata.datapipes.iter.Mapper | 27 |
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader
myzip.write(self.temp_files[2], arcname=o... | torchdata.datapipes.iter.FileLister | 28 |
from torchdata.datapipes.iter import IterDataPipe, FileLister, FileOpener, Mapper, Shuffler, Filter
) -> Tuple[IterDataPipe, List[str]]:
root = pathlib.Path(root).expanduser().resolve()
categories = sorted(entry.name for entry in os.scandir(root) if entry.is_dir())
masks: Union[List[str], str] = [f"*.{ext}... | torchdata.datapipes.iter.FileOpener | 29 |
from torchdata.datapipes.utils import StreamWrapper
self.source_datapipe: IterDataPipe[str] = source_datapipe
self.pathmgr = g_pathmgr
self.mode: str = mode
def __iter__(self) -> Iterator[Tuple[str, StreamWrapper]]:
for file_uri in self.source_datapipe:
with self.pathm... | torchdata.datapipes.iter.LineReader | 30 |
from torchdata.datapipes.iter import FileOpener, IterableWrapper
:returns: DataPipe that yields rows from QQP dataset (label (int), question1 (str), question2 (str))
:rtype: (int, str, str)
"""
if not is_module_available("torchdata"):
raise ModuleNotFoundError(
"Package `torchdata` ... | torchdata.datapipes.iter.FileLister | 31 |
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader
with open(p) as f:
self.a... | torchdata.datapipes.iter.Header | 32 |
from torchdata.datapipes.iter import BucketBatcher, Cycler, Header, IndexAdder, InMemoryCacheHolder, IterableWrapper, IterDataPipe, IterKeyZipper, LineReader, MapKeyZipper, ParagraphAggregator, Rows2Columnar, SampleMultiplexer
enum_dp = source_dp.enumerate(starting_index=10)
self.assertEqual([(10, "a")... | torchdata.datapipes.iter.FileLister | 33 |
from torchdata.datapipes.iter import FileOpener, GDriveReader, IterableWrapper
cache_compressed_dp = GDriveReader(cache_compressed_dp).end_caching(mode="wb", same_filepath_fn=True)
cache_decompressed_dp = cache_compressed_dp.on_disk_cache(
filepath_fn=lambda x: os.path.join(root, _EXTRACTED_FILES[spli... | torchdata.datapipes.iter.IoPathFileOpener | 34 |
from torchdata.datapipes.iter import BucketBatcher, Cycler, Header, IndexAdder, InMemoryCacheHolder, IterableWrapper, IterDataPipe, IterKeyZipper, LineReader, MapKeyZipper, ParagraphAggregator, Rows2Columnar, SampleMultiplexer
("file2", "Line2,1\r\n"),
("file2", "Line2,2\r\n"),
("fi... | torchdata.datapipes.iter.IterableWrapper | 35 |
from torchdata.datapipes.iter import HttpReader, IterableWrapper
# Reset Test: http_reader_dp has been read, but we reset when calling check_hash()
check_cache_dp = http_reader_dp.check_hash({file_url: expected_MD5_hash}, "md5", rewind=False)
it = iter(check_cache_dp)
path, stream = nex... | torchdata.datapipes.iter.FileLister | 36 |
from torchdata.datapipes.iter import FileOpener, IterableWrapper, HttpReader
raise ModuleNotFoundError("Package `torchdata` not found. Please install following instructions at `https://github.com/pytorch/data`")
url_dp = IterableWrapper([URL])
| torchdata.datapipes.iter.FileLister | 37 |
from torchdata.datapipes.iter import IterDataPipe, Mapper, Demultiplexer, Filter, IterKeyZipper, LineReader
def _generate_categories(self, root: pathlib.Path) -> Tuple[str, ...]:
dp = self.resources(self.default_config)[0].load(pathlib.Path(root) / self.name)
dp = Filter(dp, path_comparator("name"... | torchdata.datapipes.iter.FileOpener | 38 |
from torchdata.datapipes.iter import FileOpener, HttpReader, IterableWrapper
hash_dict={os.path.join(root, split + ".csv"): MD5[split]},
hash_type="md5"
)
cache_dp = HttpReader(cache_dp)
cache_dp = cache_dp.end_caching(mode="wb", same_filepath_fn=True)
cache_dp = FileOpener(cache_dp, mo... | torchdata.datapipes.iter.FileOpener | 39 |
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader
tar.add(self.temp_files[1])
t... | torchdata.datapipes.iter.CSVDictParser | 40 |
from torchdata.datapipes import functional_datapipe
| torchdata.datapipes.iter.LineReader | 41 |
from torchdata.dataloader2.graph import find_dps, remove_dp, replace_dp, traverse
exp_g1 = {
dp: {
m2: {c1: {dm: {ub: {new_dp1: {m1: {src_dp: {}}}}}}},
c2: {dm: {ub: {new_dp1: {m1: {src_dp: {}}}}}},
}
}
self.assertEqual(graph, exp_g1)
... | torchdata.datapipes.iter.FileLister | 42 |
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader
self.assertEqual(f.read(), gz_stream.... | torchdata.datapipes.iter.Mapper | 43 |
from torchdata.datapipes.iter import IterableWrapper
),
(iterdp.SampleMultiplexer, {IterableWrapper([0] * 10): 0.5, IterableWrapper([1] * 10): 0.5}, (), {}),
| torchdata.datapipes.iter.FileOpener | 44 |
from torchdata.datapipes.iter import BucketBatcher, Cycler, Header, IndexAdder, InMemoryCacheHolder, IterableWrapper, IterDataPipe, IterKeyZipper, LineReader, MapKeyZipper, ParagraphAggregator, Rows2Columnar, SampleMultiplexer
# Reset Test
source_dp = IterableWrapper([{i: i} for i in "abcdefg"])
... | torchdata.dataloader2.graph.remove_dp | 45 |
from torchdata.datapipes.iter import IterDataPipe, Mapper, Filter, CSVDictParser, Zipper, Demultiplexer
resource_dps[0], 2, self._classify_train_archive, drop_none=True, buffer_size=INFINITE_BUFFER_SIZE
)
else:
images_dp, ann_dp = resource_dps
images_dp = Fil... | torchdata.datapipes.iter.FileOpener | 46 |
from torchdata.datapipes.iter import FileOpener, IterableWrapper
if not is_module_available("torchdata"):
raise ModuleNotFoundError(
"Package `torchdata` not found. Please install following instructions at `https://github.com/pytorch/data`"
)
url_dp = IterableWrapper([URL])
cac... | torchdata.datapipes.iter.Mapper | 47 |
from torchdata.datapipes.iter import S3FileLister, S3FileLoader
s3_urls = list(filter(is_final_day, s3_urls))
rank = (
dist.get_rank()
if stage == "val"
else dist.get_rank() + dist.get_world_size()
)
world_size = dist.get_world_size() * 2
s3_urls... | torchdata.datapipes.iter.Mapper | 48 |
from torchdata.datapipes.iter import BucketBatcher, Cycler, Header, IndexAdder, InMemoryCacheHolder, IterableWrapper, IterDataPipe, IterKeyZipper, LineReader, MapKeyZipper, ParagraphAggregator, Rows2Columnar, SampleMultiplexer
self.assertEqual(expected_res, list(result_dp))
# Functional Test: ensure t... | torchdata.dataloader2.graph.find_dps | 49 |
from torchdata.datapipes.iter import FileOpener, IterableWrapper
- train: 87599
- dev: 10570
Args:
root: Directory where the datasets are saved. Default: os.path.expanduser('~/.torchtext/cache')
split: split or splits to be returned. Can be a string or tuple of strings. Default: (`... | torchdata.datapipes.iter.TFRecordLoader | 50 |
from torchdata.dataloader2 import DataLoader2, MultiProcessingReadingService, ReadingServiceInterface
mp.set_start_method("fork")
rs = MultiProcessingReadingService(2, persistent_workers=True)
dl = DataLoader2(dp, reading_service=rs)
d1 = list(dl)
d2 = list(dl)
self.as... | torchdata.datapipes.iter.Mapper | 51 |
from torchdata.datapipes.iter import FileOpener, HttpReader, IterableWrapper
cache_decompressed_dp = cache_compressed_dp.on_disk_cache(filepath_fn=lambda x: os.path.join(root, _EXTRACTED_FILES[split]))
# Extract zip and filter the appropriate split file
cache_decompressed_dp = FileOpener(cache_decompressed... | torchdata.datapipes.iter.IoPathFileOpener | 52 |
from torchdata.datapipes.iter import FileOpener, HttpReader, IterableWrapper
@_wrap_split_argument(("train", "test"))
def AG_NEWS(root: str, split: Union[Tuple[str], str]):
if not is_module_available("torchdata"):
raise ModuleNotFoundError("Package `torchdata` not found. Please install following instructio... | torchdata.datapipes.iter.Decompressor | 53 |
from torchdata.datapipes.iter import IterDataPipe, Mapper, Filter, IterKeyZipper, Demultiplexer, JsonParser, UnBatcher
archive_dp = resource_dps[0]
images_dp, scenes_dp = Demultiplexer(
| torchdata.datapipes.functional_datapipe | 54 |
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader
with self.assertRaisesRegex(TypeError, "insta... | torchdata.datapipes.iter.Mapper | 55 |
from torchdata.datapipes.iter import IterDataPipe, LineReader, IterKeyZipper, Mapper, Filter, Demultiplexer, TarArchiveLoader, Enumerator
dp = hint_shuffling(dp)
dp = hint_sharding(dp)
dp = Mapper(dp, self._prepare_train_data if self._split == "train" else self._prepare_test_data)
... | torchdata.datapipes.iter.S3FileLister | 56 |
from torchdata.datapipes.iter import IterDataPipe, Mapper, UnBatcher
self,
resource_dps: List[IterDataPipe],
*,
config: DatasetConfig,
decoder: Optional[Callable[[io.IOBase], torch.Tensor]],
) -> IterDataPipe[Dict[str, Any]]:
dp = resource_dps[0]
dp = Mapper(... | torchdata.datapipes.iter.IoPathFileLister | 57 |
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader
def _write_test_xz_files(self):
for path... | torchdata.datapipes.iter.Demultiplexer | 58 |
from torchdata.datapipes.iter import IterDataPipe, Mapper, Filter, IterKeyZipper, Demultiplexer, JsonParser, UnBatcher
scenes_dp = JsonParser(scenes_dp)
scenes_dp = Mapper(scenes_dp, getitem(1, "scenes"))
scenes_dp = UnBatcher(scenes_dp)
| torchdata.datapipes.iter.HttpReader | 59 |
from torchdata.datapipes.iter import IterDataPipe, Filter, Mapper
def _make_datapipe(
self,
resource_dps: List[IterDataPipe],
*,
config: DatasetConfig,
decoder: Optional[Callable[[io.IOBase], torch.Tensor]],
) -> IterDataPipe[Dict[str, Any]]:
dp = resource_dps[0]... | torchdata.datapipes.iter.FileOpener | 60 |
from torchdata.datapipes.iter import BucketBatcher, Cycler, Header, IndexAdder, InMemoryCacheHolder, IterableWrapper, IterDataPipe, IterKeyZipper, LineReader, MapKeyZipper, ParagraphAggregator, Rows2Columnar, SampleMultiplexer
# Functional Test: aggregate lines correctly with different joiner
para_agg... | torchdata.dataloader2.graph.remove_dp | 61 |
from torchdata.datapipes.iter import IterableWrapper
(iterdp.BatchMapper, IterableWrapper([(0, 0), (0, 0), (0, 0), (0, 0)]), (_fake_batch_fn, 2, 1), {}),
(iterdp.BucketBatcher, IterableWrapper([0, 0, 0, 0, 0, 0, 0]), (5,), {}),
| torchdata.datapipes.iter.HashChecker | 62 |
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader
hash_func = hashlib.sha256()
... | torchdata.datapipes.iter.Shuffler | 63 |
from torchdata.datapipes.iter import BucketBatcher, Cycler, Header, IndexAdder, InMemoryCacheHolder, IterableWrapper, IterDataPipe, IterKeyZipper, LineReader, MapKeyZipper, ParagraphAggregator, Rows2Columnar, SampleMultiplexer
self.assertEqual(30, len(header_dp))
self.assertEqual(len(wa), 1)
... | torchdata.datapipes.iter.XzFileLoader | 64 |
from torchdata.dataloader2.graph import find_dps, remove_dp, replace_dp, traverse
m1,
ub,
dm,
c1,
c2,
m2,
dp,
) = self._get_datapipes()
graph = remove_dp(graph, m1)
exp_g1 = {dp: {m2: {c1: {dm: {ub: {src_dp: {}... | torchdata.dataloader2.graph.remove_dp | 65 |
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader
lister_dp = FileLister(self.temp_dir.name, "*... | torchdata.datapipes.iter.Cycler | 66 |
from torchdata.datapipes.iter import BucketBatcher, Cycler, Header, IndexAdder, InMemoryCacheHolder, IterableWrapper, IterDataPipe, IterKeyZipper, LineReader, MapKeyZipper, ParagraphAggregator, Rows2Columnar, SampleMultiplexer
# Functional Test: raises error when the elements of source_dp is not of type Dict
... | torchdata.datapipes.iter.FileOpener | 67 |
from torchdata.datapipes.iter import IterDataPipe, FileLister, FileOpener, Mapper, Shuffler, Filter
def from_image_folder(
root: Union[str, pathlib.Path],
*,
decoder: Optional[Callable[[io.IOBase], torch.Tensor]] = pil,
valid_extensions: Collection[str] = ("jpg", "jpeg", "png", "ppm", "bmp", "pgm", "... | torchdata.datapipes.iter.Filter | 68 |
from torchdata.datapipes import functional_datapipe
def __iter__(self) -> Iterator[str]:
if self.pathmgr.isfile(self.root):
yield self.root
else:
for file_name in self.pathmgr.ls(self.root):
yield os.path.join(self.root, file_name)
@functional_datapipe("loa... | torchdata.datapipes.iter.Mapper | 69 |
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader
tar_gz_load_dp = FileOpener(tar_gz_file_dp, m... | torchdata.datapipes.iter.Mapper | 70 |
from torchdata.mpii import MPII_Joint_Names, MpiiData
def run_gui(preds, subset, model=None):
mpii_data = MpiiData('/datasets/mpii')
| torchdata.datapipes.iter.ZipArchiveReader | 71 |
from torchdata.datapipes.iter import IterDataPipe, LineReader, IterKeyZipper, Mapper, Filter, Demultiplexer, TarArchiveLoader, Enumerator
dict(zip(("label", "wnid"), label_data if label_data else (None, None))),
path=path,
image=EncodedImage.from_file(buffer),
)
def _ma... | torchdata.datapipes.iter.CSVDictParser | 72 |
from torchdata.datapipes.iter import FileLister, FileOpener, FSSpecFileLister, FSSpecFileOpener, FSSpecSaver, IterableWrapper, TFRecordLoader
}
@torch.no_grad()
def test_tfrecord_loader_example_iterdatapipe(self):
filename = f"{self.temp_dir}/example.tfrecord"
datapipe1 = IterableW... | torchdata.datapipes.iter.RarArchiveLoader | 73 |
from torchdata.datapipes.iter import IterDataPipe, LineReader, IterKeyZipper, Mapper, Filter, Demultiplexer, TarArchiveLoader, Enumerator
key_fn=getitem(0),
ref_key_fn=self._val_test_image_key,
buffer_size=INFINITE_BUFFER_SIZE,
)
dp = Mapper(dp, s... | torchdata.datapipes.iter.HttpReader | 74 |
from torchdata.datapipes.iter import IterDataPipe, Mapper, Shuffler, Filter, IterKeyZipper, Demultiplexer, LineReader, CSVParser
) -> IterDataPipe[Dict[str, Any]]:
archive_dp = resource_dps[0]
splits_dp, joint_categories_dp, images_dp = Demultiplexer(
archive_dp, 3, self._classify_arch... | torchdata.datapipes.iter.IterableWrapper | 75 |
from torchdata.datapipes.iter import BucketBatcher, Cycler, Header, IndexAdder, InMemoryCacheHolder, IterableWrapper, IterDataPipe, IterKeyZipper, LineReader, MapKeyZipper, ParagraphAggregator, Rows2Columnar, SampleMultiplexer
self.assertEqual(2, len(result_dp))
def test_sample_multiplexer_iterdatapipe(se... | torchdata.dataloader2.graph.replace_dp | 76 |
from torchdata.datapipes.iter import IterDataPipe, FileLister, FileOpener, Mapper, Shuffler, Filter
dp = FileLister(str(root), recursive=recursive, masks=masks)
dp: IterDataPipe = Filter(dp, functools.partial(_is_not_top_level_file, root=root))
dp = hint_sharding(dp)
dp = Shuffler(dp, buffer_size=INFIN... | torchdata.datapipes.iter.IterableWrapper | 77 |
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader
with self.assertRaisesRegex(TypeError, "insta... | torchdata.datapipes.iter.IterableWrapper | 78 |
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader
tar.add(self.temp_files[0])
t... | torchdata.datapipes.iter.RarArchiveLoader | 79 |
from torchdata.datapipes.iter import IterableWrapper
),
(),
{},
),
(iterdp.Cycler, None, (2,), {}),
(iterdp.DataFrameMaker, IterableWrapper([(i,) for i in range(3)]), (), {"dtype": DTYPE}),
(iterdp.Decompressor, None, (), {... | torchdata.datapipes.iter.HashChecker | 80 |
from torchdata.datapipes.iter import HttpReader
@_create_dataset_directory(dataset_name=DATASET_NAME)
@_wrap_split_argument(("train", "test"))
def AG_NEWS(root, split):
"""Demonstrating streaming use case
This might be useful when we do not want to cache or download the data.
The limitation is that we do n... | torchdata.datapipes.iter.FileOpener | 81 |
from torchdata.datapipes.iter import Mapper
train_dp = train_img_dp.zip(train_height_dp).zip(train_label_dp)
val_dp = val_img_dp.zip(val_height_dp).zip(val_label_dp)
test_dp = test_img_dp.zip(test_height_dp).zip(test_label_dp)
'''tfs = transforms.Compose(transforms.Resize((256,... | torchdata.datapipes.iter.GDriveReader | 82 |
from torchdata.datapipes.iter import FileLister, FileOpener, FSSpecFileLister, FSSpecFileOpener, FSSpecSaver, IterableWrapper, TFRecordLoader
# Functional Test: raises error if missing spec feature
with self.assertRaises(RuntimeError):
tfrecord_parser = datapipe2.load_from_tfrecord(
... | torchdata.datapipes.iter.IterableWrapper | 83 |
from torchdata.datapipes.iter import FileOpener, HttpReader, IterableWrapper
_EXTRACTED_FILES = {
'train': os.path.join('wikitext-2', 'wiki.train.tokens'),
'test': os.path.join('wikitext-2', 'wiki.test.tokens'),
'valid': os.path.join('wikitext-2', 'wiki.valid.tokens'),
}
@_add_docstring_header(num_lines=... | torchdata.datapipes.iter.JsonParser | 84 |
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader
("2.json", {"__complex__": True, "real": ... | torchdata.datapipes.iter.IterableWrapper | 85 |
from torchdata.datapipes import functional_datapipe
from torchdata.datapipes import functional_datapipe
from torchdata.datapipes.iter import IterDataPipe
try: # TODO: Create dependency on TorchArrow?
import pyarrow.parquet as parquet
import torcharrow
except ImportError:
torcharrow = None
parquet = N... | torchdata.datapipes.iter.IterableWrapper | 86 |
from torchdata.datapipes.iter import BucketBatcher, Cycler, Header, IndexAdder, InMemoryCacheHolder, IterableWrapper, IterDataPipe, IterKeyZipper, LineReader, MapKeyZipper, ParagraphAggregator, Rows2Columnar, SampleMultiplexer
# Functional Test: using sort_key, without in_batch_shuffle
batch_dp = sourc... | torchdata.dataloader2.graph.traverse | 87 |
from torchdata.dataloader2.graph import find_dps, remove_dp, replace_dp, traverse
self.assertEqual(traverse(dp, only_datapipe=True), exp_g2)
def test_remove_dps(self) -> None:
# pyre-fixme[23]: Unable to unpack 3 values, 2 were expected.
graph, (
src_dp,
m1,
... | torchdata.datapipes.iter.FileLister | 88 |
from torchdata.datapipes.iter import IterDataPipe, Mapper, Filter, Demultiplexer, IterKeyZipper, LineReader
)
return Mapper(dp, self._prepare_sample)
| torchdata.datapipes.iter.Demultiplexer | 89 |
from torchdata.datapipes.iter import S3FileLister, S3FileLoader
) -> DataLoader:
s3_urls = S3FileLister([args.s3_criteo_prefix])
| torchdata.datapipes.iter.Mapper | 90 |
from torchdata.datapipes.iter import BucketBatcher, Cycler, Header, IndexAdder, InMemoryCacheHolder, IterableWrapper, IterDataPipe, IterKeyZipper, LineReader, MapKeyZipper, ParagraphAggregator, Rows2Columnar, SampleMultiplexer
# Functional Test: Cache DP should just return the data without changing the values
... | torchdata.datapipes.iter.S3FileLister | 91 |
from torchdata.datapipes.iter import IoPathFileLister, IoPathFileOpener, IterDataPipe, ShardingFilter, Shuffler
def _make_sharded_datapipe(root: str, dataset_size: int) -> IterDataPipe[Dict[str, Any]]:
dp = IoPathFileLister(root=root)
dp = SharderDataPipe(dp)
dp = dp.shuffle(buffer_size=INFINITE_BUFFER_S... | torchdata.datapipes.iter.IterKeyZipper | 92 |
from torchdata.datapipes.iter import FileLister, FileOpener, FSSpecFileLister, FSSpecFileOpener, FSSpecSaver, IterableWrapper, TFRecordLoader
@torch.no_grad()
def test_tfrecord_loader_example_iterdatapipe(self):
filename = f"{self.temp_dir}/example.tfrecord"
datapipe1 = IterableWrapper([filena... | torchdata.datapipes.iter.CSVParser | 93 |
from torchdata.datapipes.iter import IterDataPipe, LineReader, IterKeyZipper, Mapper, Filter, Demultiplexer, TarArchiveLoader, Enumerator
dict(zip(("label", "wnid"), label_data if label_data else (None, None))),
path=path,
image=EncodedImage.from_file(buffer),
)
def _da... | torchdata.maps.Flatten | 94 |
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader
with self.assertRaisesRegex(TypeError, "has n... | torchdata.datapipes.iter.ShardingFilter | 95 |
from torchdata.datapipes.iter import IterDataPipe, Mapper, CSVDictParser
return dict(
image=Image(torch.tensor([int(idx) for idx in data["pixels"].split()], dtype=torch.uint8).reshape(48, 48)),
label=Label(int(label_id), categories=self.categories) if label_id is not None else None,
... | torchdata.datapipes.iter.FileOpener | 96 |
from torchdata.datapipes.iter import IterableWrapper
]
),
(),
{},
),
(
iterdp.LineReader,
IterableWrapper(
[("file1", StringIO("Line1\nLine2")), ("file2", StringIO("Line2,1\r\... | torchdata.datapipes.iter.FileOpener | 97 |
from torchdata.datapipes.iter import CSVDictParser, CSVParser, Decompressor, FileLister, FileOpener, HashChecker, IoPathFileLister, IoPathFileOpener, IoPathSaver, IterableWrapper, JsonParser, RarArchiveLoader, Saver, TarArchiveLoader, XzFileLoader, ZipArchiveLoader
list(rar_loader_dp_2)
# Nested R... | torchdata.datapipes.iter.Mapper | 98 |
from torchdata.datapipes.iter import BucketBatcher, Cycler, Header, IndexAdder, InMemoryCacheHolder, IterableWrapper, IterDataPipe, IterKeyZipper, LineReader, MapKeyZipper, ParagraphAggregator, Rows2Columnar, SampleMultiplexer
# Reset Test:
result_dp = Rows2Columnar(source_dp, column_names_list)
... | torchdata.datapipes.iter.FileLister | 99 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.