| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | """PASS dataset.""" |
| |
|
| | import os |
| | from datetime import datetime |
| |
|
| | import numpy as np |
| | import pandas as pd |
| |
|
| | import datasets |
| |
|
| |
|
| | _DESCRIPTION = """\ |
| | PASS (Pictures without humAns for Self-Supervision) is a large-scale dataset of 1,440,191 images that does not include any humans |
| | and which can be used for high-quality pretraining while significantly reducing privacy concerns. |
| | The PASS images are sourced from the YFCC-100M dataset. |
| | """ |
| |
|
| | _CITATION = """\ |
| | @Article{asano21pass, |
| | author = "Yuki M. Asano and Christian Rupprecht and Andrew Zisserman and Andrea Vedaldi", |
| | title = "PASS: An ImageNet replacement for self-supervised pretraining without humans", |
| | journal = "NeurIPS Track on Datasets and Benchmarks", |
| | year = "2021" |
| | } |
| | """ |
| |
|
| | _HOMEPAGE = "https://www.robots.ox.ac.uk/~vgg/research/pass/" |
| |
|
| | _LICENSE = "Creative Commons Attribution 4.0 International" |
| |
|
| | _IMAGE_ARCHIVE_DOWNLOAD_URL_TEMPLATE = "https://zenodo.org/record/6615455/files/PASS.{idx}.tar?download=1" |
| |
|
| | _METADATA_DOWNLOAD_URL = "https://zenodo.org/record/6615455/files/pass_metadata.csv?download=1" |
| |
|
| |
|
| | class PASS(datasets.GeneratorBasedBuilder): |
| | """PASS dataset.""" |
| |
|
| | |
| | |
| | VERSION = datasets.Version("2.0.0") |
| |
|
| | def _info(self): |
| | return datasets.DatasetInfo( |
| | description=_DESCRIPTION, |
| | features=datasets.Features( |
| | { |
| | "image": datasets.Image(), |
| | "creator_username": datasets.Value("string"), |
| | "hash": datasets.Value("string"), |
| | "gps_latitude": datasets.Value("float32"), |
| | "gps_longitude": datasets.Value("float32"), |
| | "date_taken": datasets.Value("timestamp[us]"), |
| | } |
| | ), |
| | homepage=_HOMEPAGE, |
| | license=_LICENSE, |
| | citation=_CITATION, |
| | ) |
| |
|
| | def _split_generators(self, dl_manager): |
| | """Returns SplitGenerators.""" |
| | metadata_file, *image_dirs = dl_manager.download( |
| | [_METADATA_DOWNLOAD_URL] + [_IMAGE_ARCHIVE_DOWNLOAD_URL_TEMPLATE.format(idx=i) for i in range(10)] |
| | ) |
| | metadata = pd.read_csv(metadata_file, encoding="utf-8") |
| | metadata = metadata.replace(np.NaN, pd.NA).where(metadata.notnull(), None) |
| | metadata = metadata.set_index("hash") |
| | return [ |
| | datasets.SplitGenerator( |
| | name=datasets.Split.TRAIN, |
| | gen_kwargs={ |
| | "metadata": metadata, |
| | "image_archives": [dl_manager.iter_archive(image_dir) for image_dir in image_dirs], |
| | }, |
| | ) |
| | ] |
| |
|
| | def _generate_examples(self, metadata, image_archives): |
| | """Yields examples.""" |
| | for image_archive in image_archives: |
| | for path, file in image_archive: |
| | img_hash = os.path.basename(path).split(".")[0] |
| | img_meta = metadata.loc[img_hash] |
| | yield img_hash, { |
| | "image": {"path": path, "bytes": file.read()}, |
| | "creator_username": img_meta["unickname"], |
| | "hash": img_hash, |
| | "gps_latitude": img_meta["latitude"], |
| | "gps_longitude": img_meta["longitude"], |
| | "date_taken": datetime.strptime(img_meta["datetaken"], "%Y-%m-%d %H:%M:%S.%f") |
| | if img_meta["datetaken"] is not None |
| | else None, |
| | } |
| |
|