wit / wit.py
nateraw's picture
Update wit.py
f669f4f
raw
history blame
3.78 kB
import json
import datasets
from datasets import Value, Sequence, Features
_CITATION = """\\n@article{srinivasan2021wit,
title={WIT: Wikipedia-based Image Text Dataset for Multimodal Multilingual Machine Learning},
author={Srinivasan, Krishna and Raman, Karthik and Chen, Jiecao and Bendersky, Michael and Najork, Marc},
journal={arXiv preprint arXiv:2103.01913},
year={2021}
}
"""
_DESCRIPTION = """\\nWikipedia-based Image Text (WIT) Dataset is a large multimodal multilingual dataset. WIT is composed of a curated set
of 37.6 million entity rich image-text examples with 11.5 million unique images across 108 Wikipedia languages. Its
size enables WIT to be used as a pretraining dataset for multimodal machine learning models.
"""
_HOMEPAGE = "https://github.com/google-research-datasets/wit"
_URL = "https://storage.googleapis.com/huggingface-nlp/datasets/wit/"
_URLS = {
# TODO - This should be in range(400). Haven't mirrored all the files yet.
'train': [_URL + f"part-{'%05d' % i}-48a6f07e-bb86-4735-aac7-883349f41a28-c000.json.gz" for i in range(10)]
}
class Wit(datasets.GeneratorBasedBuilder):
"""WIT: Wikipedia-based Image Text Dataset for Multimodal Multilingual Machine Learning"""
def _info(self):
return datasets.DatasetInfo(
description=_DESCRIPTION,
features=Features({
'b64_bytes': Value('string'),
'embedding': Sequence(Value('float64')),
'image_url': Value('string'),
'metadata_url': Value('string'),
'wit_features': Sequence({
"language": Value('string'),
"page_url": Value('string'),
"image_url": Value('string'),
"attribution_passes_lang_id": Value("string"),
"caption_alt_text_description": Value('string'),
"caption_attribution_description": Value('string'),
"caption_reference_description": Value('string'),
"caption_title_and_reference_description": Value('string'),
"context_page_description": Value('string'),
"context_section_description": Value('string'),
"hierarchical_section_title": Value('string'),
"is_main_image": Value('string'),
"mime_type": Value('string'),
"original_height": Value('string'),
"original_width": Value('string'),
"page_changed_recently": Value('string'),
"page_title": Value('string'),
"section_title": Value('string'),
})
}),
homepage=_HOMEPAGE,
citation=_CITATION,
)
def _split_generators(self, dl_manager):
"""Returns SplitGenerators."""
urls_to_download = _URLS
downloaded_files = dl_manager.download_and_extract(urls_to_download)
return [
datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"filepaths": downloaded_files["train"]}),
]
def _generate_examples(self, filepaths):
"""Yields examples."""
wit_feature_names = self.info.features['wit_features'].feature.keys()
for filepath in filepaths:
with open(filepath, "rb") as f:
for i, line in enumerate(f):
line = line.strip()
row_data = json.loads(line, encoding='utf-8')
for feature in row_data['wit_features']:
for fname in wit_feature_names:
if fname not in feature:
feature[fname] = None
yield str(i), row_data