|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
"""Openpi V2: A Dataset for tracking state changes in prcedural text by using an unrestricted library""" |
|
|
|
|
|
import json |
|
|
import os |
|
|
import sys |
|
|
import textwrap |
|
|
|
|
|
import numpy as np |
|
|
import datasets |
|
|
|
|
|
logger = datasets.logging.get_logger(__name__) |
|
|
|
|
|
_OPENPI_V2_CITATION = """\ |
|
|
@inproceedings{ |
|
|
title={{OPENPI V2}: } |
|
|
author={} |
|
|
note={} |
|
|
year={2022} |
|
|
} |
|
|
""" |
|
|
|
|
|
_OPENPI_V2_DESCRIPTION = """\ |
|
|
TEMPORARY DESCRIPTION |
|
|
""" |
|
|
|
|
|
_LICENSE = "CC BY 4.0" |
|
|
_VERSION = "1.0.0" |
|
|
_HOMEPAGE = "https://allenai.org/data/openpi" |
|
|
_URL = "https://huggingface.co/datasets/abhinavk/openpi_v2/resolve/main/data/" |
|
|
_URLS = {"train": _URL + "train-data.json", |
|
|
"dev": _URL + "dev-data.json", |
|
|
"test": _URL + "test-data.json"} |
|
|
|
|
|
|
|
|
class OpenpiConfig(datasets.BuilderConfig): |
|
|
"""BuilderConfig for Openpi V2.""" |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
features, |
|
|
data_url, |
|
|
citation, |
|
|
url, |
|
|
process_label = lambda x: x, |
|
|
**kwargs, |
|
|
): |
|
|
|
|
|
super(OpenpiConfig, self).__init__(version = datasets.Version(_VERSION), **kwargs) |
|
|
self.features = features |
|
|
self.data_url = data_url |
|
|
self.citation = citation |
|
|
self.url = url |
|
|
self.process_label = process_label |
|
|
|
|
|
|
|
|
class OpenpiV2(datasets.GeneratorBasedBuilder): |
|
|
|
|
|
BUILDER_CONFIGS = [ |
|
|
OpenpiConfig( |
|
|
name = "openpi_text", |
|
|
description = textwrap.dedent( |
|
|
"""\ |
|
|
""" |
|
|
), |
|
|
features = datasets.Features({ |
|
|
"goal": datasets.Value("string"), |
|
|
"steps": [datasets.Value("string")], |
|
|
"topics": datasets.Value("string"), |
|
|
"image_urls": [datasets.Value("string")], |
|
|
"states": [{ |
|
|
"answers_openpiv1_metadata": { |
|
|
"entity": datasets.Value("string"), |
|
|
"attribute": datasets.Value("string"), |
|
|
"answers": [datasets.Value("string")], |
|
|
"modality": [datasets.Value("string")] |
|
|
}, |
|
|
"entity": datasets.Value("string"), |
|
|
"attribute": datasets.Value("string"), |
|
|
"answers": [datasets.Value("string")], |
|
|
"saliency": datasets.Value("float32") |
|
|
}] |
|
|
}), |
|
|
data_url = _URLS, |
|
|
citation = textwrap.dedent( |
|
|
"""\ |
|
|
@inproceedings{ |
|
|
title={}, |
|
|
author={}, |
|
|
booktitle={}, |
|
|
year={} |
|
|
}""" |
|
|
), |
|
|
url = _HOMEPAGE |
|
|
), |
|
|
OpenpiConfig( |
|
|
name = "Task 1", |
|
|
description = textwrap.dedent( |
|
|
"""\ |
|
|
Given paragraph (e.g., with 5 steps), identify entities that change |
|
|
(challenge: implicit entities, some explicit entities that don’t change).""" |
|
|
), |
|
|
features = datasets.Features({ |
|
|
"steps": [datasets.Value("string")], |
|
|
"entity_changes": [[datasets.Value("string")]] |
|
|
}), |
|
|
data_url = _URLS, |
|
|
citation = textwrap.dedent( |
|
|
"""\ |
|
|
@inproceedings{ |
|
|
title={}, |
|
|
author={}, |
|
|
booktitle={}, |
|
|
year={} |
|
|
}""" |
|
|
), |
|
|
url = _HOMEPAGE |
|
|
), |
|
|
OpenpiConfig( |
|
|
name = "Task 3", |
|
|
description = textwrap.dedent( |
|
|
"""\ |
|
|
Given paragraph (e.g., with 5 steps), identify the attributes of entity that change |
|
|
(challenge: implicit entities, attributes & many combinations).""" |
|
|
), |
|
|
features = datasets.Features({ |
|
|
"steps": [datasets.Value("string")], |
|
|
"attr_entity_changes": [datasets.Value("string")] |
|
|
}), |
|
|
data_url = _URLS, |
|
|
citation = textwrap.dedent( |
|
|
"""\ |
|
|
@inproceedings{ |
|
|
title={}, |
|
|
author={}, |
|
|
booktitle={}, |
|
|
year={} |
|
|
}""" |
|
|
), |
|
|
url = _HOMEPAGE |
|
|
), |
|
|
OpenpiConfig( |
|
|
name = "Task 4", |
|
|
description = textwrap.dedent( |
|
|
"""\ |
|
|
Task 4: Given paragraph & an entity, identify the sequence of attribute value changes |
|
|
(challenge: implicit attributes).""" |
|
|
), |
|
|
features = datasets.Features({ |
|
|
"steps": [datasets.Value("string")], |
|
|
"entity": datasets.Value("string"), |
|
|
"attribute_changes": [[datasets.Value("string")]] |
|
|
}), |
|
|
data_url = _URLS, |
|
|
citation = textwrap.dedent( |
|
|
"""\ |
|
|
@inproceedings{ |
|
|
title={}, |
|
|
author={}, |
|
|
booktitle={}, |
|
|
year={} |
|
|
}""" |
|
|
), |
|
|
url = _HOMEPAGE |
|
|
), |
|
|
OpenpiConfig( |
|
|
name = "Task 7", |
|
|
description = textwrap.dedent( |
|
|
"""\ |
|
|
Task 7: Given image url, identify the visual attributes of entity and |
|
|
non-visual attributes of entity that change.""" |
|
|
), |
|
|
features = datasets.Features({ |
|
|
"image_url": datasets.Value("string"), |
|
|
"visual_attr": [datasets.Value("string")], |
|
|
"non_visual_attr": [datasets.Value("string")] |
|
|
}), |
|
|
data_url = _URLS, |
|
|
citation = textwrap.dedent( |
|
|
"""\ |
|
|
@inproceedings{ |
|
|
title={}, |
|
|
author={}, |
|
|
booktitle={}, |
|
|
year={} |
|
|
}""" |
|
|
), |
|
|
url = _HOMEPAGE |
|
|
), |
|
|
] |
|
|
|
|
|
def _info(self): |
|
|
return datasets.DatasetInfo( |
|
|
description = _OPENPI_V2_DESCRIPTION, |
|
|
features = self.config.features, |
|
|
supervised_keys = None, |
|
|
homepage = self.config.url, |
|
|
citation = self.config.citation + "\n" + _OPENPI_V2_CITATION |
|
|
) |
|
|
|
|
|
def _split_generators(self, dl_manager): |
|
|
downloaded_files = dl_manager.download_and_extract(self.config.data_url) |
|
|
|
|
|
return [ |
|
|
datasets.SplitGenerator( |
|
|
name = datasets.Split.TRAIN, |
|
|
gen_kwargs = { |
|
|
"filepath": downloaded_files["train"] |
|
|
}, |
|
|
), |
|
|
datasets.SplitGenerator( |
|
|
name = datasets.Split.VALIDATION, |
|
|
gen_kwargs = { |
|
|
"filepath": downloaded_files["dev"] |
|
|
}, |
|
|
), |
|
|
datasets.SplitGenerator( |
|
|
name = datasets.Split.TEST, |
|
|
gen_kwargs = { |
|
|
"filepath": downloaded_files["test"] |
|
|
} |
|
|
), |
|
|
] |
|
|
|
|
|
|
|
|
@staticmethod |
|
|
def change_occur(dataset): |
|
|
for step in dataset: |
|
|
if len(step) > 0: |
|
|
return True |
|
|
|
|
|
return False |
|
|
|
|
|
@staticmethod |
|
|
def find_change(key, dataset): |
|
|
res = [] |
|
|
|
|
|
for state in dataset: |
|
|
if OpenpiV2.change_occur(state["answers"]): |
|
|
list_key = state[key].split(" | ") |
|
|
res.append([el for el in list_key]) |
|
|
|
|
|
return (res) |
|
|
|
|
|
@staticmethod |
|
|
def find_attr_entity_change(dataset): |
|
|
attr_change = [] |
|
|
|
|
|
for state in dataset: |
|
|
if OpenpiV2.change_occur(state["answers"]): |
|
|
change_str = "[" + state["attribute"] + "] of [" + state["entity"] + "] changed" |
|
|
attr_change.append(change_str) |
|
|
|
|
|
return attr_change |
|
|
|
|
|
def _generate_examples(self, filepath): |
|
|
logger.info("generating examples from = %s", filepath) |
|
|
|
|
|
if self.config.name == "openpi_text": |
|
|
with open(filepath) as f: |
|
|
dataset = json.load(f) |
|
|
|
|
|
for id_ in dataset: |
|
|
yield int(id_), { |
|
|
"goal": dataset[id_]["goal"], |
|
|
"steps": dataset[id_]["steps"], |
|
|
"topics": dataset[id_]["topics"], |
|
|
"image_urls": dataset[id_]["image_urls"], |
|
|
"states": dataset[id_]["states"], |
|
|
} |
|
|
|
|
|
elif self.config.name == "Task 1": |
|
|
with open(filepath) as f: |
|
|
dataset = json.load(f) |
|
|
|
|
|
for id_ in dataset: |
|
|
steps_ar = dataset[id_]["steps"] |
|
|
entity_changes = OpenpiV2.find_change("entity", dataset[id_]["states"]) |
|
|
|
|
|
yield int(id_), { |
|
|
"steps": steps_ar, |
|
|
"entity_changes": entity_changes |
|
|
} |
|
|
|
|
|
elif self.config.name == "Task 3": |
|
|
with open(filepath) as f: |
|
|
dataset = json.load(f) |
|
|
|
|
|
for id_ in dataset: |
|
|
steps_ar = dataset[id_]["steps"] |
|
|
attr_entity_changes = OpenpiV2.find_attr_entity_change(dataset[id_]["states"]) |
|
|
|
|
|
yield int(id_), { |
|
|
"steps": steps_ar, |
|
|
"attr_entity_changes": attr_entity_changes |
|
|
} |
|
|
|
|
|
elif self.config.name == "Task 4": |
|
|
with open(filepath) as f: |
|
|
dataset = json.load(f) |
|
|
|
|
|
for id_ in dataset: |
|
|
steps_ar = dataset[id_]["steps"] |
|
|
|
|
|
for state in dataset[id_]["states"]: |
|
|
for el in state["entity"].split(" | "): |
|
|
entity = el |
|
|
attribute_changes = [] |
|
|
|
|
|
for state2 in dataset[id_]["states"]: |
|
|
flag = False |
|
|
|
|
|
for el2 in state2["entity"].split(" | "): |
|
|
if entity == el2: |
|
|
flag = True |
|
|
break |
|
|
|
|
|
if flag == False: |
|
|
continue |
|
|
|
|
|
if OpenpiV2.change_occur(state2["answers"]): |
|
|
list_attribute = state2["attribute"].split(" | ") |
|
|
attribute_changes.append([el for el in list_attribute]) |
|
|
|
|
|
yield int(id_), { |
|
|
"steps": steps_ar, |
|
|
"entity": entity, |
|
|
"attribute_changes": attribute_changes |
|
|
} |
|
|
|
|
|
elif self.config.name == "Task 7": |
|
|
with open(filepath) as f: |
|
|
dataset = json.load(f) |
|
|
|
|
|
for id_ in dataset: |
|
|
N = len(dataset[id_]["image_urls"]) |
|
|
|
|
|
for i in range(N): |
|
|
image_url = dataset[id_]["image_urls"][i] |
|
|
visual_attr = [] |
|
|
non_visual_attr = [] |
|
|
|
|
|
for state in dataset[id_]["states"]: |
|
|
if len(state["answers"][i]) > 0: |
|
|
visual = False |
|
|
non_visual = False |
|
|
|
|
|
for el in state["answers_openpiv1_metadata"]["modality"][i].split(" | "): |
|
|
visual = (visual or (el == "with_image")) |
|
|
non_visual = (non_visual or (el == "without_image")) |
|
|
|
|
|
change_str = "[" + state["attribute"] + "] of [" + state["entity"] + "] changed" |
|
|
|
|
|
if visual: |
|
|
visual_attr.append(change_str) |
|
|
if non_visual: |
|
|
non_visual_attr.append(change_str) |
|
|
|
|
|
yield int(id_), { |
|
|
"image_url": image_url, |
|
|
"visual_attr": visual_attr, |
|
|
"non_visual_attr": non_visual_attr |
|
|
} |
|
|
|