SyntacticAgreement / SyntacticAgreement.py
alba.taboas
updated SyntacticAgreement.py for nested .zip files
7997120
import os
import json
from datasets import (
GeneratorBasedBuilder,
DatasetInfo,
Features,
Value,
Sequence,
Split,
SplitGenerator,
BuilderConfig,
)
# -------------------------
# Features specification
# -------------------------
SUITE_DATASET_CONDITION_SPEC = {
"condition_name": Value("string"),
"content": Value("string"),
"regions": Sequence({
"region_number": Value("int32"),
"content": Value("string"),
}),
}
SUITE_DATASET_SPEC = Features({
"suite_name": Value("string"),
"item_number": Value("int32"),
"conditions": Sequence(SUITE_DATASET_CONDITION_SPEC),
"predictions": Sequence(Value("string")),
})
# -------------------------
# Dataset builder
# -------------------------
class SyntacticAgreement(GeneratorBasedBuilder):
BUILDER_CONFIG_CLASS = BuilderConfig
BUILDER_CONFIGS = [
BuilderConfig(
name="spanish",
description="Spanish syntax test suites",
),
BuilderConfig(
name="italian",
description="Italian syntax test suites",
),
BuilderConfig(
name="portuguese",
description="Portuguese syntax test suites",
),
BuilderConfig(
name="russian",
description="Russian syntax test suites",
),
]
def _info(self):
return DatasetInfo(
description="Agreement syntax test suites in multiple languages.",
features=SUITE_DATASET_SPEC,
supervised_keys=None,
homepage="",
license="",
citation="https://aclanthology.org/2025.depling-1.4/",
)
def _split_generators(self, dl_manager):
# URL to the zip archive of the selected language
base_url = "https://huggingface.co/datasets/albalbalba/SyntacticAgreement/resolve/main"
archive_url = f"{base_url}/{self.config.name}.zip"
extracted_dir = dl_manager.download_and_extract(archive_url)
# Handle case where extraction produces a nested folder (italian/italian/*.json)
possible_nested = os.path.join(extracted_dir, self.config.name)
if os.path.exists(possible_nested):
data_dir = possible_nested
else:
data_dir = extracted_dir
return [
SplitGenerator(
name=Split.TRAIN,
gen_kwargs={"data_dir": data_dir},
),
]
def _generate_examples(self, data_dir):
for filename in sorted(os.listdir(data_dir)):
if not filename.endswith(".json"):
continue
filepath = os.path.join(data_dir, filename)
with open(filepath, encoding="utf-8") as f:
data = json.load(f)
suite_name = data["meta"]["name"]
predictions = [p["formula"] for p in data.get("predictions", [])]
for item in data.get("items", []):
item_number = item["item_number"]
conditions = []
for condition in item.get("conditions", []):
content_all = " ".join(
[region["content"] for region in condition["regions"]]
).lstrip(" ")
content_all = (
content_all.replace(" ", " ")
.replace(" .", ".")
.replace(" ,", ",")
.replace("¿ ", "¿")
)
conditions.append({
"condition_name": condition["condition_name"],
"content": content_all,
"regions": [
{
"region_number": region["region_number"],
"content": region["content"],
}
for region in condition["regions"]
],
})
key = f"{suite_name}_{item_number}"
yield key, {
"suite_name": suite_name,
"item_number": item_number,
"conditions": conditions,
"predictions": predictions,
}