import argparse import os from datasets import DatasetDict, concatenate_datasets, load_from_disk SOURCE_SPECS = { "qqp": {"sentence1": "question1", "sentence2": "question2", "keep": lambda label: label == 1}, "mrpc": {"sentence1": "sentence1", "sentence2": "sentence2", "keep": lambda label: label == 1}, "stsb": {"sentence1": "sentence1", "sentence2": "sentence2", "keep": lambda label: label >= 3}, "rte": {"sentence1": "sentence1", "sentence2": "sentence2", "keep": lambda label: label == 1}, "anli": {"sentence1": "sentence1", "sentence2": "sentence2", "keep": lambda label: label == 0}, "smf": {"sentence1": "sentence1", "sentence2": "sentence2", "keep": lambda label: label == 0}, } def parse_args(): parser = argparse.ArgumentParser() parser.add_argument("--data_dir", default="./data") parser.add_argument("--sources", default="qqp,mrpc,stsb") parser.add_argument("--output", default="ling_conversion") return parser.parse_args() def rename_dev_split(data): if "dev" not in data and "validation" in data: data["dev"] = data["validation"] del data["validation"] return data def normalize_columns(data, source_name, sentence1_name, sentence2_name): for split in data.keys(): split_data = data[split] rename_map = {} if sentence1_name in split_data.column_names and sentence1_name != "sentence1": rename_map[sentence1_name] = "sentence1" if sentence2_name in split_data.column_names and sentence2_name != "sentence2": rename_map[sentence2_name] = "sentence2" if rename_map: split_data = split_data.rename_columns(rename_map) if "source" not in split_data.column_names: split_data = split_data.add_column("source", [source_name] * len(split_data)) keep_columns = [column for column in split_data.column_names if column.startswith("sentence") or column == "source"] data[split] = split_data.remove_columns(sorted(set(split_data.column_names) - set(keep_columns))) return data def main(): args = parse_args() sources = [source.strip() for source in args.sources.split(",") if source.strip()] datasets = {} for source_name in sources: if source_name not in SOURCE_SPECS: raise ValueError(f"Unsupported source dataset: {source_name}") spec = SOURCE_SPECS[source_name] dataset = load_from_disk(os.path.join(args.data_dir, source_name)) dataset = rename_dev_split(dataset) dataset = dataset.filter(lambda row: spec["keep"](row["label"])) dataset = normalize_columns(dataset, source_name, spec["sentence1"], spec["sentence2"]) datasets[source_name] = dataset merged = DatasetDict( { "train": concatenate_datasets([dataset["train"] for dataset in datasets.values() if "train" in dataset]), "dev": concatenate_datasets([dataset["dev"] for dataset in datasets.values() if "dev" in dataset]), "test": concatenate_datasets([dataset["test"] for dataset in datasets.values() if "test" in dataset]), } ) merged.save_to_disk(os.path.join(args.data_dir, args.output)) if __name__ == "__main__": main()