You need to agree to share your contact information to access this dataset

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this dataset content.

DEEPSYNTH: A Benchmark for Deep Information Synthesis

DEEPSYNTH Bench
Published at ICLR 2026  |  📄 Paper  |  💻 Code  |  🌐 Project Page

Image

Overview

DEEPSYNTH-Bench is a challenging benchmark for evaluating deep information synthesis — the ability of AI systems to integrate, reason over, and consolidate multi-source information into precise, structured answers.

Unlike benchmarks focused on retrieval or single-hop reasoning, DEEPSYNTH-Bench requires models to:

  • Chain multiple reasoning steps across heterogeneous sources
  • Produce structured JSON outputs with specific keys and values
  • Demonstrate analytical depth, not just surface-level extraction

The benchmark includes a public dev set of 40 tasks with gold answers, full decompositions, and intermediate steps for iterative development, and a test set of 80 tasks (questions only) for clean evaluation — 120 tasks in total.


Repository Structure

deepsynth-bench/
├── README.md                  # This dataset card
├── data/
│   ├── test.jsonl   # Full test set (80 tasks)
│   └── dev.jsonl    # Dev/Lite split for prototyping ((40 tasks))
├── evaluation/
│   ├── evaluate.py            # Evaluation script (F1, EM, LLM-Judge)
│   └── llm_judge_prompt.txt   # Prompt used for LLM-as-a-judge metric
├── assets/
│   └── octopus_logo.png       # Project logo
└── LICENSE                    # CC-BY-4.0

Dataset Files

File Split Size Description
dev.json Dev 40 tasks Questions, gold answers, reasoning plans, and full decompositions with intermediate steps
test.json Test 80 tasks Questions only — submit answers for evaluation

Loading the Data

import json
from huggingface_hub import hf_hub_download

# Dev set — includes gold answers
dev_path = hf_hub_download(
    repo_id="DeepSynthesisTeam/deepsynth-bench",
    filename="data/dev.json",
    repo_type="dataset"
)
with open(dev_path, "r") as f:
    dev_set = json.load(f)

# Test set — questions only
test_path = hf_hub_download(
    repo_id="DeepSynthesisTeam/deepsynth-bench",
    filename="data/test.json",
    repo_type="dataset"
)
with open(test_path, "r") as f:
    test_set = json.load(f)

Prediction Format

Model predictions should be a JSON file mapping task IDs to answer dictionaries:

{
  "001": {"Sweden": 1.2, "Finland": 0.8},
  "002": {"Brunei": -0.67, "Singapore": -0.34}
}

Evaluation

Evaluation scripts are available in the GitHub repository.

Metric Description
Exact Match (EM) All keys and values must be exactly correct
F1 Score Partial credit for correct key-value pairs
LLM Judge Semantic equivalence; allows small numerical margins (1–5.5%)
# Clone the repository to access evaluation scripts
git clone https://github.com/agentdeepsynthesis/deepsynth-bench.git
cd deepsynth-bench

# Run EM + F1 evaluation
python scripts/evaluation/eval_static_score.py your_predictions.json

# Run LLM-as-judge evaluation
python scripts/evaluation/llm_judge.py your_predictions.json

🧩 Decompositions & Validation Schemas

Decomposition Files (decompositions/*.json)

Each file (e.g., 001.json) maps the logical sub-steps required to solve the corresponding question. These decompositions support step-by-step evaluation and can be used to guide or audit model reasoning chains.

Validation Schemas (intermediate_answers_schemas/)

Each decomposition has a matching JSON Schema (e.g., 001.schema.json) that defines the expected format for intermediate answer fields. Use these to programmatically validate whether a model's intermediate outputs conform to the expected structure.


Citation

If you use DEEPSYNTH-Bench in your research, please cite:

@inproceedings{paul-etal-2026-deepinfosynth,
  title = {A Benchmark for Deep Information Synthesis},
  author = {Paul, Debjit and Murphy, Daniel and Gritta, Milan and Cardenas, Ronald and Prokhorov, Victor and Bolliger, Lena Sophia and Toker, Aysim and Miles, Roy and Oncescu, Andreea-Maria and Sivakumar, Jasivan Alex and Borchert, Philipp and Elezi, Ismail and Zhang, Meiru and Lee, Ka Yiu and Zhang, Guchun and Wang, Jun and Lampouras, Gerasimos},
  booktitle = {The Fourteenth International Conference on Learning Representations},
  month = apr,
  year = {2026},
}

License

We follow Apache License Version 2.0. Please see the License file for more information.

Disclaimer: This open source project is not an official Huawei product, Huawei is not expected to provide support for this project.

Downloads last month
52