replicatorbench / README.md
domsoos's picture
Upload folder using huggingface_hub
204f58d verified

LLM Benchmarking Project — Dataset (Scientific Replication Benchmark)

This repository contains the data-only portion of the Center for Open Science (COS) LLM Benchmarking Project. The dataset supports benchmarking LLM agents on core parts of the scientific research lifecycle—especially replication—including:

  • Information extraction from scientific papers into structured JSON
  • Research design and analysis planning
  • (Optional) execution support using provided replication datasets and code
  • Scientific interpretation using human reference materials and expected outputs

Each numbered folder corresponds to one study instance in the benchmark.

Dataset contents (per study)

Each study folder typically contains:

  • original_paper.pdf
    The published paper used as the primary input.

  • initial_details.txt
    Brief notes to orient the replication attempt (e.g., key outcomes, hints, pointers).

  • replication_data/
    Data and scripts required to reproduce analyses (common formats: .csv, .dta, .rds, .R, .do, etc.).

  • human_preregistration.(pdf|docx)
    Human-created preregistration describing the replication plan.

  • human_report.(pdf|docx)
    Human-created replication report describing analyses and findings.

  • expected_post_registration*.json
    Expert-annotated ground truth structured outputs used for evaluation.

    • expected_post_registration.json is the primary reference.
    • expected_post_registration_2.json, _3.json, etc. are acceptable alternative variants where applicable.

Some studies include multiple acceptable ground-truth variants to capture permissible differences in annotation or representation.

Repository structure

At the dataset root, folders like 1/, 2/, 10/, 11/, etc. are study IDs.

Example:

text
.
├── 1/
│   ├── expected_post_registration.json
│   ├── expected_post_registration_2.json
│   ├── human_preregistration.pdf
│   ├── human_report.pdf
│   ├── initial_details.txt
│   ├── original_paper.pdf
│   └── replication_data/
│       ├── <data files>
│       └── <analysis scripts>

Intended uses

This dataset is intended for:

  • Benchmarking LLM agents that extract structured study metadata from papers
  • Evaluating LLM systems that generate replication plans and analysis specifications
  • Comparing model outputs against expert-annotated expected JSON and human reference docs

Not intended for

  • Clinical or other high-stakes decision-making
  • Producing definitive judgments about the original papers
  • Training models to reproduce copyrighted texts verbatim

Quickstart (local)

Iterate over studies and load ground truth

python
from pathlib import Path
import json

root = Path(".")
study_dirs = sorted(
    [p for p in root.iterdir() if p.is_dir() and p.name.isdigit()],
    key=lambda p: int(p.name)
)

for study in study_dirs:
    gt = study / "expected_post_registration.json"
    if gt.exists():
        data = json.loads(gt.read_text(encoding="utf-8"))
        print(study.name, "ground truth keys:", list(data.keys())[:10])

Using with the main pipeline repository (recommended)

If you are using the LLM Benchmarking Project codebase, point the pipeline/evaluators at a given study directory:

bash
make evaluate-extract STUDY=/path/to/llm-benchmarking-data/1

The expected JSON format is defined by the main repository’s templates/schemas. Use those schemas to validate or format model outputs.

Notes on multiple expected JSON variants

Some studies include expected_post_registration_2.json, expected_post_registration_3.json, etc. This is intentional:

  • Some fields allow multiple equivalent representations
  • Annotation may vary slightly without changing meaning
  • Evaluators may accept any variant depending on scoring rules

If you implement your own scorer, consider:

  • Exact matching for strictly defined fields
  • More tolerant matching for lists, notes, or fields with legitimate paraphrase/format variation

File formats

You may encounter:

  • R artifacts: .R, .rds
  • Stata artifacts: .do, .dta
  • CSV/tabular data: .csv
  • Documents: .pdf, .docx
  • Structured evaluation targets: .json

Reproducing analyses may require R and/or Stata depending on the study.

Licensing, copyright, and redistribution (important)

This repository is released under Apache 2.0 for COS-authored materials and annotations (for example: benchmark scaffolding, expected JSON outputs, and other COS-created files).

However, some contents may be third-party materials, including (but not limited to):

  • original_paper.pdf (publisher copyright may apply)
  • replication_data/ (may have its own license/terms from the original authors)
  • external scripts or datasets included for replication

You are responsible for ensuring you have the right to redistribute third-party files publicly (e.g., GitHub / Hugging Face).

Common options if redistribution is restricted:

  • Remove third-party PDFs and provide DOI/URL references instead
  • Keep restricted files in a private location and publish only COS-authored annotations
  • Add per-study LICENSE / NOTICE files inside each study folder where terms are known

Large files (Git LFS recommendation)

If hosting on GitHub, consider Git LFS for PDFs and large datasets:

bash
git lfs install
git lfs track "*.pdf" "*.dta" "*.rds"
git add .gitattributes

Citation

If you use this dataset in academic work, please cite it as:

bibtex
@dataset{cos_llm_benchmarking_data_2026,
  author    = {Center for Open Science},
  title     = {LLM Benchmarking Project: Scientific Replication Benchmark Data},
  year      = {2026},
  publisher = {Center for Open Science},
  note      = {Benchmark dataset for evaluating LLM agents on scientific replication tasks}
}

Acknowledgements

This project is funded by Coefficient Giving as part of its “Benchmarking LLM Agents on Consequential Real-World Tasks” program. We thank the annotators who contributed to the ground-truth post-registrations for the extraction stage.

Contact

For questions about this dataset:

Shakhlo Nematova
Research Scientist, Center for Open Science
shakhlo@cos.io