File size: 1,819 Bytes
6f022aa |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 |
---
license: mit
task_categories:
- question-answering
- text-generation
language:
- en
tags:
- science
- physics
- biology
- chemistry
- experimental-prediction
- benchmark
size_categories:
- n<1K
---
# SciPredict: Can LLMs Predict the Outcomes of Research Experiments?
**Paper:** SciPredict: Can LLMs Predict the Outcomes of Research Experiments in Natural Sciences?
## Overview
SciPredict is a benchmark evaluating whether AI systems can predict experimental outcomes in physics, biology, and chemistry. The dataset comprises **405 questions** derived from recently published empirical studies (post-March 2025), spanning **33 subdomains**.
## Dataset Structure
- **Total Questions:** 405 (5,716 rows including model responses)
- **Domains:** Physics (9 subdomains), Chemistry (10 subdomains), Biology (14 subdomains)
- **Question Formats:** Multiple-choice (MCQ), Free-format, Numerical
### Key Fields
- `DOMAIN`: Scientific domain (Physics, Biology, Chemistry)
- `FIELD`: Specific field within the domain
- `PQ_FORMAT`: Question format (MCQ, Free-Format, Numerical)
- `TITLE`: Paper title
- `URL`: Paper URL
- `PUBLISHING_DATE`: Publication date
- `EXPERIMENTAL_SETUP`: Description of the experimental configuration
- `MEASUREMENT_TAKEN`: What was measured in the experiment
- `OUTCOME_PREDICTION_QUESTION`: The prediction task
- `GTA`: Ground truth answer
- `BACKGROUND_KNOWLEDGE`: Expert-curated background knowledge
- `RELATED_PAPERS_DATA`: Related papers information
## Key Findings
- **Model accuracy:** 14-26% (vs. ~20% human expert accuracy)
- **Poor calibration:** Models cannot distinguish reliable from unreliable predictions
- **Background knowledge helps:** Providing expert-curated context improves performance
- **Format matters:** Performance degrades from MCQ → Free-form → Numerical
|