utkarsh4430 commited on
Commit
6f022aa
·
verified ·
1 Parent(s): 50ef6a9

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +53 -3
README.md CHANGED
@@ -1,3 +1,53 @@
1
- ---
2
- license: cc-by-nc-4.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ task_categories:
4
+ - question-answering
5
+ - text-generation
6
+ language:
7
+ - en
8
+ tags:
9
+ - science
10
+ - physics
11
+ - biology
12
+ - chemistry
13
+ - experimental-prediction
14
+ - benchmark
15
+ size_categories:
16
+ - n<1K
17
+ ---
18
+
19
+ # SciPredict: Can LLMs Predict the Outcomes of Research Experiments?
20
+
21
+ **Paper:** SciPredict: Can LLMs Predict the Outcomes of Research Experiments in Natural Sciences?
22
+
23
+ ## Overview
24
+
25
+ SciPredict is a benchmark evaluating whether AI systems can predict experimental outcomes in physics, biology, and chemistry. The dataset comprises **405 questions** derived from recently published empirical studies (post-March 2025), spanning **33 subdomains**.
26
+
27
+ ## Dataset Structure
28
+
29
+ - **Total Questions:** 405 (5,716 rows including model responses)
30
+ - **Domains:** Physics (9 subdomains), Chemistry (10 subdomains), Biology (14 subdomains)
31
+ - **Question Formats:** Multiple-choice (MCQ), Free-format, Numerical
32
+
33
+ ### Key Fields
34
+
35
+ - `DOMAIN`: Scientific domain (Physics, Biology, Chemistry)
36
+ - `FIELD`: Specific field within the domain
37
+ - `PQ_FORMAT`: Question format (MCQ, Free-Format, Numerical)
38
+ - `TITLE`: Paper title
39
+ - `URL`: Paper URL
40
+ - `PUBLISHING_DATE`: Publication date
41
+ - `EXPERIMENTAL_SETUP`: Description of the experimental configuration
42
+ - `MEASUREMENT_TAKEN`: What was measured in the experiment
43
+ - `OUTCOME_PREDICTION_QUESTION`: The prediction task
44
+ - `GTA`: Ground truth answer
45
+ - `BACKGROUND_KNOWLEDGE`: Expert-curated background knowledge
46
+ - `RELATED_PAPERS_DATA`: Related papers information
47
+
48
+ ## Key Findings
49
+
50
+ - **Model accuracy:** 14-26% (vs. ~20% human expert accuracy)
51
+ - **Poor calibration:** Models cannot distinguish reliable from unreliable predictions
52
+ - **Background knowledge helps:** Providing expert-curated context improves performance
53
+ - **Format matters:** Performance degrades from MCQ → Free-form → Numerical