saurabh-gnirut commited on
Commit
650f99c
·
verified ·
1 Parent(s): e7bf092

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +203 -3
README.md CHANGED
@@ -1,3 +1,203 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ language:
4
+ - en
5
+ tags:
6
+ - chemistry
7
+ - physics
8
+ - math
9
+ - biology
10
+
11
+ pretty_name: sci-rl
12
+ size_categories:
13
+ - n<1K
14
+ task_categories:
15
+ - question-answering
16
+ ---
17
+
18
+ # STEM Verifiable QA (Sci-RL)
19
+
20
+ [![License: MIT](https://img.shields.io/badge/License-MIT-green.svg)](https://opensource.org/licenses/MIT)
21
+ [![Turing](https://img.shields.io/badge/Org-Turing-blue)](https://turing.com)
22
+
23
+ ---
24
+
25
+ ## Dataset Summary
26
+
27
+ This dataset contains **self-contained, verifiable, and unambiguous STEM reasoning problems** across Physics, Mathematics, Biology, and Chemistry.
28
+
29
+ Each problem:
30
+
31
+ * Requires multi-step reasoning
32
+ * Involves symbolic manipulation and/or numerical computation
33
+ * Has a deterministic, objectively verifiable final answer
34
+
35
+ The problems were evaluated against contemporary large language models. Observed pass rates indicate that the tasks are **non-trivial yet solvable**, placing them within reach of advanced models while still exposing meaningful reasoning gaps.
36
+
37
+ This makes the dataset particularly suitable for:
38
+
39
+ * Reinforcement learning (RL) fine-tuning
40
+ * Reward modeling
41
+ * Outcome-supervised training
42
+ * Verifiable reasoning benchmarks
43
+
44
+ ---
45
+
46
+ ## Dataset Structure
47
+
48
+ | Field | Type | Description |
49
+ | ----------------- | ------ | ----------------------------------------- |
50
+ | `conversation_id` | string | Unique identifier for each QA pair. |
51
+ | `domain` | string | Physics, Math, Chemistry, Biology. |
52
+ | `sub-domain` | string | Specific discipline. |
53
+ | `question` | string | STEM problem statement (LaTeX supported). |
54
+ | `answer` | string | Deterministic ground-truth solution. |
55
+
56
+ ---
57
+
58
+ ## Example
59
+
60
+ ```json
61
+ {
62
+ "conversation_id": "217998",
63
+ "domain": "Physics",
64
+ "sub-domain": "Astrophysics",
65
+ "question": "Consider a Navarro–Frenk–White (NFW) dark matter halo profile where...",
66
+ "answer": "\( \frac{4GM_{0}}{r_{0}} + \frac{16\pi Gk}{r_{0}}\left[ \ln\left(\frac{r_{0}}{r_{s}}\right) + 0.31 \right] \)"
67
+ }
68
+ ```
69
+
70
+ ---
71
+
72
+ ## Verifiability and Automatic Grading
73
+
74
+ A core design principle of this dataset is **objective verifiability**.
75
+
76
+ Each problem is constructed such that:
77
+
78
+ * The final answer is deterministic
79
+ * Correctness can be evaluated programmatically
80
+ * No subjective interpretation is required
81
+ * There is a clear separation between reasoning steps and final outcome
82
+
83
+ ### Answer Types
84
+
85
+ The dataset includes answers that are:
86
+
87
+ * Closed-form symbolic expressions
88
+ * Numerical scalars
89
+ * Algebraic identities
90
+ * Simplified analytic forms
91
+ * Canonical LaTeX representations
92
+
93
+ Because answers are deterministic, evaluation can be performed via:
94
+
95
+ * Exact string matching (after normalization)
96
+ * Symbolic equivalence checking (e.g., SymPy)
97
+ * Numerical tolerance comparison
98
+ * Unit consistency validation (where applicable)
99
+
100
+ ---
101
+
102
+ ## Reinforcement Learning and Outcome Supervision
103
+
104
+ This dataset is designed to support **outcome-based reinforcement learning** for reasoning models.
105
+
106
+ In contrast to preference-based RL (RLHF), which relies on subjective ranking signals, this dataset enables:
107
+
108
+ * Outcome-supervised reinforcement learning (OSRL)
109
+ * Deterministic reward assignment
110
+ * Binary or graded correctness rewards
111
+ * Scalable automated evaluation
112
+
113
+ ### Example RL Setup
114
+
115
+ Given:
116
+
117
+ * Prompt: `question`
118
+ * Model output: predicted final answer
119
+
120
+ Reward can be computed as:
121
+
122
+ * `+1` if the final answer matches ground truth
123
+ * `0` or `-1` otherwise
124
+ * Optional partial credit via symbolic or numerical closeness
125
+
126
+ This allows:
127
+
128
+ * Policy gradient methods (e.g., PPO)
129
+ * Direct optimization against correctness signals
130
+ * Reward model bootstrapping
131
+ * Iterative self-improvement pipelines
132
+
133
+ ### Calibration Regime
134
+
135
+ The problems were stress-tested against advanced language models and found to be:
136
+
137
+ * Not trivially solved
138
+ * Not universally failed
139
+ * Within the capability frontier of modern LLMs
140
+
141
+ This places them in a **learning-efficient regime**:
142
+
143
+ * Hard enough to produce gradient signal
144
+ * Solvable enough to avoid reward sparsity
145
+ * Suitable for curriculum-style training
146
+
147
+ ---
148
+
149
+
150
+ ## Future Directions: NuRL and Structured Nudging
151
+
152
+ We plan to extend this dataset with additional problem sets and a structured **"nudge" augmentation layer** inspired by the paper *["Nudging the Boundaries of LLM Reasoning"](https://arxiv.org/html/2509.25666v1)*.
153
+
154
+ ### Motivation
155
+
156
+ Standard online RL algorithms (e.g., GRPO-style approaches) can only learn from problems where the model occasionally produces correct rollouts. For sufficiently difficult problems with a **0% pass rate**, no reward signal is generated, and therefore no gradient updates occur. As a result, such problems cannot contribute to expanding the model’s reasoning frontier.
157
+
158
+ ### NuRL-Style Nudging
159
+
160
+ To address this limitation, future versions of this dataset will include:
161
+
162
+ * Abstract, high-level **hints ("nudges")**
163
+ * Hints generated conditionally using the gold answer
164
+ * Carefully designed cues that reduce problem difficulty without revealing the solution
165
+
166
+ Under a NuRL-style training pipeline:
167
+
168
+ 1. Rollouts are first generated without hints.
169
+ 2. If pass rate > 0%, standard RL proceeds.
170
+ 3. If pass rate = 0%, a structured hint is injected.
171
+ 4. A new batch of trajectories is generated with the hint.
172
+
173
+ This enables:
174
+
175
+ * Previously unsolvable samples to produce non-zero rewards
176
+ * Learning signal from frontier-level problems
177
+ * Expansion of the model’s upper reasoning bound
178
+
179
+ ### Design Principles for Effective Nudges
180
+
181
+ Planned nudges will follow empirical findings from prior work:
182
+
183
+ * Hints should be **abstract and knowledge-oriented**, not answer-revealing
184
+ * Hints should preserve distributional alignment with base policy reasoning
185
+ * Hints should be injected only when necessary
186
+ * Nudges are most effective after base RL convergence
187
+
188
+ ---
189
+
190
+ This evolution positions the dataset not only as a verifiable benchmark, but as a controlled testbed for **upper-bound expansion in reinforcement learning for reasoning models**.
191
+
192
+ ---
193
+
194
+ ## Citation
195
+
196
+ ```bibtex
197
+ @dataset{saurabh_2025_sci_rl,
198
+ title = {STEM Verifiable QA (Sci-RL)},
199
+ author = {Saurabh Patil and Anshuman Lall and Marko Pavlovic and Tejas Ukarde and Chinmayee Shukla and Mahesh Joshi and Kihwan Han},
200
+ year = {2026},
201
+ url = {https://huggingface.co/datasets/TuringEnterprises/Turing-Open-Reasoning/}
202
+ }
203
+ ```