--- license: mit language: - en tags: - chemistry - physics - math - biology - science pretty_name: open-rl size_categories: - n<1K task_categories: - question-answering --- # Open-RL [![License: MIT](https://img.shields.io/badge/License-MIT-green.svg)](https://opensource.org/licenses/MIT) [![Turing](https://img.shields.io/badge/Org-Turing-blue)](https://turing.com) --- ## Dataset Summary This dataset contains **self-contained, verifiable, and unambiguous STEM reasoning problems** across Physics, Mathematics, Biology, and Chemistry. Each problem: * Requires multi-step reasoning * Involves symbolic manipulation and/or numerical computation * Has a deterministic, objectively verifiable final answer The problems were evaluated against contemporary large language models. Observed pass rates indicate that the tasks are **non-trivial yet solvable**, placing them within reach of advanced models while still exposing meaningful reasoning gaps. This makes the dataset particularly suitable for: * Reinforcement learning (RL) fine-tuning * Reward modeling * Outcome-supervised training * Verifiable reasoning benchmarks --- ## Dataset Structure | Field | Type | Description | | ----------------- | ------ | ----------------------------------------- | | `conversation_id` | string | Unique identifier for each QA pair. | | `domain` | string | Physics, Math, Chemistry, Biology. | | `sub_domain` | string | Specific discipline. | | `question` | string | STEM problem statement (LaTeX supported). | | `answer` | string | Deterministic ground-truth solution. | --- ## Example ```json { "conversation_id": "217998", "domain": "Physics", "sub_domain": "Astrophysics", "question": "Consider a Navarro–Frenk–White (NFW) dark matter halo profile where...", "answer": "\( \frac{4GM_{0}}{r_{0}} + \frac{16\pi Gk}{r_{0}}\left[ \ln\left(\frac{r_{0}}{r_{s}}\right) + 0.31 \right] \)" } ``` --- ## Verifiability and Automatic Grading A core design principle of this dataset is **objective verifiability**. Each problem is constructed such that: * The final answer is deterministic * Correctness can be evaluated programmatically * No subjective interpretation is required * There is a clear separation between reasoning steps and final outcome ### Answer Types The dataset includes answers that are: * Closed-form symbolic expressions * Numerical scalars * Algebraic identities * Simplified analytic forms * Canonical LaTeX representations Because answers are deterministic, evaluation can be performed via: * Exact string matching (after normalization) * Symbolic equivalence checking (e.g., SymPy) * Numerical tolerance comparison * Unit consistency validation (where applicable) --- ## Data Quality Assurance Process To ensure scientific validity of the answer, all tasks are prepared and reviewed twice by PhD experts. Key quality rubrics include: * Prompt and answer accuracy * Clarity of prompt and underlying reasoning * Expert-verified model breaking cases due to model’s incorrect reasoning process * Google-proof originality validation. --- ## Reinforcement Learning and Outcome Supervision This dataset is designed to support **outcome-based reinforcement learning** for reasoning models. In contrast to preference-based RL (RLHF), which relies on subjective ranking signals, this dataset enables: * Outcome-supervised reinforcement learning (OSRL) * Deterministic reward assignment * Binary or graded correctness rewards * Scalable automated evaluation ### Example RL Setup Given: * Prompt: `question` * Model output: predicted final answer Reward can be computed as: * `+1` if the final answer matches ground truth * `0` or `-1` otherwise * Optional partial credit via symbolic or numerical closeness This allows: * Policy gradient methods (e.g., PPO) * Direct optimization against correctness signals * Reward model bootstrapping * Iterative self-improvement pipelines ### Calibration Regime The problems were stress-tested against advanced language models and found to be: * Not trivially solved * Not universally failed * Within the capability frontier of modern LLMs This places them in a **learning-efficient regime**: * Hard enough to produce gradient signal * Solvable enough to avoid reward sparsity * Suitable for curriculum-style training --- ## Future Directions: NuRL and Structured Nudging We plan to extend this dataset with additional problem sets and a structured **"nudge" augmentation layer** inspired by the paper *["Nudging the Boundaries of LLM Reasoning"](https://arxiv.org/html/2509.25666v1)*. ### Motivation Standard online RL algorithms (e.g., GRPO-style approaches) can only learn from problems where the model occasionally produces correct rollouts. For sufficiently difficult problems with a **0% pass rate**, no reward signal is generated, and therefore no gradient updates occur. As a result, such problems cannot contribute to expanding the model’s reasoning frontier. ### NuRL-Style Nudging To address this limitation, future versions of this dataset will include: * Abstract, high-level **hints ("nudges")** * Hints generated conditionally using the gold answer * Carefully designed cues that reduce problem difficulty without revealing the solution Under a NuRL-style training pipeline: 1. Rollouts are first generated without hints. 2. If pass rate > 0%, standard RL proceeds. 3. If pass rate = 0%, a structured hint is injected. 4. A new batch of trajectories is generated with the hint. This enables: * Previously unsolvable samples to produce non-zero rewards * Learning signal from frontier-level problems * Expansion of the model’s upper reasoning bound ### Design Principles for Effective Nudges Planned nudges will follow empirical findings from prior work: * Hints should be **abstract and knowledge-oriented**, not answer-revealing * Hints should preserve distributional alignment with base policy reasoning * Hints should be injected only when necessary * Nudges are most effective after base RL convergence --- This evolution positions the dataset not only as a verifiable benchmark, but as a controlled testbed for **upper-bound expansion in reinforcement learning for reasoning models**. --- ## Citation ```bibtex @dataset{turing_2026_open_rl, title = {Open-RL }, author = {Saurabh Patil, Anshuman Lall, Marko Pavlovic , Chinmayee Shukla, Seetesh Pande, Tejass Mohan Ukarde , Amanda Gollo Bertollo, Mahesh Joshi, Kihwan Han}, year = {2026}, url = {https://huggingface.co/datasets/TuringEnterprises/Open-RL} } ```