s-pp / README.md
glowsenior's picture
Upload folder using huggingface_hub
7cf9e66 verified
---
license: apache-2.0
base_model: Qwen/Qwen3-32B
tags:
- reinforcement-learning
- rl
- ppo
- skyrl
- code
- reasoning
datasets:
- penfever/r2egym_gpt5_codex_solved_tasks_256_subset
pipeline_tag: text-generation
model-index:
- name: Qwen3-32B-R2EGYM-256-3epochs
results: []
---
# Qwen3-32B-R2EGYM-256-3epochs
This model is a reinforcement learning fine-tuned version of [Qwen/Qwen3-32B](https://huggingface.co/Qwen/Qwen3-32B), trained using the SkyRL framework with fully asynchronous PPO on coding and reasoning tasks from the R2EGYM benchmark.
## Training Details
### Framework
- **Training Framework**: [SkyRL](https://github.com/NovaSkyAI/SkyRL) (fully async PPO)
- **Parallelism Strategy**: FSDP2 with CPU offload
- **Agent**: Terminus-2 (terminal-based coding agent with thinking enabled)
### Dataset
- **Dataset**: [penfever/r2egym_gpt5_codex_solved_tasks_256_subset](https://huggingface.co/datasets/penfever/r2egym_gpt5_codex_solved_tasks_256_subset)
- **Number of tasks**: 256
- **Evaluation set**: OpenThoughts-TB-dev (70 tasks)
### Hyperparameters
| Parameter | Value |
|---|---|
| **Epochs** | 3 |
| **Total steps** | 12 (4 steps/epoch) |
| **Learning rate** | 1e-5 |
| **Weight decay** | 0.0 |
| **Train batch size** | 64 |
| **Micro train batch size per GPU** | 1 |
| **Advantage estimator** | rloo_n |
| **KL loss** | disabled |
| **Samples per prompt** | 8 |
| **Max prompt length** | 2,048 |
| **Max generate length** | 30,720 |
| **RoPE scaling** | yarn (factor=4.0, original_max_position_embeddings=32,768) |
### Infrastructure
| Component | Configuration |
|---|---|
| **Policy nodes** | 4 nodes x 4 GPUs |
| **Reference model nodes** | 4 nodes x 4 GPUs |
| **Inference engines** | 26 (tensor parallelism = 2) |
| **Parallel generation workers** | 96 |
| **Concurrent sandbox trials** | 96 |
| **Total training nodes** | 17 |
### Training Notes
- Training was resumed from a step-9 checkpoint
- The model uses Terminus-2, a terminal-based coding agent that interacts with sandboxed Docker environments to solve programming tasks
- Thinking mode was enabled during training (`--enable_thinking`)
## Usage
This model can be used as a drop-in replacement for Qwen3-32B with improved coding and reasoning capabilities.
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("laion/Qwen3-32B-R2EGYM-256-3epochs")
tokenizer = AutoTokenizer.from_pretrained("laion/Qwen3-32B-R2EGYM-256-3epochs")
```