Update README.md
Browse files
README.md
CHANGED
|
@@ -28,7 +28,7 @@ cost-efficient to deploy.
|
|
| 28 |
- **16K context length**
|
| 29 |
- Optimized for reasoning-centric tasks
|
| 30 |
- Designed for modern GPU inference runtimes
|
| 31 |
-
- **Architecture:** Qwen3-compatible (DeepBrainz-R series
|
| 32 |
|
| 33 |
---
|
| 34 |
|
|
@@ -72,7 +72,7 @@ print(tok.decode(out[0], skip_special_tokens=True))
|
|
| 72 |
## Training Summary
|
| 73 |
|
| 74 |
The model was produced using a multi-stage optimization process involving
|
| 75 |
-
large-scale
|
| 76 |
quality and robustness. Specific training details are intentionally
|
| 77 |
abstracted in this public release.
|
| 78 |
|
|
|
|
| 28 |
- **16K context length**
|
| 29 |
- Optimized for reasoning-centric tasks
|
| 30 |
- Designed for modern GPU inference runtimes
|
| 31 |
+
- **Architecture:** Qwen3-compatible (DeepBrainz-R series post-trained and optimized for reasoning-centric workloads)
|
| 32 |
|
| 33 |
---
|
| 34 |
|
|
|
|
| 72 |
## Training Summary
|
| 73 |
|
| 74 |
The model was produced using a multi-stage optimization process involving
|
| 75 |
+
large-scale on-policy optimization and iterative refinement to improve reasoning
|
| 76 |
quality and robustness. Specific training details are intentionally
|
| 77 |
abstracted in this public release.
|
| 78 |
|