Update README.md
Browse files
README.md
CHANGED
|
@@ -29,7 +29,7 @@ The resulting model is the **InfiR2-7B-Instruct-FP8**.
|
|
| 29 |
|
| 30 |
**Training Recipe**:
|
| 31 |
<p align="center">
|
| 32 |
-
<img src="fp8_recipe.png" width="
|
| 33 |
<p>
|
| 34 |
|
| 35 |
- Stable and Reproducible Performance
|
|
@@ -47,7 +47,7 @@ The InfiR2 framework offers multiple variants model with different size and trai
|
|
| 47 |
- **7B**
|
| 48 |
- [InfiR2-7B-base-FP8](https://huggingface.co/InfiX-ai/InfiR2-7B-base-FP8): *Continue pretrain on Qwen2.5-7B-base*
|
| 49 |
- [InfiR2-7B-Instruct-FP8](https://huggingface.co/InfiX-ai/InfiR2-7B-Instruct-FP8): *Supervised fine-tuning on InfiR2-7B-base-FP8 with [InfiAlign dataset](https://huggingface.co/papers/2508.05496)*
|
| 50 |
-
- [InfiR2-R1-7B-FP8](https://huggingface.co/InfiX-ai/InfiR2-R1-7B-FP8): *
|
| 51 |
|
| 52 |
## 📊 Model Performance
|
| 53 |
Below is the performance comparison of InfiR2-7B-Instruct-FP8 on reasoning benchmarks. Note: 'w. InfiAlign' denotes Supervised Fine-Tuning (SFT) using the InfiAlign dataset.
|
|
|
|
| 29 |
|
| 30 |
**Training Recipe**:
|
| 31 |
<p align="center">
|
| 32 |
+
<img src="fp8_recipe.png" width="80%"/>
|
| 33 |
<p>
|
| 34 |
|
| 35 |
- Stable and Reproducible Performance
|
|
|
|
| 47 |
- **7B**
|
| 48 |
- [InfiR2-7B-base-FP8](https://huggingface.co/InfiX-ai/InfiR2-7B-base-FP8): *Continue pretrain on Qwen2.5-7B-base*
|
| 49 |
- [InfiR2-7B-Instruct-FP8](https://huggingface.co/InfiX-ai/InfiR2-7B-Instruct-FP8): *Supervised fine-tuning on InfiR2-7B-base-FP8 with [InfiAlign dataset](https://huggingface.co/papers/2508.05496)*
|
| 50 |
+
- [InfiR2-R1-7B-FP8-Preview](https://huggingface.co/InfiX-ai/InfiR2-R1-7B-FP8-Preview): *Multi-stage FP8 Reinforcement Learning*
|
| 51 |
|
| 52 |
## 📊 Model Performance
|
| 53 |
Below is the performance comparison of InfiR2-7B-Instruct-FP8 on reasoning benchmarks. Note: 'w. InfiAlign' denotes Supervised Fine-Tuning (SFT) using the InfiAlign dataset.
|