Update README.md
Browse files
README.md
CHANGED
|
@@ -27,7 +27,7 @@ library_name: transformers
|
|
| 27 |
|
| 28 |
## 1. Introduction
|
| 29 |
|
| 30 |
-
**Step 3.5 Flash** ([visit website](https://static.stepfun.com/blog/step-3.5-flash/)) is our most capable open-source foundation model, engineered to deliver frontier reasoning and agentic capabilities with exceptional efficiency. We also open-sourced the training codebase, with support for continue pretrain, SFT, RL (WIP), and evaluation (WIP), and will open-source the SFT data. Built on a sparse Mixture of Experts (MoE) architecture, it selectively activates only 11B of its 196B parameters per token. This "intelligence density" allows it to rival the reasoning depth of top-tier proprietary models, while maintaining the agility required for real-time interaction.
|
| 31 |
|
| 32 |
## 2. Key Capabilities
|
| 33 |
|
|
|
|
| 27 |
|
| 28 |
## 1. Introduction
|
| 29 |
|
| 30 |
+
**Step 3.5 Flash** ([visit website](https://static.stepfun.com/blog/step-3.5-flash/)) is our most capable open-source foundation model, engineered to deliver frontier reasoning and agentic capabilities with exceptional efficiency. We also open-sourced the training codebase ([SteptronOss](https://github.com/stepfun-ai/SteptronOss)), with support for continue pretrain, SFT, RL (WIP), and evaluation (WIP), and will open-source the SFT data. Built on a sparse Mixture of Experts (MoE) architecture, it selectively activates only 11B of its 196B parameters per token. This "intelligence density" allows it to rival the reasoning depth of top-tier proprietary models, while maintaining the agility required for real-time interaction.
|
| 31 |
|
| 32 |
## 2. Key Capabilities
|
| 33 |
|