WinstonDeng commited on
Commit
5462fc4
·
verified ·
1 Parent(s): e8565d1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -1
README.md CHANGED
@@ -15,6 +15,7 @@ base_model:
15
 
16
  [![Hugging Face](https://img.shields.io/badge/%F0%9F%A4%97%20HF-StepFun/STEP3p5-preview)](https://huggingface.co/stepfun-ai/Step-3.5-Flash)
17
  [![ModelScope](https://img.shields.io/badge/ModelScope-StepFun/STEP3p5-preview)](https://huggingface.co/stepfun-ai/step3p5_preview/tree/main)
 
18
  [![Paper](https://img.shields.io/badge/Paper-Arxiv-red)](https://huggingface.co/stepfun-ai/Step-3.5-Flash)
19
  [![License](https://img.shields.io/badge/License-Apache%202.0-green)]()
20
 
@@ -22,7 +23,7 @@ base_model:
22
 
23
  ## 1. Introduction
24
 
25
- **Step 3.5 Flash** is our most capable open-source foundation model, engineered to deliver frontier reasoning and agentic capabilities with exceptional efficiency. Built on a sparse Mixture of Experts (MoE) architecture, it selectively activates only 11B of its 196B parameters per token. This "intelligence density" allows it to rival the reasoning depth of top-tier proprietary models, while maintaining the agility required for real-time interaction.
26
 
27
  ## 2. Key Capabilities
28
 
 
15
 
16
  [![Hugging Face](https://img.shields.io/badge/%F0%9F%A4%97%20HF-StepFun/STEP3p5-preview)](https://huggingface.co/stepfun-ai/Step-3.5-Flash)
17
  [![ModelScope](https://img.shields.io/badge/ModelScope-StepFun/STEP3p5-preview)](https://huggingface.co/stepfun-ai/step3p5_preview/tree/main)
18
+ [![Webpage](https://img.shields.io/badge/Webpage-Blog-blue)](https://static.stepfun.com/blog/step-3.5-flash/)
19
  [![Paper](https://img.shields.io/badge/Paper-Arxiv-red)](https://huggingface.co/stepfun-ai/Step-3.5-Flash)
20
  [![License](https://img.shields.io/badge/License-Apache%202.0-green)]()
21
 
 
23
 
24
  ## 1. Introduction
25
 
26
+ **Step 3.5 Flash** ([visit website](https://static.stepfun.com/blog/step-3.5-flash/)) is our most capable open-source foundation model, engineered to deliver frontier reasoning and agentic capabilities with exceptional efficiency. Built on a sparse Mixture of Experts (MoE) architecture, it selectively activates only 11B of its 196B parameters per token. This "intelligence density" allows it to rival the reasoning depth of top-tier proprietary models, while maintaining the agility required for real-time interaction.
27
 
28
  ## 2. Key Capabilities
29