loubnabnl HF Staff commited on
Commit
0ff8759
·
verified ·
1 Parent(s): 951e73c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -3
README.md CHANGED
@@ -31,11 +31,10 @@ language:
31
 
32
  SmolLM3 is a 3B parameter language model designed to push the boundaries of small models. It supports 6 languages, advanced reasoning and long context. SmolLM3 is a fully open model that offers strong performance at the 3B–4B scale.
33
 
34
- The model is a decoder-only transformer using GQA and NoRope, it was trained on 11.2T tokens with a staged curriculum of web, code, math and reasoning data. Post-training included midtraining on 100B reasoning followed by supervised fine-tuning and alignment via Anchored Preference Optimization.
35
-
36
-
37
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/61c141342aac764ce1654e43/Zcm_016pWeyFr_uIkT7Ki.png)
38
 
 
 
39
  ### Key features
40
  - Instruct model optimized for **hybrid reasoning**
41
  - **Fully open model**: open weights + full training details including public data mixture and training configs
 
31
 
32
  SmolLM3 is a 3B parameter language model designed to push the boundaries of small models. It supports 6 languages, advanced reasoning and long context. SmolLM3 is a fully open model that offers strong performance at the 3B–4B scale.
33
 
 
 
 
34
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/61c141342aac764ce1654e43/Zcm_016pWeyFr_uIkT7Ki.png)
35
 
36
+ The model is a decoder-only transformer using GQA and NoRope, it was trained on 11.2T tokens with a staged curriculum of web, code, math and reasoning data. Post-training included midtraining on 100B reasoning followed by supervised fine-tuning and alignment via Anchored Preference Optimization.
37
+
38
  ### Key features
39
  - Instruct model optimized for **hybrid reasoning**
40
  - **Fully open model**: open weights + full training details including public data mixture and training configs