mlabonne commited on
Commit
8dc4e68
·
verified ·
1 Parent(s): 47c527b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -1
README.md CHANGED
@@ -201,7 +201,8 @@ LFM2.5-1.2B-Instruct offers extremely fast inference speed on CPUs with a low me
201
 
202
  ![image](https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/dbbI-15p9re2ROhAkqnZm.png)
203
 
204
- In addition, we are partnering with AMD, Qualcomm, and Nexa AI to bring the LFM2.5 family to NPUs. These optimized models are available through our partners, enabling highly efficient on-device inference.
 
205
 
206
  | Device | Inference | Framework | Model | Prefill (tok/s) | Decode (tok/s) | Memory (GB) |
207
  | ---------------------------------------------------- | --------- | ---------------- | -------------------- | --------------- | -------------- | ----------- |
 
201
 
202
  ![image](https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/dbbI-15p9re2ROhAkqnZm.png)
203
 
204
+ In addition, we are partnering with AMD, Qualcomm, and Nexa AI to bring the LFM2.5 family to NPUs. These optimized models are available through our partners, enabling highly efficient on-device inference.
205
+ The following numbers have been calculated using 1K prefill and 100 decode tokens:
206
 
207
  | Device | Inference | Framework | Model | Prefill (tok/s) | Decode (tok/s) | Memory (GB) |
208
  | ---------------------------------------------------- | --------- | ---------------- | -------------------- | --------------- | -------------- | ----------- |