Update README.md
Browse files
README.md
CHANGED
|
@@ -201,7 +201,8 @@ LFM2.5-1.2B-Instruct offers extremely fast inference speed on CPUs with a low me
|
|
| 201 |
|
| 202 |

|
| 203 |
|
| 204 |
-
In addition, we are partnering with AMD, Qualcomm, and Nexa AI to bring the LFM2.5 family to NPUs. These optimized models are available through our partners, enabling highly efficient on-device inference.
|
|
|
|
| 205 |
|
| 206 |
| Device | Inference | Framework | Model | Prefill (tok/s) | Decode (tok/s) | Memory (GB) |
|
| 207 |
| ---------------------------------------------------- | --------- | ---------------- | -------------------- | --------------- | -------------- | ----------- |
|
|
|
|
| 201 |
|
| 202 |

|
| 203 |
|
| 204 |
+
In addition, we are partnering with AMD, Qualcomm, and Nexa AI to bring the LFM2.5 family to NPUs. These optimized models are available through our partners, enabling highly efficient on-device inference.
|
| 205 |
+
The following numbers have been calculated using 1K prefill and 100 decode tokens:
|
| 206 |
|
| 207 |
| Device | Inference | Framework | Model | Prefill (tok/s) | Decode (tok/s) | Memory (GB) |
|
| 208 |
| ---------------------------------------------------- | --------- | ---------------- | -------------------- | --------------- | -------------- | ----------- |
|