Update README.md
Browse files
README.md
CHANGED
|
@@ -48,8 +48,6 @@ NexaQuant on Reasoning Benchmarks Compared to BF16 and LMStudio's Q4_K_M
|
|
| 48 |
<img src="https://cdn-uploads.huggingface.co/production/uploads/66abfd6f65beb23afa427d8a/Cyh1zVvDHNBT598IkLHkd.png" width="80%" alt="Example" />
|
| 49 |
</div>
|
| 50 |
|
| 51 |
-
The general capacity has also greatly improved:
|
| 52 |
-
|
| 53 |
**General Capacity:**
|
| 54 |
|
| 55 |
| Benchmark | Full 16-bit | llama.cpp (4-bit) | NexaQuant (4-bit)|
|
|
@@ -114,9 +112,9 @@ Get the latest version from the [official website](https://lmstudio.ai/).
|
|
| 114 |
|
| 115 |
## What's next
|
| 116 |
|
| 117 |
-
1.
|
| 118 |
|
| 119 |
-
2.
|
| 120 |
|
| 121 |
### Follow us
|
| 122 |
|
|
|
|
| 48 |
<img src="https://cdn-uploads.huggingface.co/production/uploads/66abfd6f65beb23afa427d8a/Cyh1zVvDHNBT598IkLHkd.png" width="80%" alt="Example" />
|
| 49 |
</div>
|
| 50 |
|
|
|
|
|
|
|
| 51 |
**General Capacity:**
|
| 52 |
|
| 53 |
| Benchmark | Full 16-bit | llama.cpp (4-bit) | NexaQuant (4-bit)|
|
|
|
|
| 112 |
|
| 113 |
## What's next
|
| 114 |
|
| 115 |
+
1. This model is built for complex problem-solving, which is why it sometimes takes a long thinking process even for simple questions. We recognize this and are working on improving it in the next update.
|
| 116 |
|
| 117 |
+
2. Inference Nexa Quantized Deepseek-R1 distilled model on NPU.
|
| 118 |
|
| 119 |
### Follow us
|
| 120 |
|