Add Nota AI quantization model, remove Public API Access section
Browse files
README.md
CHANGED
|
@@ -29,6 +29,7 @@ arxiv: 2601.07022
|
|
| 29 |
|
| 30 |
* **MoE Architecture (102B / 12B):** Built on a Mixture-of-Experts architecture with **102B total / 12B active parameters**. This design delivers the knowledge depth of a massive model with the inference speed and cost-efficiency of a much smaller model.
|
| 31 |
* **Massive Training Scale:** Pre-trained on **19.7 trillion tokens**, ensuring broad knowledge coverage and robust reasoning capabilities across various domains.
|
|
|
|
| 32 |
|
| 33 |
## Model Overview
|
| 34 |
|
|
@@ -213,13 +214,6 @@ vllm serve upstage/Solar-Open-100B \
|
|
| 213 |
--tensor-parallel-size 8
|
| 214 |
```
|
| 215 |
|
| 216 |
-
## Public API Access
|
| 217 |
-
|
| 218 |
-
The official API service for Solar Open is scheduled to launch publicly in **January**.
|
| 219 |
-
|
| 220 |
-
* **Access:** Upstage Console (TBA)
|
| 221 |
-
* **Documentation:** Upstage Console (TBA)
|
| 222 |
-
|
| 223 |
## Citation
|
| 224 |
|
| 225 |
If you use Solar Open in your research, please cite:
|
|
|
|
| 29 |
|
| 30 |
* **MoE Architecture (102B / 12B):** Built on a Mixture-of-Experts architecture with **102B total / 12B active parameters**. This design delivers the knowledge depth of a massive model with the inference speed and cost-efficiency of a much smaller model.
|
| 31 |
* **Massive Training Scale:** Pre-trained on **19.7 trillion tokens**, ensuring broad knowledge coverage and robust reasoning capabilities across various domains.
|
| 32 |
+
* **Quantized Version Available:** An official INT4 quantized model is provided by NotaAI and available at [`nota-ai/Solar-Open-100B-NotaMoEQuant-Int4`](https://huggingface.co/nota-ai/Solar-Open-100B-NotaMoEQuant-Int4).
|
| 33 |
|
| 34 |
## Model Overview
|
| 35 |
|
|
|
|
| 214 |
--tensor-parallel-size 8
|
| 215 |
```
|
| 216 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 217 |
## Citation
|
| 218 |
|
| 219 |
If you use Solar Open in your research, please cite:
|