siyoungpark commited on
Commit
a383611
·
verified ·
1 Parent(s): 4ad167d

Add Nota AI quantization model, remove Public API Access section

Browse files
Files changed (1) hide show
  1. README.md +1 -7
README.md CHANGED
@@ -29,6 +29,7 @@ arxiv: 2601.07022
29
 
30
  * **MoE Architecture (102B / 12B):** Built on a Mixture-of-Experts architecture with **102B total / 12B active parameters**. This design delivers the knowledge depth of a massive model with the inference speed and cost-efficiency of a much smaller model.
31
  * **Massive Training Scale:** Pre-trained on **19.7 trillion tokens**, ensuring broad knowledge coverage and robust reasoning capabilities across various domains.
 
32
 
33
  ## Model Overview
34
 
@@ -213,13 +214,6 @@ vllm serve upstage/Solar-Open-100B \
213
  --tensor-parallel-size 8
214
  ```
215
 
216
- ## Public API Access
217
-
218
- The official API service for Solar Open is scheduled to launch publicly in **January**.
219
-
220
- * **Access:** Upstage Console (TBA)
221
- * **Documentation:** Upstage Console (TBA)
222
-
223
  ## Citation
224
 
225
  If you use Solar Open in your research, please cite:
 
29
 
30
  * **MoE Architecture (102B / 12B):** Built on a Mixture-of-Experts architecture with **102B total / 12B active parameters**. This design delivers the knowledge depth of a massive model with the inference speed and cost-efficiency of a much smaller model.
31
  * **Massive Training Scale:** Pre-trained on **19.7 trillion tokens**, ensuring broad knowledge coverage and robust reasoning capabilities across various domains.
32
+ * **Quantized Version Available:** An official INT4 quantized model is provided by NotaAI and available at [`nota-ai/Solar-Open-100B-NotaMoEQuant-Int4`](https://huggingface.co/nota-ai/Solar-Open-100B-NotaMoEQuant-Int4).
33
 
34
  ## Model Overview
35
 
 
214
  --tensor-parallel-size 8
215
  ```
216
 
 
 
 
 
 
 
 
217
  ## Citation
218
 
219
  If you use Solar Open in your research, please cite: