siyoungpark commited on
Commit
5751e6f
·
verified ·
1 Parent(s): 8efc246

Add model card for Solar Open 100B

Browse files
Files changed (1) hide show
  1. README.md +60 -5
README.md CHANGED
@@ -1,5 +1,60 @@
1
- ---
2
- license: other
3
- license_name: solar-apache-license
4
- license_link: LICENSE
5
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ - ko
5
+ license: other
6
+ license_name: solar-apache-2.0
7
+ tags:
8
+ - upstage
9
+ - solar
10
+ - moe
11
+ - 100b
12
+ - llm
13
+ ---
14
+
15
+ # **Solar Open**
16
+
17
+ **Solar Open** is Upstage's flagship large language model with **102B parameters**, now available under the **Solar-Apache License 2.0** (see LICENSE file). It is a **Mixture-of-Experts (MoE)** model designed to empower the open-source community with enterprise-grade reasoning, instruction-following, and agentic capabilities.
18
+
19
+ ## Highlights
20
+ * **MoE Architecture (102B / 12B):** Built on a Mixture-of-Experts architecture with **102B total / 12B active parameters**. This design delivers the knowledge depth of a massive model with the inference speed and cost-efficiency of a much smaller model.
21
+ * **Agentic Specialist with Parallel Tool Calling:** Engineered to handle complex agentic workflows. It supports **Parallel Tool Calling**, allowing the model to generate multiple function calls in a single turn to execute tasks efficiently.
22
+ * **Massive Training Scale:** Pre-trained on **19.7 trillion tokens**, ensuring broad knowledge coverage and robust reasoning capabilities across various domains.
23
+
24
+ ## Model Overview
25
+ * **Model Name:** Solar Open 100B
26
+ * **Hugging Face ID:** Upstage/Solar-Open-100B
27
+ * **Architecture:** Mixture-of-Experts (MoE)
28
+ * **Total Parameters:** 102.6B
29
+ * **Active Parameters:** 12B (per token)
30
+ * **Experts:** 129 Experts (top 8 among 128 Routed + 1 Shared)
31
+ * **Pre-training Tokens:** 19.7 Trillion
32
+ * **Context Length:** 128k
33
+ * **Training Hardware:** NVIDIA B200 GPUs
34
+ * **License:** **Solar-Apache License 2.0** (See `LICENSE` file)
35
+
36
+ ## Performance
37
+ *Detailed benchmarks and performance metrics will be updated upon the official release on December 31, 2025.*
38
+
39
+ ## Quickstart
40
+ *Python code snippets and usage examples will be available upon the official release on December 31, 2025.*
41
+
42
+ ## Agentic Use & Parallel Tool Calling
43
+ Solar Open excels at **Parallel Tool Calling**, enabling the model to request multiple actions simultaneously within a single turn. This reduces latency and improves the efficiency of AI agents.
44
+
45
+ ## Public API Access
46
+ The official API service for Solar Open is scheduled to launch publicly on **January 1, 2026**.
47
+ * **Access:** Upstage Console (Available starting Jan 1, 2026)
48
+ * **Documentation:** [**Upstage Console Docs**](https://console.upstage.ai/docs/getting-started)
49
+
50
+ ## Citation
51
+ If you use Solar Open in your research, please cite:
52
+ ```bibtex
53
+ @misc{solar-open-2025,
54
+ title={Solar Open: Scaling Upstage's LLM Capabilities with MoE},
55
+ author={Upstage AI},
56
+ year={2025},
57
+ url={https://huggingface.co/Upstage/Solar-Open-100B}
58
+ }
59
+ ```
60
+