Add model card for Solar Open 100B
Browse files
README.md
CHANGED
|
@@ -1,5 +1,60 @@
|
|
| 1 |
-
---
|
| 2 |
-
|
| 3 |
-
|
| 4 |
-
|
| 5 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
language:
|
| 3 |
+
- en
|
| 4 |
+
- ko
|
| 5 |
+
license: other
|
| 6 |
+
license_name: solar-apache-2.0
|
| 7 |
+
tags:
|
| 8 |
+
- upstage
|
| 9 |
+
- solar
|
| 10 |
+
- moe
|
| 11 |
+
- 100b
|
| 12 |
+
- llm
|
| 13 |
+
---
|
| 14 |
+
|
| 15 |
+
# **Solar Open**
|
| 16 |
+
|
| 17 |
+
**Solar Open** is Upstage's flagship large language model with **102B parameters**, now available under the **Solar-Apache License 2.0** (see LICENSE file). It is a **Mixture-of-Experts (MoE)** model designed to empower the open-source community with enterprise-grade reasoning, instruction-following, and agentic capabilities.
|
| 18 |
+
|
| 19 |
+
## Highlights
|
| 20 |
+
* **MoE Architecture (102B / 12B):** Built on a Mixture-of-Experts architecture with **102B total / 12B active parameters**. This design delivers the knowledge depth of a massive model with the inference speed and cost-efficiency of a much smaller model.
|
| 21 |
+
* **Agentic Specialist with Parallel Tool Calling:** Engineered to handle complex agentic workflows. It supports **Parallel Tool Calling**, allowing the model to generate multiple function calls in a single turn to execute tasks efficiently.
|
| 22 |
+
* **Massive Training Scale:** Pre-trained on **19.7 trillion tokens**, ensuring broad knowledge coverage and robust reasoning capabilities across various domains.
|
| 23 |
+
|
| 24 |
+
## Model Overview
|
| 25 |
+
* **Model Name:** Solar Open 100B
|
| 26 |
+
* **Hugging Face ID:** Upstage/Solar-Open-100B
|
| 27 |
+
* **Architecture:** Mixture-of-Experts (MoE)
|
| 28 |
+
* **Total Parameters:** 102.6B
|
| 29 |
+
* **Active Parameters:** 12B (per token)
|
| 30 |
+
* **Experts:** 129 Experts (top 8 among 128 Routed + 1 Shared)
|
| 31 |
+
* **Pre-training Tokens:** 19.7 Trillion
|
| 32 |
+
* **Context Length:** 128k
|
| 33 |
+
* **Training Hardware:** NVIDIA B200 GPUs
|
| 34 |
+
* **License:** **Solar-Apache License 2.0** (See `LICENSE` file)
|
| 35 |
+
|
| 36 |
+
## Performance
|
| 37 |
+
*Detailed benchmarks and performance metrics will be updated upon the official release on December 31, 2025.*
|
| 38 |
+
|
| 39 |
+
## Quickstart
|
| 40 |
+
*Python code snippets and usage examples will be available upon the official release on December 31, 2025.*
|
| 41 |
+
|
| 42 |
+
## Agentic Use & Parallel Tool Calling
|
| 43 |
+
Solar Open excels at **Parallel Tool Calling**, enabling the model to request multiple actions simultaneously within a single turn. This reduces latency and improves the efficiency of AI agents.
|
| 44 |
+
|
| 45 |
+
## Public API Access
|
| 46 |
+
The official API service for Solar Open is scheduled to launch publicly on **January 1, 2026**.
|
| 47 |
+
* **Access:** Upstage Console (Available starting Jan 1, 2026)
|
| 48 |
+
* **Documentation:** [**Upstage Console Docs**](https://console.upstage.ai/docs/getting-started)
|
| 49 |
+
|
| 50 |
+
## Citation
|
| 51 |
+
If you use Solar Open in your research, please cite:
|
| 52 |
+
```bibtex
|
| 53 |
+
@misc{solar-open-2025,
|
| 54 |
+
title={Solar Open: Scaling Upstage's LLM Capabilities with MoE},
|
| 55 |
+
author={Upstage AI},
|
| 56 |
+
year={2025},
|
| 57 |
+
url={https://huggingface.co/Upstage/Solar-Open-100B}
|
| 58 |
+
}
|
| 59 |
+
```
|
| 60 |
+
|