add solar open technical report. add benchmark results.

#19
Files changed (3) hide show
  1. .gitattributes +1 -0
  2. README.md +44 -3
  3. solar-open-technical-report.pdf +3 -0
.gitattributes CHANGED
@@ -17,6 +17,7 @@
17
  *.ot filter=lfs diff=lfs merge=lfs -text
18
  *.parquet filter=lfs diff=lfs merge=lfs -text
19
  *.pb filter=lfs diff=lfs merge=lfs -text
 
20
  *.pickle filter=lfs diff=lfs merge=lfs -text
21
  *.pkl filter=lfs diff=lfs merge=lfs -text
22
  *.pt filter=lfs diff=lfs merge=lfs -text
 
17
  *.ot filter=lfs diff=lfs merge=lfs -text
18
  *.parquet filter=lfs diff=lfs merge=lfs -text
19
  *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pdf filter=lfs diff=lfs merge=lfs -text
21
  *.pickle filter=lfs diff=lfs merge=lfs -text
22
  *.pkl filter=lfs diff=lfs merge=lfs -text
23
  *.pt filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -28,7 +28,7 @@ tags:
28
  ## Model Overview
29
 
30
  * **Model Name:** Solar Open 100B
31
- * **Hugging Face ID:** Upstage/Solar-Open-100B
32
  * **Architecture:** Mixture-of-Experts (MoE)
33
  * **Total Parameters:** 102.6B
34
  * **Active Parameters:** 12B (per token)
@@ -36,10 +36,12 @@ tags:
36
  * **Pre-training Tokens:** 19.7 Trillion
37
  * **Context Length:** 128k
38
  * **Training Hardware:** NVIDIA B200 GPUs
39
- * **License:** **Solar-Apache License 2.0** (See [LICENSE](./LICENSE))
40
  * **Hardware Requirements:**
41
  * **Minimum:** 4x NVIDIA A100 (80GB)
42
 
 
 
43
  ## License
44
  This repository contains both model weights and code,
45
  which are licensed under different terms:
@@ -54,7 +56,46 @@ which are licensed under different terms:
54
 
55
  ## Performance
56
 
57
- TBA
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
58
 
59
  ## Inference Quickstart
60
 
 
28
  ## Model Overview
29
 
30
  * **Model Name:** Solar Open 100B
31
+ * **Hugging Face ID:** `Upstage/Solar-Open-100B`
32
  * **Architecture:** Mixture-of-Experts (MoE)
33
  * **Total Parameters:** 102.6B
34
  * **Active Parameters:** 12B (per token)
 
36
  * **Pre-training Tokens:** 19.7 Trillion
37
  * **Context Length:** 128k
38
  * **Training Hardware:** NVIDIA B200 GPUs
39
+ * **License:** **Solar-Apache License 2.0** (See [LICENSE](#license))
40
  * **Hardware Requirements:**
41
  * **Minimum:** 4x NVIDIA A100 (80GB)
42
 
43
+ For more details, please refer to [Solar Open Technical Report](solar-open-technical-report.pdf).
44
+
45
  ## License
46
  This repository contains both model weights and code,
47
  which are licensed under different terms:
 
56
 
57
  ## Performance
58
 
59
+ ### Korean Benchmarks
60
+
61
+ | Category | Benchmarks | Model Name (102B) | gpt-oss-120b (117B, high) | gpt-oss-120b (117B, medium) | GLM-4.5-Air (110B) |
62
+ | :--- | :--- | :---: | :---: | :---: | :---: |
63
+ | **General** | KMMLU | 73.0 | 72.7 | 70.3 | 70.2 |
64
+ | | KMMLU-Pro | 64.0 | 62.6 | 60.5 | 60.7 |
65
+ | | CLIcK | 78.9 | 77.2 | 72.9 | 48.3 |
66
+ | | HAE-RAE v1.1 | 73.3 | 70.8 | 69.6 | 42.6 |
67
+ | | KoBALT | 44.3 | 52.6 | 45.0 | 40.3 |
68
+ | **Finance** | KBankMMLU (in-house) | 65.5 | 62.5 | 61.5 | 64.7 |
69
+ | **Law** | KBL | 65.5 | 62.8 | 60.1 | 60.6 |
70
+ | **Medical** | KorMedMCQA | 84.4 | 75.8 | 76.3 | 80.5 |
71
+ | **Math** | Ko-AIME 2024 (in-house) | 80.3 | 90.0 | 76.7 | 80.0 |
72
+ | | Ko-AIME 2025 (in-house) | 80.0 | 90.0 | 70.0 | 83.3 |
73
+ | | HRM8K | 87.6 | 89.5 | 84.8 | 86.0 |
74
+ | **IF** | Ko-IFEval | 87.5 | 93.2 | 86.7 | 79.5 |
75
+ | **Preference** | Ko Arena Hard v2 (in-house) | 79.9 | 79.5 | 73.8 | 60.4 |
76
+
77
+
78
+ ### English Benchmarks
79
+
80
+ | Category | Benchmarks | Model Name (102B) | gpt-oss-120b (117B, high) | gpt-oss-120b (117B, medium) | GLM-4.5-Air (110B) |
81
+ | :--- | :--- | :---: | :---: | :---: | :---: |
82
+ | **General** | MMLU | 88.2 | 88.6 | 87.9 | 83.3 |
83
+ | | MMLU-Pro | 80.4 | 80.4 | 78.6 | 81.4 |
84
+ | | GPQA-Diamond | 68.1 | 78.0 | 69.4 | 75.8 |
85
+ | | HLE (text only) | 10.5 | 18.4 | 7.23 | 10.8 |
86
+ | **Math** | AIME 2024 | 91.7 | 94.3 | 77.7 | 88.7 |
87
+ | | AIME 2025 | 84.3 | 91.7 | 75.0 | 82.7 |
88
+ | | HMMT 2025 (Feb) | 73.3 | 80.0 | 63.3 | 66.7 |
89
+ | | HMMT 2025 (Nov) | 80.0 | 73.3 | 66.7 | 70.0 |
90
+ | **Code** | LiveCodeBench (v1–v6 cumul) | 74.2 | 89.9 | 82.8 | 71.9 |
91
+ | **IF** | IFBench | 53.7 | 70.8 | 61.2 | 37.8 |
92
+ | | IFEval | 88.0 | 91.4 | 86.5 | 86.5 |
93
+ | **Preference** | Arena Hard v2 | 74.8 | 79.6 | 72.7 | 62.5 |
94
+ | | Writing Bench | 7.51 | 6.61 | 6.55 | 7.40 |
95
+ | **Agent** | Tau² Airline | 52.4 | 56.0 | 52.8 | 60.8 |
96
+ | | Tau² Telecom | 55.6 | 57.7 | 47.4 | 28.1 |
97
+ | | Tau² Retail | 59.3 | 76.5 | 68.4 | 71.9 |
98
+ | **Long** | AA-LCR | 35.0 | 48.3 | 45.0 | 37.3 |
99
 
100
  ## Inference Quickstart
101
 
solar-open-technical-report.pdf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05d6664d644f12a4eff2deaa1d061e377aa11fce81e679c2837a8bfdecf509cd
3
+ size 366668