sealical commited on
Commit
5cc4626
·
verified ·
1 Parent(s): 3002396

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +20 -87
README.md CHANGED
@@ -9,7 +9,7 @@ pipeline_tag: image-text-to-text
9
  <img src="skywork-logo.png" alt="Introduction Image" width="500" height="400">
10
  </div>
11
 
12
- ## 📖 [R1V2 Report](https://arxiv.org/abs/2504.16656) | 💻 [GitHub](https://github.com/SkyworkAI/Skywork-R1V) | 🌐 [ModelScope](https://modelscope.cn/models/Skywork/Skywork-R1V2-38B)
13
 
14
  <div align="center">
15
 
@@ -29,151 +29,72 @@ pipeline_tag: image-text-to-text
29
  <th>Model</th>
30
  <th align="center">MMMU</th>
31
  <th align="center">MathVista</th>
32
- <th align="center">MathVision</th>
33
- <th align="center">Olympiad Bench</th>
34
- <th align="center">AIME 24</th>
35
- <th align="center">LiveCode bench</th>
36
- <th align="center">Live Bench</th>
37
- <th align="center">IFEVAL</th>
38
  </tr>
39
  </thead>
40
  <tbody>
41
  <tr>
42
- <td colspan="9" align="center"><i>Proprietary Models</i></td>
43
  </tr>
44
  <tr>
45
  <td>Claude-3.5-Sonnet</td>
46
  <td align="center">70.4</td>
47
- <td align="center">67.7</td>
48
- <td align="center">-</td>
49
- <td align="center">-</td>
50
- <td align="center">-</td>
51
- <td align="center">-</td>
52
- <td align="center">-</td>
53
  <td align="center">-</td>
54
  </tr>
55
  <tr>
56
  <td>Gemini-2-Flash</td>
57
  <td align="center">70.7</td>
58
- <td align="center">73.1</td>
59
- <td align="center">41.3</td>
60
- <td align="center">-</td>
61
- <td align="center">-</td>
62
- <td align="center">-</td>
63
- <td align="center">-</td>
64
  <td align="center">-</td>
65
  </tr>
66
  <tr>
67
  <td>Kimi-k1.5-longcot</td>
68
  <td align="center">70.0</td>
69
- <td align="center">74.9</td>
70
- <td align="center">53.3</td>
71
- <td align="center">-</td>
72
- <td align="center">-</td>
73
- <td align="center">-</td>
74
- <td align="center">-</td>
75
  <td align="center">-</td>
76
  </tr>
77
  <tr>
78
  <td>OpenAI-o1</td>
79
  <td align="center">-</td>
80
  <td align="center">-</td>
81
- <td align="center">-</td>
82
- <td align="center">-</td>
83
- <td align="center">74.3</td>
84
- <td align="center">63.4</td>
85
- <td align="center">72.2</td>
86
- <td align="center">-</td>
87
  </tr>
88
  <tr>
89
  <td>OpenAI-o4-mini</td>
90
  <td align="center"><b>81.6</b></td>
91
- <td align="center"><b>84.3</b></td>
92
- <td align="center"><b>58.0</b></td>
93
- <td align="center">-</td>
94
- <td align="center"><b>93.4</b></td>
95
- <td align="center"><b>74.6</b></td>
96
- <td align="center"><b>78.1</b></td>
97
  <td align="center">-</td>
98
  </tr>
99
  <tr>
100
- <td colspan="9" align="center"><i>Open-Source Models</i></td>
101
  </tr>
102
  <tr>
103
  <td>Skywork-R1V1</td>
104
  <td align="center">68.0</td>
105
- <td align="center">67.0</td>
106
- <td align="center">-</td>
107
- <td align="center">-</td>
108
- <td align="center">72.0</td>
109
- <td align="center">57.2</td>
110
- <td align="center">54.6</td>
111
  <td align="center">72.5</td>
112
  </tr>
113
  <tr>
114
  <td>DeepseekR1-671B</td>
115
  <td align="center">-</td>
116
- <td align="center">-</td>
117
- <td align="center">-</td>
118
- <td align="center">-</td
119
- >
120
- <td align="center"><b>79.8</b></td>
121
- <td align="center"><b>65.9</b></td>
122
- <td align="center">71.6</td>
123
  <td align="center"><b>83.3</b></td>
124
  </tr>
125
  <tr>
126
  <td>InternVL3-38B</td>
127
  <td align="center">70.1</td>
128
- <td align="center">75.1</td>
129
- <td align="center">34.2</td>
130
- <td align="center">-</td>
131
- <td align="center">-</td>
132
- <td align="center">-</td>
133
- <td align="center">-</td>
134
  <td align="center">-</td>
135
  </tr>
136
  <tr>
137
  <td>Qwen2.5-VL-72B</td>
138
  <td align="center">70.2</td>
139
- <td align="center">74.8</td>
140
- <td align="center">38.1</td>
141
- <td align="center">40.4</td>
142
- <td align="center">-</td>
143
- <td align="center">-</td>
144
- <td align="center">-</td>
145
  <td align="center">-</td>
146
  </tr>
147
  <tr>
148
  <td>QvQ-Preview-72B</td>
149
  <td align="center">70.3</td>
150
- <td align="center">71.4</td>
151
- <td align="center">35.9</td>
152
- <td align="center">33.2</td>
153
- <td align="center">-</td>
154
- <td align="center">-</td>
155
- <td align="center">-</td>
156
  <td align="center">-</td>
157
  </tr>
158
  <tr>
159
- <td>Skywork-R1V2</td>
160
- <td align="center"><b>73.6</b></td>
161
- <td align="center">74.0</td>
162
- <td align="center"><b>49.0</b></td>
163
- <td align="center"><b>62.6</b></td>
164
- <td align="center">78.9</td>
165
- <td align="center">63.6</td>
166
  <td align="center"><b>73.2</b></td>
167
  <td align="center">82.9</td>
168
  </tr>
169
  <tr>
170
- <td>Skywork-R1V2-AWQ</td>
171
- <td align="center">64.4</td>
172
- <td align="center">64.8</td>
173
- <td align="center">42.9</td>
174
- <td align="center">54.8</td>
175
- <td align="center">77.3</td>
176
- <td align="center">55.7</td>
177
  <td align="center">64.1</td>
178
  <td align="center">72.5</td>
179
  </tr>
@@ -191,7 +112,7 @@ You can use the quantized model with different inference frameworks:
191
  import os
192
  from vllm import LLM, SamplingParams
193
  from vllm.entrypoints.chat_utils import load_chat_template
194
- model_name = "Skywork/Skywork-R1V2-38B-AWQ" # or local path
195
  llm = LLM(model_name,
196
  dtype='float16',
197
  quantization="awq",
@@ -205,7 +126,7 @@ llm = LLM(model_name,
205
  #### OpenAI-compatible API Server
206
 
207
  ```bash
208
- MODEL_ID="Skywork/Skywork-R1V2-38B-AWQ" # or local path
209
  CUDA_VISIBLE_DEVICES=0 \
210
  python -m vllm.entrypoints.openai.api_server \
211
  --model $MODEL_ID \
@@ -223,7 +144,7 @@ CUDA_VISIBLE_DEVICES=0 \
223
  import os
224
  from lmdeploy import pipeline, TurbomindEngineConfig, ChatTemplateConfig
225
  from lmdeploy.vl import load_image
226
- model_path = "Skywork/Skywork-R1V2-38B-AWQ" # or local path
227
  engine_config = TurbomindEngineConfig(cache_max_entry_count=0.75)
228
  chat_template_config = ChatTemplateConfig(model_name=model_path)
229
  pipe = pipeline(model_path,
@@ -270,3 +191,15 @@ If you use this model in your research, please cite:
270
  url={https://arxiv.org/abs/2504.16656},
271
  }
272
  ```
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  <img src="skywork-logo.png" alt="Introduction Image" width="500" height="400">
10
  </div>
11
 
12
+ ## 📖 [R1V3 Report](https://arxiv.org/abs/2507.06167) | 💻 [GitHub](https://github.com/SkyworkAI/Skywork-R1V) | 🌐 [ModelScope](https://modelscope.cn/models/Skywork/Skywork-R1V3-38B)
13
 
14
  <div align="center">
15
 
 
29
  <th>Model</th>
30
  <th align="center">MMMU</th>
31
  <th align="center">MathVista</th>
 
 
 
 
 
 
32
  </tr>
33
  </thead>
34
  <tbody>
35
  <tr>
36
+ <td colspan="3" align="center"><i>Proprietary Models</i></td>
37
  </tr>
38
  <tr>
39
  <td>Claude-3.5-Sonnet</td>
40
  <td align="center">70.4</td>
 
 
 
 
 
 
41
  <td align="center">-</td>
42
  </tr>
43
  <tr>
44
  <td>Gemini-2-Flash</td>
45
  <td align="center">70.7</td>
 
 
 
 
 
 
46
  <td align="center">-</td>
47
  </tr>
48
  <tr>
49
  <td>Kimi-k1.5-longcot</td>
50
  <td align="center">70.0</td>
 
 
 
 
 
 
51
  <td align="center">-</td>
52
  </tr>
53
  <tr>
54
  <td>OpenAI-o1</td>
55
  <td align="center">-</td>
56
  <td align="center">-</td>
 
 
 
 
 
 
57
  </tr>
58
  <tr>
59
  <td>OpenAI-o4-mini</td>
60
  <td align="center"><b>81.6</b></td>
 
 
 
 
 
 
61
  <td align="center">-</td>
62
  </tr>
63
  <tr>
64
+ <td colspan="3" align="center"><i>Open-Source Models</i></td>
65
  </tr>
66
  <tr>
67
  <td>Skywork-R1V1</td>
68
  <td align="center">68.0</td>
 
 
 
 
 
 
69
  <td align="center">72.5</td>
70
  </tr>
71
  <tr>
72
  <td>DeepseekR1-671B</td>
73
  <td align="center">-</td>
 
 
 
 
 
 
 
74
  <td align="center"><b>83.3</b></td>
75
  </tr>
76
  <tr>
77
  <td>InternVL3-38B</td>
78
  <td align="center">70.1</td>
 
 
 
 
 
 
79
  <td align="center">-</td>
80
  </tr>
81
  <tr>
82
  <td>Qwen2.5-VL-72B</td>
83
  <td align="center">70.2</td>
 
 
 
 
 
 
84
  <td align="center">-</td>
85
  </tr>
86
  <tr>
87
  <td>QvQ-Preview-72B</td>
88
  <td align="center">70.3</td>
 
 
 
 
 
 
89
  <td align="center">-</td>
90
  </tr>
91
  <tr>
92
+ <td>Skywork-R1V3</td>
 
 
 
 
 
 
93
  <td align="center"><b>73.2</b></td>
94
  <td align="center">82.9</td>
95
  </tr>
96
  <tr>
97
+ <td>Skywork-R1V3-AWQ</td>
 
 
 
 
 
 
98
  <td align="center">64.1</td>
99
  <td align="center">72.5</td>
100
  </tr>
 
112
  import os
113
  from vllm import LLM, SamplingParams
114
  from vllm.entrypoints.chat_utils import load_chat_template
115
+ model_name = "Skywork/Skywork-R1V3-38B-AWQ" # or local path
116
  llm = LLM(model_name,
117
  dtype='float16',
118
  quantization="awq",
 
126
  #### OpenAI-compatible API Server
127
 
128
  ```bash
129
+ MODEL_ID="Skywork/Skywork-R1V3-38B-AWQ" # or local path
130
  CUDA_VISIBLE_DEVICES=0 \
131
  python -m vllm.entrypoints.openai.api_server \
132
  --model $MODEL_ID \
 
144
  import os
145
  from lmdeploy import pipeline, TurbomindEngineConfig, ChatTemplateConfig
146
  from lmdeploy.vl import load_image
147
+ model_path = "Skywork/Skywork-R1V3-38B-AWQ" # or local path
148
  engine_config = TurbomindEngineConfig(cache_max_entry_count=0.75)
149
  chat_template_config = ChatTemplateConfig(model_name=model_path)
150
  pipe = pipeline(model_path,
 
191
  url={https://arxiv.org/abs/2504.16656},
192
  }
193
  ```
194
+
195
+ ```bibtex
196
+ @misc{shen2025skyworkr1v3technicalreport,
197
+ title={Skywork-R1V3 Technical Report},
198
+ author={Wei Shen and Jiangbo Pei and Yi Peng and Xuchen Song and Yang Liu and Jian Peng and Haofeng Sun and Yunzhuo Hao and Peiyu Wang and Jianhao Zhang and Yahui Zhou},
199
+ year={2025},
200
+ eprint={2507.06167},
201
+ archivePrefix={arXiv},
202
+ primaryClass={cs.CL},
203
+ url={https://arxiv.org/abs/2507.06167},
204
+ }
205
+ ```