shawn0wang commited on
Commit
cd73fb8
·
verified ·
1 Parent(s): e1c99eb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -33
README.md CHANGED
@@ -57,7 +57,6 @@ The model follows a connection pattern of Vision Encoder → MLP Adapter → Lan
57
  <th align="center"><strong>GPQA</strong></th>
58
  <th align="center"><strong>MathVista(mini)</strong></th>
59
  <th align="center"><strong>MMMU(Val)</strong></th>
60
- <th align="center"><strong>CSVQA</strong></th>
61
  </tr>
62
  <tr>
63
  <th></th>
@@ -67,7 +66,6 @@ The model follows a connection pattern of Vision Encoder → MLP Adapter → Lan
67
  <th align="center">pass@1</th>
68
  <th align="center">pass@1</th>
69
  <th align="center">pass@1</th>
70
- <th align="center">pass@1</th>
71
  </tr>
72
  </thead>
73
  <tbody>
@@ -79,7 +77,6 @@ The model follows a connection pattern of Vision Encoder → MLP Adapter → Lan
79
  <td align="center">49.0</td>
80
  <td align="center">-</td>
81
  <td align="center">-</td>
82
- <td align="center">-</td>
83
  </tr>
84
  <tr>
85
  <td>Deepseek V3</td>
@@ -89,7 +86,6 @@ The model follows a connection pattern of Vision Encoder → MLP Adapter → Lan
89
  <td align="center">59.1</td>
90
  <td align="center">-</td>
91
  <td align="center">-</td>
92
- <td align="center">-</td>
93
  </tr>
94
  <tr>
95
  <td>Deepseek R1</td>
@@ -99,7 +95,6 @@ The model follows a connection pattern of Vision Encoder → MLP Adapter → Lan
99
  <td align="center">71.5</td>
100
  <td align="center">-</td>
101
  <td align="center">-</td>
102
- <td align="center">-</td>
103
  </tr>
104
  <tr>
105
  <td>Claude 3.5 Sonnet</td>
@@ -109,7 +104,6 @@ The model follows a connection pattern of Vision Encoder → MLP Adapter → Lan
109
  <td align="center">65.0</td>
110
  <td align="center">67.7</td>
111
  <td align="center">68.3</td>
112
- <td align="center">-</td>
113
  </tr>
114
  <tr>
115
  <td>GPT-4o</td>
@@ -119,7 +113,6 @@ The model follows a connection pattern of Vision Encoder → MLP Adapter → Lan
119
  <td align="center">53.6</td>
120
  <td align="center">63.8</td>
121
  <td align="center">69.1</td>
122
- <td align="center">-</td>
123
  </tr>
124
  <tr>
125
  <td>Kimi k1.5</td>
@@ -129,7 +122,6 @@ The model follows a connection pattern of Vision Encoder → MLP Adapter → Lan
129
  <td align="center">-</td>
130
  <td align="center">74.9</td>
131
  <td align="center">70.0</td>
132
- <td align="center">-</td>
133
  </tr>
134
  <tr>
135
  <td>Qwen2.5-VL-72B-Instruct</td>
@@ -139,7 +131,6 @@ The model follows a connection pattern of Vision Encoder → MLP Adapter → Lan
139
  <td align="center">-</td>
140
  <td align="center">74.8</td>
141
  <td align="center">70.2</td>
142
- <td align="center">-</td>
143
  </tr>
144
  <tr>
145
  <td>LLaVA-Onevision-72B</td>
@@ -149,7 +140,6 @@ The model follows a connection pattern of Vision Encoder → MLP Adapter → Lan
149
  <td align="center">-</td>
150
  <td align="center">67.5</td>
151
  <td align="center">56.8</td>
152
- <td align="center">-</td>
153
  </tr>
154
  <tr>
155
  <td>InternVL2-Llama3-76B</td>
@@ -159,7 +149,6 @@ The model follows a connection pattern of Vision Encoder → MLP Adapter → Lan
159
  <td align="center">-</td>
160
  <td align="center">65.5</td>
161
  <td align="center">58.3</td>
162
- <td align="center">-</td>
163
  </tr>
164
  <tr>
165
  <td>InternVL2.5-78B</td>
@@ -169,7 +158,6 @@ The model follows a connection pattern of Vision Encoder → MLP Adapter → Lan
169
  <td align="center">-</td>
170
  <td align="center">72.3</td>
171
  <td align="center">70.1</td>
172
- <td align="center">-</td>
173
  </tr>
174
  <tr>
175
  <td>Skywork-R1V-38B</td>
@@ -179,7 +167,6 @@ The model follows a connection pattern of Vision Encoder → MLP Adapter → Lan
179
  <td align="center">61.6</td>
180
  <td align="center">67.5</td>
181
  <td align="center">69.0</td>
182
- <td align="center">-</td>
183
  </tr>
184
  </tbody>
185
  </table>
@@ -239,7 +226,7 @@ The model follows a connection pattern of Vision Encoder → MLP Adapter → Lan
239
  <td align="center">71.9</td>
240
  <td align="center">49.5</td>
241
  <td align="center">63.7</td>
242
- <td align="center">71.0</td>
243
  </tr>
244
  <tr>
245
  <td>MMMU(Val)</td>
@@ -247,15 +234,7 @@ The model follows a connection pattern of Vision Encoder → MLP Adapter → Lan
247
  <td align="center">63.9</td>
248
  <td align="center">55.1</td>
249
  <td align="center">55.2</td>
250
- <td align="center">68.1</td>
251
- </tr>
252
- <tr>
253
- <td>CSVQA</td>
254
- <td align="center">-</td>
255
- <td align="center"></td>
256
- <td align="center"></td>
257
- <td align="center"></td>
258
- <td align="center"></td>
259
  </tr>
260
  </tbody>
261
  </table>
@@ -295,20 +274,13 @@ python demo.py
295
 
296
  ---
297
 
298
- ## 6. Additional Resources
299
-
300
- - [📂 GitHub Repository](https://github.com/SkyworkAI/Skywork-R1V)
301
- - [🚀 Quick Start](#快速入门)
302
- - [📖 Full Documentation](#)
303
-
304
-
305
- ## 7. Citation
306
  If you use Skywork-R1V in your research, please cite:
307
 
308
  ```
309
  @article{skywork2025r1v,
310
- title = {Skywork-R1V: Bridging Vision and Language for Advanced Multimodal Reasoning},
311
- author = {SkyworkVL Team},
312
  year = {2025},
313
  journal = {arXiv preprint arXiv:XXXX.XXXXX},
314
  url = {https://github.com/skywork-ai/Skywork-R1V}
 
57
  <th align="center"><strong>GPQA</strong></th>
58
  <th align="center"><strong>MathVista(mini)</strong></th>
59
  <th align="center"><strong>MMMU(Val)</strong></th>
 
60
  </tr>
61
  <tr>
62
  <th></th>
 
66
  <th align="center">pass@1</th>
67
  <th align="center">pass@1</th>
68
  <th align="center">pass@1</th>
 
69
  </tr>
70
  </thead>
71
  <tbody>
 
77
  <td align="center">49.0</td>
78
  <td align="center">-</td>
79
  <td align="center">-</td>
 
80
  </tr>
81
  <tr>
82
  <td>Deepseek V3</td>
 
86
  <td align="center">59.1</td>
87
  <td align="center">-</td>
88
  <td align="center">-</td>
 
89
  </tr>
90
  <tr>
91
  <td>Deepseek R1</td>
 
95
  <td align="center">71.5</td>
96
  <td align="center">-</td>
97
  <td align="center">-</td>
 
98
  </tr>
99
  <tr>
100
  <td>Claude 3.5 Sonnet</td>
 
104
  <td align="center">65.0</td>
105
  <td align="center">67.7</td>
106
  <td align="center">68.3</td>
 
107
  </tr>
108
  <tr>
109
  <td>GPT-4o</td>
 
113
  <td align="center">53.6</td>
114
  <td align="center">63.8</td>
115
  <td align="center">69.1</td>
 
116
  </tr>
117
  <tr>
118
  <td>Kimi k1.5</td>
 
122
  <td align="center">-</td>
123
  <td align="center">74.9</td>
124
  <td align="center">70.0</td>
 
125
  </tr>
126
  <tr>
127
  <td>Qwen2.5-VL-72B-Instruct</td>
 
131
  <td align="center">-</td>
132
  <td align="center">74.8</td>
133
  <td align="center">70.2</td>
 
134
  </tr>
135
  <tr>
136
  <td>LLaVA-Onevision-72B</td>
 
140
  <td align="center">-</td>
141
  <td align="center">67.5</td>
142
  <td align="center">56.8</td>
 
143
  </tr>
144
  <tr>
145
  <td>InternVL2-Llama3-76B</td>
 
149
  <td align="center">-</td>
150
  <td align="center">65.5</td>
151
  <td align="center">58.3</td>
 
152
  </tr>
153
  <tr>
154
  <td>InternVL2.5-78B</td>
 
158
  <td align="center">-</td>
159
  <td align="center">72.3</td>
160
  <td align="center">70.1</td>
 
161
  </tr>
162
  <tr>
163
  <td>Skywork-R1V-38B</td>
 
167
  <td align="center">61.6</td>
168
  <td align="center">67.5</td>
169
  <td align="center">69.0</td>
 
170
  </tr>
171
  </tbody>
172
  </table>
 
226
  <td align="center">71.9</td>
227
  <td align="center">49.5</td>
228
  <td align="center">63.7</td>
229
+ <td align="center">67.5</td>
230
  </tr>
231
  <tr>
232
  <td>MMMU(Val)</td>
 
234
  <td align="center">63.9</td>
235
  <td align="center">55.1</td>
236
  <td align="center">55.2</td>
237
+ <td align="center">69.0</td>
 
 
 
 
 
 
 
 
238
  </tr>
239
  </tbody>
240
  </table>
 
274
 
275
  ---
276
 
277
+ ## 6. Citation
 
 
 
 
 
 
 
278
  If you use Skywork-R1V in your research, please cite:
279
 
280
  ```
281
  @article{skywork2025r1v,
282
+ title = {Skywork R1V: Bridging Vision and Language for Advanced Multimodal Reasoning},
283
+ author = {Yi Peng, Chris, Xiaokun Wang, Yichen Wei, Jiangbo Pei, Weijie Qiu, Ai Jian, Yunzhuo Hao, Jiachun Pan, Tianyidan Xie, Li Ge, Rongxian Zhuang, Xuchen Song, Yang Liu, Yahui Zhou},
284
  year = {2025},
285
  journal = {arXiv preprint arXiv:XXXX.XXXXX},
286
  url = {https://github.com/skywork-ai/Skywork-R1V}