image imagewidth (px) 240 640 | text stringclasses 68 values | mask listlengths 180 640 | image_id stringlengths 3 6 | ann_id listlengths 1 160 | img_height int32 180 640 | img_width int32 240 640 | bbox listlengths 1 160 |
|---|---|---|---|---|---|---|---|
All vase | [[false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,fa(...TRUNCATED) | 139 | [
"139_19",
"139_20",
"139_21",
"139_22",
"139_29"
] | 426 | 640 | [
[
549,
307,
37,
94
],
[
351,
209,
11,
22
],
[
241,
197,
12,
15
],
[
338,
199,
8,
15
],
[
167,
233,
19,
34
]
] | |
All potted plant | [[false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,fa(...TRUNCATED) | 632 | [
"632_9",
"632_11",
"632_23"
] | 483 | 640 | [
[
183,
137,
61,
75
],
[
345,
212,
85,
109
],
[
487,
89,
31,
18
]
] | |
All teddy bear | [[false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,fa(...TRUNCATED) | 776 | [
"776_1",
"776_2",
"776_3"
] | 640 | 428 | [
[
0,
53,
422,
497
],
[
0,
270,
320,
369
],
[
98,
0,
329,
428
]
] | |
All skis | [[false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,fa(...TRUNCATED) | 785 | [
"785_4",
"785_5"
] | 425 | 640 | [
[
263,
360,
352,
40
],
[
206,
338,
239,
44
]
] | |
All person | [[false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,fa(...TRUNCATED) | 872 | [
"872_6",
"872_7"
] | 640 | 621 | [
[
150,
101,
285,
455
],
[
164,
130,
261,
478
]
] | |
All person | [[false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,fa(...TRUNCATED) | 885 | [
"885_3",
"885_4",
"885_5",
"885_7",
"885_8",
"885_9",
"885_10",
"885_11"
] | 427 | 640 | [[277,187,139,210],[596,26,43,224],[419,0,48,13],[543,0,67,14],[499,0,76,14],[288,90,117,161],[188,0(...TRUNCATED) | |
All backpack | [[false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,fa(...TRUNCATED) | 1000 | [
"1000_16",
"1000_19",
"1000_20"
] | 480 | 640 | [
[
42,
224,
56,
60
],
[
257,
164,
10,
18
],
[
197,
223,
70,
119
]
] | |
All boat | [[false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,fa(...TRUNCATED) | 1268 | [
"1268_11",
"1268_12",
"1268_19"
] | 427 | 640 | [
[
126,
125,
138,
17
],
[
0,
130,
105,
15
],
[
292,
121,
120,
22
]
] | |
All person | [[false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,fa(...TRUNCATED) | 1353 | [
"1353_4",
"1353_5",
"1353_6",
"1353_7",
"1353_8",
"1353_9",
"1353_10"
] | 500 | 375 | [[247,234,24,49],[149,137,75,52],[164,198,103,185],[140,183,67,69],[193,184,28,57],[212,150,69,68],[(...TRUNCATED) | |
All mouse | [[false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,fa(...TRUNCATED) | 1503 | [
"1503_7",
"1503_10"
] | 240 | 320 | [
[
121,
178,
38,
22
],
[
306,
154,
14,
10
]
] |
Dr. Seg: Revisiting GRPO Training for Visual Large Language Models through Perception-Oriented Design
This repository contains the weights for Dr. Seg-7B, as presented in the paper Dr. Seg: Revisiting GRPO Training for Visual Large Language Models through Perception-Oriented Design.
Dr. Seg is a plug-and-play GRPO-based framework designed to adapt Visual Large Language Models (VLLMs) for visual perception tasks such as reasoning segmentation and object detection. It introduces two key components: a Look-to-Confirm mechanism and a Distribution-Ranked Reward module, requiring no architectural modifications and integrating seamlessly with existing GRPO-based VLLMs.
Links
- Paper: arXiv:2603.00152
- Dataset: COCONut
- Code: GitHub Repository
Model Description
Dr. Seg-7B is fine-tuned from Qwen2.5-VL-7B-Instruct using perception-oriented designs. While standard GRPO is often tailored for language reasoning, Dr. Seg addresses the specific needs of visual perception by providing a broader output space and fine-grained, stable reward signals. Experiments demonstrate that Dr. Seg improves performance in complex visual scenarios while maintaining strong generalization.
Citation
If you find this work useful, please cite:
@article{sun2026dr,
title={Dr. Seg: Revisiting GRPO Training for Visual Large Language Models through Perception-Oriented Design},
author={Sun, Haoxiang and Wang, Tao and Tang, Chenwei and Yuan, Li and Lv, Jiancheng},
journal={arXiv preprint arXiv:2603.00152},
year={2026}
}
Acknowledgements
This project builds upon several open-source efforts, including VisionReasoner, Seg-Zero, EasyR1, veRL, and COCONut-PanCap. We also utilize pretrained models from Qwen2.5-VL and SAM2.
- Downloads last month
- 14