Dataset Viewer
Auto-converted to Parquet Duplicate
image
imagewidth (px)
240
640
text
stringclasses
68 values
mask
listlengths
180
640
image_id
stringlengths
3
6
ann_id
listlengths
1
160
img_height
int32
180
640
img_width
int32
240
640
bbox
listlengths
1
160
All vase
[[false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,fa(...TRUNCATED)
139
[ "139_19", "139_20", "139_21", "139_22", "139_29" ]
426
640
[ [ 549, 307, 37, 94 ], [ 351, 209, 11, 22 ], [ 241, 197, 12, 15 ], [ 338, 199, 8, 15 ], [ 167, 233, 19, 34 ] ]
All potted plant
[[false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,fa(...TRUNCATED)
632
[ "632_9", "632_11", "632_23" ]
483
640
[ [ 183, 137, 61, 75 ], [ 345, 212, 85, 109 ], [ 487, 89, 31, 18 ] ]
All teddy bear
[[false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,fa(...TRUNCATED)
776
[ "776_1", "776_2", "776_3" ]
640
428
[ [ 0, 53, 422, 497 ], [ 0, 270, 320, 369 ], [ 98, 0, 329, 428 ] ]
All skis
[[false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,fa(...TRUNCATED)
785
[ "785_4", "785_5" ]
425
640
[ [ 263, 360, 352, 40 ], [ 206, 338, 239, 44 ] ]
All person
[[false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,fa(...TRUNCATED)
872
[ "872_6", "872_7" ]
640
621
[ [ 150, 101, 285, 455 ], [ 164, 130, 261, 478 ] ]
All person
[[false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,fa(...TRUNCATED)
885
[ "885_3", "885_4", "885_5", "885_7", "885_8", "885_9", "885_10", "885_11" ]
427
640
[[277,187,139,210],[596,26,43,224],[419,0,48,13],[543,0,67,14],[499,0,76,14],[288,90,117,161],[188,0(...TRUNCATED)
All backpack
[[false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,fa(...TRUNCATED)
1000
[ "1000_16", "1000_19", "1000_20" ]
480
640
[ [ 42, 224, 56, 60 ], [ 257, 164, 10, 18 ], [ 197, 223, 70, 119 ] ]
All boat
[[false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,fa(...TRUNCATED)
1268
[ "1268_11", "1268_12", "1268_19" ]
427
640
[ [ 126, 125, 138, 17 ], [ 0, 130, 105, 15 ], [ 292, 121, 120, 22 ] ]
All person
[[false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,fa(...TRUNCATED)
1353
[ "1353_4", "1353_5", "1353_6", "1353_7", "1353_8", "1353_9", "1353_10" ]
500
375
[[247,234,24,49],[149,137,75,52],[164,198,103,185],[140,183,67,69],[193,184,28,57],[212,150,69,68],[(...TRUNCATED)
All mouse
[[false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,false,fa(...TRUNCATED)
1503
[ "1503_7", "1503_10" ]
240
320
[ [ 121, 178, 38, 22 ], [ 306, 154, 14, 10 ] ]
End of preview. Expand in Data Studio

Dr. Seg: Revisiting GRPO Training for Visual Large Language Models through Perception-Oriented Design

This repository contains the weights for Dr. Seg-7B, as presented in the paper Dr. Seg: Revisiting GRPO Training for Visual Large Language Models through Perception-Oriented Design.

Dr. Seg is a plug-and-play GRPO-based framework designed to adapt Visual Large Language Models (VLLMs) for visual perception tasks such as reasoning segmentation and object detection. It introduces two key components: a Look-to-Confirm mechanism and a Distribution-Ranked Reward module, requiring no architectural modifications and integrating seamlessly with existing GRPO-based VLLMs.

Links

Model Description

Dr. Seg-7B is fine-tuned from Qwen2.5-VL-7B-Instruct using perception-oriented designs. While standard GRPO is often tailored for language reasoning, Dr. Seg addresses the specific needs of visual perception by providing a broader output space and fine-grained, stable reward signals. Experiments demonstrate that Dr. Seg improves performance in complex visual scenarios while maintaining strong generalization.

Citation

If you find this work useful, please cite:

@article{sun2026dr,
  title={Dr. Seg: Revisiting GRPO Training for Visual Large Language Models through Perception-Oriented Design},
  author={Sun, Haoxiang and Wang, Tao and Tang, Chenwei and Yuan, Li and Lv, Jiancheng},
  journal={arXiv preprint arXiv:2603.00152},
  year={2026}
}

Acknowledgements

This project builds upon several open-source efforts, including VisionReasoner, Seg-Zero, EasyR1, veRL, and COCONut-PanCap. We also utilize pretrained models from Qwen2.5-VL and SAM2.

Downloads last month
14

Paper for hao05/coconut