Update model card and pipeline tag

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +33 -5
README.md CHANGED
@@ -1,14 +1,42 @@
1
  ---
2
  base_model:
3
  - Qwen/Qwen2.5-VL-3B-Instruct
4
- pipeline_tag: image-text-to-text
5
  library_name: transformers
 
6
  ---
7
 
8
- ## Bridging Semantics and Geometry: A Decoupled LVLM–SAM Framework for Reasoning Segmentation in Remote Sensing
9
 
10
- This is the 3B model of [Think2Seg-RS](https://github.com/Ricardo-XZ/Think2Seg-RS), a decoupled framework for reasoning segmentation in remote sensing (RS) imagery.
11
 
12
- Our core idea is to decouple high-level semantic reasoning from low-level geometric execution. Specifically, we train an LVLM prompter (e.g., Qwen-2.5-VL) to control a frozen Segment Anything Model (SAM2) via structured geometric prompts. Through a result-oriented reinforcement learning objective, the LVLM learns to translate abstract semantic reasoning into spatially grounded actions, achieving state-of-the-art performance on the EarthReason dataset.
13
 
14
- For more details, code, and the complete framework, please visit our [GitHub repository](https://github.com/Ricardo-XZ/Think2Seg-RS).
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  base_model:
3
  - Qwen/Qwen2.5-VL-3B-Instruct
 
4
  library_name: transformers
5
+ pipeline_tag: image-segmentation
6
  ---
7
 
8
+ # Think2Seg-RS-3B
9
 
10
+ This repository contains the 3B prompter model for **Think2Seg-RS**, a decoupled framework for reasoning segmentation in remote sensing (RS) imagery.
11
 
12
+ ## Overview
13
 
14
+ Think2Seg-RS addresses the limitations of coupling linguistic reasoning and pixel prediction in remote sensing analysis. The framework decouples high-level semantic reasoning from low-level geometric execution by training an LVLM prompter (based on Qwen2.5-VL) to control a frozen Segment Anything Model (SAM2) via structured geometric prompts.
15
+
16
+ Through a result-oriented reinforcement learning objective, the model learns to translate abstract semantic reasoning into spatially grounded actions, achieving state-of-the-art performance on the EarthReason dataset.
17
+
18
+ - **Paper:** [Bridging Semantics and Geometry: A Decoupled LVLM-SAM Framework for Reasoning Segmentation in Remote Sensing](https://huggingface.co/papers/2512.19302)
19
+ - **Repository:** [GitHub - Ricardo-XZ/Think2Seg-RS](https://github.com/Ricardo-XZ/Think2Seg-RS)
20
+ - **Base Model:** [Qwen2.5-VL-3B-Instruct](https://huggingface.co/Qwen/Qwen2.5-VL-3B-Instruct)
21
+
22
+ ## Key Features
23
+ - **Decoupled Architecture:** Separates high-level semantic reasoning from low-level geometric execution.
24
+ - **Geospatial Understanding:** Optimized for the complexities of remote sensing imagery and heterogeneous backgrounds.
25
+ - **Zero-shot Generalization:** The learned prompting policy generalizes effectively across multiple referring segmentation benchmarks.
26
+
27
+ ## Setup and Usage
28
+
29
+ For installation, training, and evaluation scripts, please visit the official [GitHub repository](https://github.com/Ricardo-XZ/Think2Seg-RS).
30
+
31
+ ## Citation
32
+
33
+ If you find this work helpful, please consider citing:
34
+
35
+ ```bibtex
36
+ @article{think2seg_rs_2025,
37
+ title={Bridging Semantics and Geometry: A Decoupled LVLM-SAM Framework for Reasoning Segmentation in Remote Sensing},
38
+ author={Luo, Junyu and Luo, Xiao and Chen, Xiusi and Xiao, Zhiping and Ju, Wei and Zhang, Ming},
39
+ journal={arXiv preprint arXiv:2512.19302},
40
+ year={2025}
41
+ }
42
+ ```