Confetti nielsr HF Staff commited on
Commit
83b4ae7
·
1 Parent(s): f501bbd

Add model card for GLAD (#1)

Browse files

- Add model card for GLAD (f746189a7521fe46baeaef613b0f2f07c0b90503)


Co-authored-by: Niels Rogge <nielsr@users.noreply.huggingface.co>

Files changed (1) hide show
  1. README.md +35 -0
README.md ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ pipeline_tag: object-detection
3
+ tags:
4
+ - vision-language-tracking
5
+ - diffusion-models
6
+ - visual-tracking
7
+ ---
8
+
9
+ # GLAD: Generative Language-Assisted Visual Tracking for Low-Semantic Templates
10
+
11
+ This repository contains the weights for **GLAD**, a vision-language tracking model introduced in the paper [GLAD: Generative Language-Assisted Visual Tracking for Low-Semantic Templates](https://huggingface.co/papers/2602.00570).
12
+
13
+ ## Overview
14
+
15
+ GLAD (Generative Language-AssisteD tracking) is a pioneering model that utilizes diffusion models for generative multi-modal fusion of text descriptions and template images.
16
+
17
+ Current vision-language trackers often struggle with "low-semantic" images (such as those with significant blur or low resolution) because traditional discriminative fusion paradigms have limited effectiveness in bridging the gap between text and degraded visual features. GLAD addresses this by leveraging the reconstruction capabilities of generative models to bolster compatibility between language and images, effectively enhancing the semantic information of the template for more robust tracking.
18
+
19
+ ## Resources
20
+
21
+ - **Paper:** [GLAD: Generative Language-Assisted Visual Tracking for Low-Semantic Templates](https://huggingface.co/papers/2602.00570)
22
+ - **GitHub Repository:** [https://github.com/Confetti-lxy/GLAD](https://github.com/Confetti-lxy/GLAD)
23
+
24
+ ## Citation
25
+
26
+ If you find this work useful in your research, please cite:
27
+
28
+ ```bibtex
29
+ @article{luo2026glad,
30
+ title={GLAD: Generative Language-Assisted Visual Tracking for Low-Semantic Templates},
31
+ author={Luo, Xingyu and Cai, Yidong and Liu, Jie and Tang, Jie and Wu, Gangshan and Wang, Limin},
32
+ journal={arXiv preprint arXiv:2602.00570},
33
+ year={2026}
34
+ }
35
+ ```