kcsayem commited on
Commit
b5bd302
·
verified ·
1 Parent(s): 65b8949

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +37 -1
README.md CHANGED
@@ -10,7 +10,43 @@ size_categories:
10
  ---
11
 
12
 
13
- # [CVPR 2026] HandVQA: Diagnosing and Improving Fine-Grained Spatial Reasoning about Hands in Vision-Language Models
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
 
15
  HandVQA is a large-scale diagnostic benchmark designed to evaluate fine-grained spatial reasoning in vision-language models (VLMs), focusing on articulated hand pose understanding.
16
 
 
10
  ---
11
 
12
 
13
+ <div align="center">
14
+
15
+ # HandVQA: Diagnosing and Improving Fine-Grained Spatial Reasoning about Hands in Vision-Language Models
16
+
17
+ <p align="center">
18
+ <strong>CVPR 2026</strong>
19
+ </p>
20
+
21
+ <p align="center">
22
+ MD Khalequzzaman Chowdhury Sayem<sup>1*</sup>,
23
+ Mubarrat Tajoar Chowdhury<sup>1*</sup>,
24
+ Yihalem Yimolal Tiruneh<sup>1</sup>,
25
+ Muneeb A. Khan<sup>1*</sup>,
26
+ Muhammad Salman Ali<sup>1*</sup>,
27
+ Binod Bhattarai<sup>2,3,4&dagger;</sup>,
28
+ Seungryul Baek<sup>1&dagger;</sup>
29
+ </p>
30
+
31
+ <p align="center">
32
+ <sup>1</sup>UNIST,
33
+ <sup>2</sup>University of Aberdeen,
34
+ <sup>3</sup>University College London,
35
+ <sup>4</sup>Fogsphere (Redev.AI Ltd)
36
+ </p>
37
+
38
+ <p align="center">
39
+ <sup>*</sup>Equal contribution.
40
+ <sup>&dagger;</sup>These authors jointly supervised this work.
41
+ </p>
42
+
43
+ <p align="center">
44
+ <a href="https://kcsayem.github.io/handvqa/"><img alt="Project Page" src="https://img.shields.io/badge/Project-Page-3b3b3b?style=for-the-badge"></a>
45
+ <a href="https://huggingface.co/datasets/kcsayem/handvqa"><img alt="Dataset" src="https://img.shields.io/badge/HuggingFace-Dataset-fcc624?style=for-the-badge"></a>
46
+ <a href="https://github.com/kcsayem/handvqa"><img alt="GitHub Code" src="https://img.shields.io/badge/GitHub-Code-2f3542?style=for-the-badge"></a>
47
+ </p>
48
+
49
+ </div>
50
 
51
  HandVQA is a large-scale diagnostic benchmark designed to evaluate fine-grained spatial reasoning in vision-language models (VLMs), focusing on articulated hand pose understanding.
52