Add dataset card, link to paper and task category

#2
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +32 -0
README.md ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ task_categories:
3
+ - image-text-to-text
4
+ ---
5
+
6
+ # Global Perceptor Data
7
+
8
+ This dataset contains the training data for the **Global Perceptor** module of the **GeoFocus** framework, as presented in the paper [GeoFocus: Blending Efficient Global-to-Local Perception for Multimodal Geometry Problem-Solving](https://huggingface.co/papers/2602.08524).
9
+
10
+ [**Project Page**](https://github.com/dle666/GeoFocus) | [**Paper**](https://huggingface.co/papers/2602.08524) | [**Code**](https://github.com/dle666/GeoFocus)
11
+
12
+ ## Introduction
13
+ GeoFocus is a novel framework designed for multimodal geometry problem-solving. It addresses the challenges of geometry reasoning by combining two core modules:
14
+ 1. **Critical Local Perceptor**: Automatically identifies and emphasizes critical local structures (e.g., angles, parallel lines) through theory-based perception templates.
15
+ 2. **Global Perceptor**: Uses **VertexLang**, a compact topology formal language, to encode global figures through vertex coordinates and connectivity relations.
16
+
17
+ This repository provides the data necessary for the global perception training component of the model.
18
+
19
+ ## Related Datasets
20
+ - [Local_Perceptor_Data](https://huggingface.co/datasets/dle666/Local_Perceptor)
21
+ - [Geo_test](https://huggingface.co/datasets/dle666/GeoFocus-test)
22
+
23
+ ## Citation
24
+ If you find this dataset useful for your research, please cite the original paper:
25
+ ```bibtex
26
+ @article{geofocus2026,
27
+ title={GeoFocus: Blending Efficient Global-to-Local Perception for Multimodal Geometry Problem-Solving},
28
+ author={Anonymous Authors},
29
+ journal={arXiv preprint arXiv:2602.08524},
30
+ year={2026}
31
+ }
32
+ ```