jquenum commited on
Commit
9ceec46
·
verified ·
1 Parent(s): ff1cd5c

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +101 -3
README.md CHANGED
@@ -1,3 +1,101 @@
1
- ---
2
- license: cc-by-nc-sa-4.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-sa-4.0
3
+ ---
4
+
5
+ ## Summary
6
+ We introduce the Geospatial Reasoning Segmentation Dataset (GRES), a collection of vision and language data designed around remote-sensing applications. GRES consists of two core components: PreGRES, a dataset consisting of over 1M remote-sensing specific visual instruction-tuning Q/A pairs for pre-training geospatial models, and GRES, a semi-synthetic dataset specialized for reasoning segmentation of remote-sensing data and consisting of 9,205 images and 27,615 natural language queries/answers within those images. From this LISAt dataset, we generate train, test, and validation splits consisting of 7,205, 1,500, and 500 images respectively.
7
+
8
+ To generate synthetic data, we use the pipeline depicted below. We start with a seed detection dataset (xView). We then filter detections for those that are both visually interesting and highly distinguishable (A). For those detection, we then generate a natural language description (B), and a pixel-wise segmentation mask (C). Finally, the natural language description is used to generate a localization query (D). Together, the segmentation mask and the query form a ground-truth pair for the [LISAT](https://huggingface.co/jquenum/LISAt-7b) reasoning segmentation fine-tuning.
9
+
10
+ <p align="center">
11
+ <img src="https://huggingface.co/datasets/jquenum/GRES/resolve/main/gres.png" width="1024"/>
12
+
13
+ </p>
14
+
15
+ ## Usage
16
+
17
+ ```python
18
+ from datasets import load_dataset
19
+
20
+ # Define the dataset repo ID
21
+ repo_id = "jquenum/GRES"
22
+
23
+ # Download the dataset
24
+ dataset = load_dataset(repo_id)
25
+
26
+ # Show the dataset details
27
+ print("Dataset loaded successfully!")
28
+ print(dataset)
29
+
30
+ # Access specific splits like train, validation, and test
31
+ train_dataset = dataset['train']
32
+ val_dataset = dataset['val']
33
+ test_dataset = dataset['test']
34
+
35
+ # Print the first example from the train split
36
+ print("\nFirst example from the train dataset:", train_dataset[0])
37
+
38
+ ```
39
+
40
+ run ```./extract_gres_images.sh /path/to/xview_train_images /path/to/xView_train.geojson .``` to get the gres image pool.
41
+
42
+
43
+ ## LISAT GRES Dataset
44
+
45
+ This repository contains the LISAT GRES dataset, which includes image files and corresponding annotation files in JSON format. The dataset is organized into three main splits: **train**, **val**, and **test**.
46
+
47
+ ## Dataset Folder Structure
48
+
49
+ This GRES dataset includes image files and corresponding annotation files in JSON format. The dataset is organized into three main splits: **train**, **val**, and **test**.
50
+
51
+ - gres_images/
52
+ - train/
53
+ - lisat_gres_000000016192.jpg
54
+ - lisat_gres_000000016195.jpg
55
+ - lisat_gres_000000017340.jpg
56
+ - ...
57
+ - val/
58
+ - lisat_gres_000000016203.jpg
59
+ - lisat_gres_000000016210.jpg
60
+ - lisat_gres_000000017500.jpg
61
+ - ...
62
+ - test/
63
+ - lisat_gres_000000016217.jpg
64
+ - lisat_gres_000000016234.jpg
65
+ - lisat_gres_000000017800.jpg
66
+ - ...
67
+ - gres_annotations/
68
+ - train/
69
+ - lisat_gres_000000016192.json
70
+ - lisat_gres_000000016195.json
71
+ - lisat_gres_000000017340.json
72
+ - ...
73
+ - train.txt
74
+ - val/
75
+ - lisat_gres_000000016203.json
76
+ - lisat_gres_000000016210.json
77
+ - lisat_gres_000000017500.json
78
+ - ...
79
+ - val.txt
80
+ - test/
81
+ - lisat_gres_000000016217.json
82
+ - lisat_gres_000000016234.json
83
+ - lisat_gres_000000017800.json
84
+ - ...
85
+ - test.txt
86
+ - large.txt
87
+ - small.txt
88
+
89
+
90
+ ## Citation
91
+
92
+ If you use LISAt in your research or applications, please cite our paper:
93
+
94
+ ```bibtex
95
+ @article{TBD,
96
+ title={LISAt: Language-Instructed Segmentation Assistant for Satellite Imagery},
97
+ author={Quenum, Jerome and Hsieh, Wen-Han and Wu, Tsung-Han and Gupta, Ritwik and Darrell, Trevor and Chan, David M},
98
+ journal={TBD},
99
+ year={2025},
100
+ url={TBD}
101
+ }