VoyagerX commited on
Commit
10a3ac0
·
1 Parent(s): 0eee4ec

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +116 -116
README.md CHANGED
@@ -1,116 +1,116 @@
1
- ---
2
- license: cc-by-sa-4.0
3
- task_categories:
4
- - text-to-image
5
- - image-to-image
6
- - other
7
- language:
8
- - en
9
- tags:
10
- - satellite-imagery
11
- - earth-observation
12
- - embeddings
13
- - geospatial
14
- - clip
15
- - majortom
16
- size_categories:
17
- - 10K<n<100K
18
- - 100K<n<1M
19
- ---
20
-
21
- <div style="display: flex; gap: 0.2em; align-items: center; justify-content: center;">
22
- <a href="https://www.modelscope.cn/studios/VoyagerX/EarthExplorer"><img src="https://img.shields.io/badge/Open in ModelScope.cn-xGPU-624aff"></a>
23
- <a href="https://www.modelscope.ai/studios/VoyagerX/EarthExplorer"><img src="https://img.shields.io/badge/Open in ModelScope.ai-CPU-624aff"></a>
24
- <a href="https://huggingface.co/spaces/ML4Sustain/EarthExplorer"><img src="https://img.shields.io/badge/Open in HF Space-CPU-FFD21E"></a>
25
- <a href="https://modelscope.cn/studios/VoyagerX/EarthExplorer/file/view/master/Tutorial.md?status=1"> <img src="https://img.shields.io/badge/Tutorial-📖-007bff"> </a>
26
- <a href="https://www.modelscope.cn/learn/3958"> <img src="https://img.shields.io/badge/中文教程-📖-007bff"> </a>
27
- </div>
28
-
29
- # EarthEmbeddings
30
-
31
- Satellite imagery embeddings dataset for the **EarthEmbeddingExplorer**, enabling natural language and location-based search of Earth observation data.
32
-
33
- ## Overview
34
-
35
- This repository contains pre-computed embeddings of satellite imagery using state-of-the-art vision-language models. These embeddings power the [EarthEmbeddingExplorer](https://huggingface.co/spaces/ML4Sustain/EarthExplorer) application, which allows users to search for satellite images using text queries, image uploads, or geographic locations.
36
-
37
- **Key features:**
38
- - Global satellite imagery from Sentinel-2 (MajorTOM Core-S2L2A)
39
- - Multiple embedding models optimized for Earth observation
40
- - Fast similarity search without raw image preprocessing
41
- - Ready-to-use Parquet format for efficient data access
42
-
43
- ## Dataset Description
44
-
45
- ### Data Source
46
- - **Base dataset**: MajorTOM Core-S2L2A (Sentinel-2 Level 2A, 2.2M+ samples)
47
- - **Processing**: Center crop (384×384 pixels) + uniform global sampling
48
-
49
- ### Embedding Models
50
-
51
- ### Embedding Models
52
-
53
- Four state-of-the-art vision models are used:
54
-
55
- | Model | Description | Training Data |
56
- | :--- | :--- | :--- |
57
- | **SigLIP** | General-purpose vision-language model | Web-scale natural image-text pairs |
58
- | **DINOv2** | Self-supervised vision transformer | Web-scale natural images (self-supervised) |
59
- | **FarSLIP** | Fine-grained satellite imagery model | Satellite image-text pairs |
60
- | **SatCLIP** | Location-based satellite model | Satellite image-location pairs |
61
-
62
- ## Dataset Splits
63
-
64
- ### 1. `uniform_sample_250k` ⚠️ Preview
65
- ```
66
- ├── uniform_sample_250k
67
- │ ├── dinov2
68
- │ │ ├── DINOv2_grid_sample_center_224x224_249k_MajorTOM.parquet
69
- │ │ └── DINOv2_grid_sample_center_384x384_244k.parquet
70
- │ ├── farslip
71
- │ │ └── FarSLIP_grid_sample_center_384x384_244k.parquet
72
- │ ├── satclip
73
- │ │ └── SatCLIP_grid_sample_center_384x384_244k.parquet
74
- │ └── siglip
75
- │ └── SigLIP_grid_sample_center_384x384_244k.parquet
76
- ```
77
-
78
- - **~250,000** globally distributed satellite images
79
- - **Current status**: Preview revision with ~244k pre-computed embeddings and ~249k embeddings sampled from [Major-TOM/Core-S2RGB-DINOv2](https://huggingface.co/datasets/Major-TOM/Core-S2RGB-DINOv2) available
80
- - **Note**: About 4-6k original image chips were lost due to network error; full version coming soon
81
- - **Crop size**: For the 1/9 sampled grids, we crop the central bbox in each grid. To ensure the image patches are the same for each model, we chose crop size of 384x384, for pre-computed embeddings, we chose the crop size at 384x384. So these embeddings could represent the same regions on Earth surface.
82
-
83
- | Filename | Embedding Model | Crop Size | Model Input Size | Embedding Dim | Source |
84
- |----------|-----------------|-----------|------------------|---------------|--------|
85
- | `DINOv2_grid_sample_center_224x224_249k_MajorTOM.parquet` | [DINOv2-large](https://huggingface.co/facebook/dinov2-large) | 224×224 | 224×224 | 1024 | [Major-TOM/Core-S2RGB-DINOv2](https://huggingface.co/datasets/Major-TOM/Core-S2RGB-DINOv2) |
86
- | `DINOv2_grid_sample_center_384x384_244k.parquet` | [DINOv2-large](https://huggingface.co/facebook/dinov2-large) | 384×384 | 224×224 | 1024 | Pre-computed |
87
- | `FarSLIP_grid_sample_center_384x384_244k.parquet` | [FarSLIP-ViT-B-16](https://huggingface.co/ZhenShiL/FarSLIP) | 384×384 | 224×224 | 512 | Pre-computed |
88
- | `SatCLIP_grid_sample_center_384x384_244k.parquet` | [SatCLIP-ViT16-L40](https://github.com/microsoft/satclip) | 384×384 | 224×224 | 256 | Pre-computed |
89
- | `SigLIP_grid_sample_center_384x384_244k.parquet` | [SigLIP-SO400M-14](https://huggingface.co/timm/ViT-SO400M-14-SigLIP-384) | 384×384 | 384×384 | 1152 | Pre-computed |
90
-
91
-
92
- ### 2. `uniform_sample_22k`
93
- - **22,000** globally distributed satellite images
94
- - **Files**: `grid_sample_center_22k_{FarSLIP,SatCLIP,SigLIP}_384x384.parquet`
95
-
96
- ### 3. `Zhejiang_samples`
97
- - **2,000** samples from Zhejiang region, China
98
- - **Files**: `zhejiang_sample_center_2k_{FarSLIP,SatCLIP,SigLIP}_384x384.parquet`
99
- - Regional case study dataset
100
-
101
- ## Data Format
102
-
103
- All embeddings are stored in **Parquet** format:
104
- - Efficient columnar storage for fast download
105
- - 384×384 pixel satellite image crops
106
-
107
-
108
- ## Related Work
109
-
110
- - **Tutorial**: [EarthEmbeddingExplorer Tutorial](https://huggingface.co/spaces/ML4Sustain/EarthExplorer/blob/main/Tutorial.md)
111
- - **Application**: [EarthEmbeddingExplorer Space](https://huggingface.co/spaces/ML4Sustain/EarthExplorer)
112
- - **Base Dataset**: [MajorTOM by ESA](https://github.com/ESA-PhiLab/MajorTOM)
113
-
114
- ## License
115
-
116
- CC-BY-SA-4.0
 
1
+ ---
2
+ license: cc-by-sa-4.0
3
+ task_categories:
4
+ - text-to-image
5
+ - image-to-image
6
+ - other
7
+ language:
8
+ - en
9
+ tags:
10
+ - satellite-imagery
11
+ - earth-observation
12
+ - embeddings
13
+ - geospatial
14
+ - clip
15
+ - majortom
16
+ size_categories:
17
+ - 10K<n<100K
18
+ - 100K<n<1M
19
+ ---
20
+
21
+ <div style="display: flex; gap: 0.2em; align-items: center; justify-content: center;">
22
+ <a href="https://www.modelscope.cn/studios/VoyagerX/EarthExplorer"><img src="https://img.shields.io/badge/Open in ModelScope.cn-xGPU-624aff"></a>
23
+ <a href="https://www.modelscope.ai/studios/VoyagerX/EarthExplorer"><img src="https://img.shields.io/badge/Open in ModelScope.ai-CPU-624aff"></a>
24
+ <a href="https://huggingface.co/spaces/ML4Sustain/EarthExplorer"><img src="https://img.shields.io/badge/Open in HF Space-CPU-FFD21E"></a>
25
+ <a href="https://modelscope.cn/studios/VoyagerX/EarthExplorer/file/view/master/Tutorial.md?status=1"> <img src="https://img.shields.io/badge/Tutorial-📖-007bff"> </a>
26
+ <a href="https://www.modelscope.cn/learn/3958"> <img src="https://img.shields.io/badge/中文教程-📖-007bff"> </a>
27
+ </div>
28
+
29
+ # EarthEmbeddings
30
+
31
+ Satellite imagery embeddings dataset for the **EarthEmbeddingExplorer**, enabling natural language and location-based search of Earth observation data.
32
+
33
+ ## Overview
34
+
35
+ This repository contains pre-computed embeddings of satellite imagery using state-of-the-art vision-language models. These embeddings power the [EarthEmbeddingExplorer](https://huggingface.co/spaces/ML4Sustain/EarthExplorer) application, which allows users to search for satellite images using text queries, image uploads, or geographic locations.
36
+
37
+ **Key features:**
38
+ - Global satellite imagery from Sentinel-2 (MajorTOM Core-S2L2A)
39
+ - Multiple embedding models optimized for Earth observation
40
+ - Fast similarity search without raw image preprocessing
41
+ - Ready-to-use Parquet format for efficient data access
42
+
43
+ ## Dataset Description
44
+
45
+ ### Data Source
46
+ - **Base dataset**: MajorTOM Core-S2L2A (Sentinel-2 Level 2A, 2.2M+ samples)
47
+ - **Processing**: Center crop (384×384 pixels) + uniform global sampling
48
+
49
+ ### Embedding Models
50
+
51
+ ### Embedding Models
52
+
53
+ Four state-of-the-art vision models are used:
54
+
55
+ | Model | Description | Training Data |
56
+ | :--- | :--- | :--- |
57
+ | **SigLIP** | General-purpose vision-language model | Web-scale natural image-text pairs |
58
+ | **DINOv2** | Self-supervised vision transformer | Web-scale natural images (self-supervised) |
59
+ | **FarSLIP** | Fine-grained satellite imagery model | Satellite image-text pairs |
60
+ | **SatCLIP** | Location-based satellite model | Satellite image-location pairs |
61
+
62
+ ## Dataset Splits
63
+
64
+ ### 1. `uniform_sample_250k` ⚠️ Preview
65
+ ```
66
+ ├── uniform_sample_250k
67
+ │ ├── dinov2
68
+ │ │ ├── DINOv2_grid_sample_center_224x224_249k_MajorTOM.parquet
69
+ │ │ └── DINOv2_grid_sample_center_384x384_244k.parquet
70
+ │ ├── farslip
71
+ │ │ └── FarSLIP_grid_sample_center_384x384_244k.parquet
72
+ │ ├── satclip
73
+ │ │ └── SatCLIP_grid_sample_center_384x384_244k.parquet
74
+ │ └── siglip
75
+ │ └── SigLIP_grid_sample_center_384x384_244k.parquet
76
+ ```
77
+
78
+ - **~250,000** globally distributed satellite images
79
+ - **Current status**: Preview revision with ~244k pre-computed embeddings and ~249k embeddings sampled from [Major-TOM/Core-S2RGB-DINOv2](https://huggingface.co/datasets/Major-TOM/Core-S2RGB-DINOv2) available
80
+ - **Note**: About 4-6k original image chips were lost due to network error; full version coming soon
81
+ - **Crop size**: For the 1/9 sampled grids, we crop the central bbox in each grid. To ensure the image patches are the same for each model, we chose crop size of 384x384, for pre-computed embeddings, we chose the crop size at 384x384. So these embeddings could represent the same regions on Earth surface.
82
+
83
+ | Filename | Embedding Model | Crop Size | Model Input Size | Embedding Dim | Source |
84
+ |----------|-----------------|-----------|------------------|---------------|--------|
85
+ | `DINOv2_grid_sample_center_224x224_249k_MajorTOM.parquet` | [DINOv2-large](https://huggingface.co/facebook/dinov2-large) | 224×224 | 224×224 | 1024 | [Major-TOM/Core-S2RGB-DINOv2](https://huggingface.co/datasets/Major-TOM/Core-S2RGB-DINOv2) |
86
+ | `DINOv2_grid_sample_center_384x384_244k.parquet` | [DINOv2-large](https://huggingface.co/facebook/dinov2-large) | 384×384 | 224×224 | 1024 | Pre-computed |
87
+ | `FarSLIP_grid_sample_center_384x384_244k.parquet` | [FarSLIP-ViT-B-16](https://huggingface.co/ZhenShiL/FarSLIP) | 384×384 | 224×224 | 512 | Pre-computed |
88
+ | `SatCLIP_grid_sample_center_384x384_244k.parquet` | [SatCLIP-ViT16-L40](https://github.com/microsoft/satclip) | 384×384 | 224×224 | 256 | Pre-computed |
89
+ | `SigLIP_grid_sample_center_384x384_244k.parquet` | [SigLIP-SO400M-14](https://huggingface.co/timm/ViT-SO400M-14-SigLIP-384) | 384×384 | 384×384 | 1152 | Pre-computed |
90
+
91
+
92
+ ### 2. `uniform_sample_22k`
93
+ - **22,000** globally distributed satellite images
94
+ - **Files**: `grid_sample_center_22k_{FarSLIP,SatCLIP,SigLIP}_384x384.parquet`
95
+
96
+ ### 3. `Zhejiang_samples`
97
+ - **2,000** samples from Zhejiang region, China
98
+ - **Files**: `zhejiang_sample_center_2k_{FarSLIP,SatCLIP,SigLIP}_384x384.parquet`
99
+ - Regional case study dataset
100
+
101
+ ## Data Format
102
+
103
+ All embeddings are stored in **Parquet** format:
104
+ - Efficient columnar storage for fast download
105
+ - 384×384 pixel satellite image crops
106
+
107
+
108
+ ## Related Work
109
+
110
+ - **Tutorial**: [EarthEmbeddingExplorer Tutorial](https://huggingface.co/spaces/ML4Sustain/EarthExplorer/blob/main/Tutorial.md)
111
+ - **Application**: [EarthEmbeddingExplorer Space](https://huggingface.co/spaces/ML4Sustain/EarthExplorer)
112
+ - **Base Dataset**: [MajorTOM by ESA](https://github.com/ESA-PhiLab/MajorTOM)
113
+
114
+ ## License
115
+
116
+ CC-BY-SA-4.0