zhou777 commited on
Commit
e125dce
ยท
verified ยท
1 Parent(s): c9a6f9e

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +70 -0
README.md ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ - zh
6
+ pipeline_tag: image-text-to-text
7
+ tags:
8
+ - remote-sensing
9
+ - geospatial-reasoning
10
+ - qwen2.5-vl
11
+ - sft
12
+ - chain-of-thought
13
+ - ms-swift
14
+ ---
15
+
16
+ # LandAI-Base: Activating Geospatial Chain-of-Thought Reasoning
17
+
18
+ <div align="center">
19
+
20
+ **[Paper (Under Review)]**
21
+
22
+ </div>
23
+
24
+ ## ๐Ÿ“– Introduction
25
+
26
+ **LandAI-Base** is the foundational **Supervised Fine-Tuned (SFT)** model of the LandAI family. Built upon the **Qwen2.5-VL-7B-Instruct** architecture, it is specifically designed to activate domain-specific logical reasoning in Earth Observation tasks.
27
+
28
+ Unlike general-purpose multimodal models, LandAI-Base has been fine-tuned on a composite corpus of approximately **334,000 reasoning chains**, including the novel **Geo-Base-Thinking-14K** dataset. This process instills the model with the "epistemological authority" of geography experts, enabling it to decompose complex spatial problems before engaging in visual recognition.
29
+
30
+ **LandAI-Base serves two primary purposes:**
31
+ 1. **A robust baseline** for geospatial reasoning tasks (Q&A, analysis).
32
+ 2. **The "Cold Start" initialization** for the [LandAI-L1](https://huggingface.co/zhou777/LandAI-L1) model (trained via GRPO-L1).
33
+
34
+ ## ๐Ÿš€ Key Features
35
+
36
+ * **Domain-Specific Cognitive Activation**: Fine-tuned to simulate the reasoning patterns of geography experts, moving from rote memorization to logical deduction.
37
+ * **High-Quality Training Data**: Trained on a curated mix of:
38
+ * **Geo-Base-Thinking-14K**: ~14.7k distillations from geography entrance exams and textbooks.
39
+ * **General Reasoning Corpus**: Subsets from OpenR1-Math, OpenThoughts, and Chinese-Data-R1 to enhance mathematical and scientific logic.
40
+ * **Strong Zero-Shot Performance**: Significantly outperforms the vanilla Qwen2.5-VL-7B on geographic benchmark exams.
41
+ * **MS-Swift Compatibility**: Fully compatible with the [ms-swift](https://github.com/modelscope/swift) training framework.
42
+
43
+ ## ๐Ÿ“Š Performance Benchmarks
44
+
45
+ LandAI-Base demonstrates a substantial leap in reasoning capabilities compared to its backbone model. In the **GeoTest2025** benchmark (derived from restricted 2025 National Postgraduate Entrance Examination questions), it achieves near-commercial performance.
46
+
47
+ | Model | GeoTest2025 (Geography) | AIME 2024 (Math Logic) | HumanEval | MMMU pro |
48
+ | :--- | :---: | :---: | :---: |
49
+ | **LandAI-Base-7B (Ours)** | **93.3%** | **16.7%** | **66.4%** | **44.7%** |
50
+ | Qwen2.5-VL-7B (Baseline) | 46.7% | 3.3% | 67.3% | 41.2% |
51
+ | GPT-4o | 92.1% | 9.3% | 90.2% | 51.9% |
52
+ | Gemini 2.5 Pro | 98.3% | 92.0% | - | 71.2% |
53
+
54
+ > **Note**: As shown in Extended Data Table 2 of the paper, LandAI-Base achieves a **2.5x improvement** in geographic reasoning and a **5x improvement** in mathematical logic over the vanilla baseline.
55
+
56
+ ## ๐Ÿ“‚ Dataset Composition
57
+
58
+ The explicit reasoning capability of LandAI-Base stems from its training data distribution:
59
+
60
+ | Dataset Source | Samples | Purpose |
61
+ | :--- | :--- | :--- |
62
+ | **Geo-Base-Thinking-14K** | ~14.7k | Domain-specific geospatial logic & knowledge |
63
+ | **OpenR1-Math** | ~96k | Mathematical reasoning infrastructure |
64
+ | **OpenThoughts** | ~114k | General scientific literacy (Physics/Chem/Bio) |
65
+ | **Chinese-Data-R1** | ~110k | Linguistic nuance and logic bridging |
66
+
67
+ ## ๐Ÿ› ๏ธ Quick Start
68
+
69
+ LandAI-Base follows the standard **Qwen2.5-VL** architecture. You can use it for geospatial Question Answering or as a base for further RL training.
70
+