LandAI-Base: Activating Geospatial Chain-of-Thought Reasoning

[Paper (Under Review)]

πŸ“– Introduction

LandAI-Base is the foundational Supervised Fine-Tuned (SFT) model of the LandAI family. Built upon the Qwen2.5-VL-7B-Instruct architecture, it is specifically designed to activate domain-specific logical reasoning in Earth Observation tasks.

Unlike general-purpose multimodal models, LandAI-Base has been fine-tuned on a composite corpus of approximately 334,000 reasoning chains, including the novel Geo-Base-Thinking-14K dataset. This process instills the model with the "epistemological authority" of geography experts, enabling it to decompose complex spatial problems before engaging in visual recognition.

LandAI-Base serves two primary purposes:

  1. A robust baseline for geospatial reasoning tasks (Q&A, analysis).
  2. The "Cold Start" initialization for the LandAI-L1 model (trained via GRPO-L1).

πŸš€ Key Features

  • Domain-Specific Cognitive Activation: Fine-tuned to simulate the reasoning patterns of geography experts, moving from rote memorization to logical deduction.
  • High-Quality Training Data: Trained on a curated mix of:
    • Geo-Base-Thinking-14K: ~14.7k distillations from geography entrance exams and textbooks.
    • General Reasoning Corpus: Subsets from OpenR1-Math, OpenThoughts, and Chinese-Data-R1 to enhance mathematical and scientific logic.
  • Strong Zero-Shot Performance: Significantly outperforms the vanilla Qwen2.5-VL-7B on geographic benchmark exams.
  • MS-Swift Compatibility: Fully compatible with the ms-swift training framework.

πŸ“Š Performance Benchmarks

LandAI-Base demonstrates a substantial leap in reasoning capabilities compared to its backbone model. In the GeoTest2025 benchmark (derived from restricted 2025 National Postgraduate Entrance Examination questions), it achieves near-commercial performance.

Model GeoTest2025 (Geography) AIME 2024 HumanEval MMMU pro
LandAI-Base-7B (Ours) 93.3% 16.7% 66.4% 44.7%
Qwen2.5-VL-7B (Baseline) 46.7% 3.3% 67.3% 41.2%
GPT-4o 92.1% 9.3% 90.2% 51.9%
Gemini 2.5 Pro 98.3% 92.0% - 71.2%

πŸ“‚ Dataset Composition

The explicit reasoning capability of LandAI-Base stems from its training data distribution:

Dataset Source Samples Purpose
Geo-Base-Thinking-14K ~14.7k Domain-specific geospatial logic & knowledge
OpenR1-Math ~96k Mathematical reasoning infrastructure
OpenThoughts ~114k General scientific literacy (Physics/Chem/Bio)
Chinese-Data-R1 ~110k Linguistic nuance and logic bridging

πŸ› οΈ Quick Start

LandAI-Base follows the standard Qwen2.5-VL architecture. You can use it for geospatial Question Answering or as a base for further RL training.

Downloads last month
9
Safetensors
Model size
8B params
Tensor type
BF16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for zhou777/LandAI-Base

Quantizations
2 models