File size: 2,497 Bytes
6028ebb |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 |
---
license: apache-2.0
language:
- en
tags:
- vision-language
- vlm
- grpo
- earthmind
- geospatial
- remote-sensing
library_name: transformers
pipeline_tag: image-text-to-text
---
# EarthMind-R1
EarthMind-R1 is a vision-language model fine-tuned using GRPO (Group Relative Policy Optimization) for geospatial and remote sensing image understanding tasks.
## Model Description
- **Base Model:** EarthMind-4B
- **Training Method:** GRPO (Group Relative Policy Optimization)
- **Training Data:** Geospatial instruction dataset
- **Fine-tuning:** LoRA adapters merged into base weights
## Usage
### Quick Start
```python
import torch
from PIL import Image
from transformers import AutoModelForCausalLM, AutoTokenizer
# Load model and tokenizer
model_id = "aadex/Earthmind-R1"
tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained(
model_id,
trust_remote_code=True,
torch_dtype=torch.bfloat16,
device_map="auto",
)
# Load an image
image = Image.open("your_image.jpg").convert("RGB")
# Ask a question
question = "Describe what you see in this satellite image."
# Use model's chat interface
response = model.chat(
tokenizer=tokenizer,
question=question,
images=[image],
generation_config={
"max_new_tokens": 512,
"temperature": 0.7,
"do_sample": True,
},
)
print(response)
```
### Expected Output Format
The model is trained to provide structured responses:
```
<think>
[Reasoning about the image content]
</think>
<answer>
[Final answer to the question]
</answer>
```
## Requirements
```
torch>=2.0
transformers>=4.40
accelerate
pillow
```
## Hardware Requirements
- **Minimum:** 16GB VRAM (with bfloat16)
- **Recommended:** 24GB VRAM for comfortable inference
## Training Details
- **Framework:** VLM-R1 + TRL
- **Optimizer:** AdamW
- **Learning Rate:** 1e-6
- **LoRA Configuration:**
- r: 32
- alpha: 64
- dropout: 0.05
- **GRPO Settings:**
- num_generations: 4
- num_iterations: 2
- beta: 0.01
## Limitations
- Optimized for geospatial/remote sensing imagery
- May not perform as well on general domain images
- Response quality depends on image resolution and clarity
## Citation
If you use this model, please cite:
```bibtex
@misc{earthmind-r1,
title={EarthMind-R1: GRPO Fine-tuned Vision-Language Model for Geospatial Understanding},
author={Your Name},
year={2024},
publisher={HuggingFace}
}
```
## License
Apache 2.0
|