--- pipeline_tag: image-feature-extraction license: mit datasets: - IGNF/FLAIR-HUB --- # RAMEN: Resolution-Adjustable Multimodal Encoder for Earth Observation [Paper](https://huggingface.co/papers/2512.05025) | [Code](https://github.com/nicolashoudre/RAMEN) RAMEN is a resolution-adjustable multimodal encoder that learns a shared visual representation across Earth Observation (EO) data in a fully sensor-agnostic manner. It treats modality and spatial/temporal resolutions as key input features, enabling coherent analysis across modalities. Its main methodological contribution is to define spatial resolution as a controllable output parameter, giving users direct control over the desired level of detail at inference and allowing explicit trade-offs between spatial precision and computational cost.

RAMEN workflow

## Key features - 🛰️ **Sensor-agnostic foundation model**: RAMEN supports any kind of multispectral, SAR or elevation maps modalities. Just specify input shape, channels and original spatial resolution (GSD) ! - 🔧 **Adjustable feature map resolution**: Customize the resolution of feature maps to suit specific downstream tasks and computational constraints. - 🌍 **Multimodal data fusion**: Effectively combine data from multiple modalities into a unified representation. ## PANGAEA Bench evaluation All downstream tasks results presented in RAMEN were conducted using the [PANGAEA](https://github.com/VMarsocci/pangaea-bench) Benchmark. We report here the main results obtained on eight tasks. | Model | BurnSr | MADOS | PASTIS | Sen1Fl11 | DEN | CTM-SS | SN7 | AI4Farms | Avg. mIoU | Avg. Rank | |-------|---------|--------|--------|----------|------|--------|------|-----------|-----------|-----------| | CROMA | 82.42 | 67.55 | 32.32 | 90.89 | 38.29 | 49.38 | 59.28 | 25.65 | 55.72 | 6.50 | | DOFA | 80.63 | 59.58 | 30.02 | 89.37 | 39.29 | 51.33 | **61.84** | 27.07 | 54.89 | 7.50 | | TerraMind-B | 82.42 | 69.52 | 40.51 | 90.62 | 37.87 | **55.80** | 60.61 | 28.12 | 58.18 | 4.25 | | TerraMind-L | 82.93 | **75.57** | **43.13** | 90.78 | 37.89 | 55.04 | 59.98 | 27.47 | 59.10 | 3.75 | | **RAMEN (ours)** | **85.02** | 69.72 | 42.29 | **91.03** | **39.85** | 53.27 | 60.31 | **38.78** | **60.03** | **2.63** | More informations on how to reproduce results and implement RAMEN in PANGAEA can be found in the [`pangaea-bench`](https://github.com/nicolashoudre/RAMEN/tree/main/pangaea-bench) folder. ## Citation If you use RAMEN, please cite our paper: ```bibtex @article{RAMEN, title={{RAMEN}: Resolution-Adjustable Multimodal Encoder for Earth Observation}, author={Nicolas Houdré and Diego Marcos and Hugo Riffaud de Turckheim and Dino Ienco and Laurent Wendling and Camille Kurtz and Sylvain Lobry}, journal={arXiv preprint arXiv:2512.05025}, year={2025} } ```