File size: 1,472 Bytes
8da7be0
 
 
 
0bcdd33
 
 
 
 
 
 
 
 
 
 
8da7be0
0bcdd33
8da7be0
0bcdd33
 
 
 
59bcb2c
2b71b7a
0bcdd33
 
59bcb2c
0bcdd33
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
---
tags:
- model_hub_mixin
- pytorch_model_hub_mixin
- computer-vision
- 3d-reconstruction
- multi-view-stereo
- depth-estimation
- camera-pose
- covisibility
- mapanything
license: apache-2.0
language:
- en
pipeline_tag: image-to-3d
---
## Overview

MapAnything is a simple, end-to-end trained transformer model that directly regresses the factored metric 3D geometry of a scene given various types of modalities as inputs. A single feed-forward model supports over 12 different 3D reconstruction tasks including multi-image sfm, multi-view stereo, monocular metric depth estimation, registration, depth completion and more.

This is the Apache 2.0 variant of the model released in September 2025, i.e., the V1 Version.

Latest Apache 2.0 model here: https://huggingface.co/facebook/map-anything-apache

## Quick Start

Please refer to our Github Repo: https://github.com/facebookresearch/map-anything

## Citation

If you find our repository useful, please consider giving it a star ⭐ and citing our paper in your work:

```bibtex
@inproceedings{keetha2026mapanything,
  title={{MapAnything}: Universal Feed-Forward Metric 3D Reconstruction},
  author={Keetha, Nikhil and M{\"u}ller, Norman and Sch{\"o}nberger, Johannes and Porzi, Lorenzo and Zhang, Yuchen and Fischer, Tobias and Knapitsch, Arno and Zauss, Duncan and Weber, Ethan and Antunes, Nelson and others},
  booktitle={International Conference on 3D Vision (3DV)},
  year={2026},
  organization={IEEE}
}
```