File size: 1,472 Bytes
76fab33 0414d50 76fab33 0414d50 515f332 8467f37 0414d50 515f332 0414d50 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 |
---
tags:
- model_hub_mixin
- pytorch_model_hub_mixin
- computer-vision
- 3d-reconstruction
- multi-view-stereo
- depth-estimation
- camera-pose
- covisibility
- mapanything
license: cc-by-nc-4.0
language:
- en
pipeline_tag: image-to-3d
---
## Overview
MapAnything is a simple, end-to-end trained transformer model that directly regresses the factored metric 3D geometry of a scene given various types of modalities as inputs. A single feed-forward model supports over 12 different 3D reconstruction tasks including multi-image sfm, multi-view stereo, monocular metric depth estimation, registration, depth completion and more.
This is the CC-BY-NC-4.0 variant of the model released in September 2025, i.e., the V1 Version.
Latest CC-BY-NC-4.0 model here: https://huggingface.co/facebook/map-anything
## Quick Start
Please refer to our Github Repo: https://github.com/facebookresearch/map-anything
## Citation
If you find our repository useful, please consider giving it a star ⭐ and citing our paper in your work:
```bibtex
@inproceedings{keetha2026mapanything,
title={{MapAnything}: Universal Feed-Forward Metric 3D Reconstruction},
author={Keetha, Nikhil and M{\"u}ller, Norman and Sch{\"o}nberger, Johannes and Porzi, Lorenzo and Zhang, Yuchen and Fischer, Tobias and Knapitsch, Arno and Zauss, Duncan and Weber, Ethan and Antunes, Nelson and others},
booktitle={International Conference on 3D Vision (3DV)},
year={2026},
organization={IEEE}
}
``` |