NikV09 commited on
Commit
4cf3561
·
verified ·
1 Parent(s): d33a87f

Push model using huggingface_hub.

Browse files
Files changed (3) hide show
  1. README.md +4 -34
  2. config.json +1 -1
  3. model.safetensors +1 -1
README.md CHANGED
@@ -2,39 +2,9 @@
2
  tags:
3
  - model_hub_mixin
4
  - pytorch_model_hub_mixin
5
- - computer-vision
6
- - 3d-reconstruction
7
- - multi-view-stereo
8
- - depth-estimation
9
- - camera-pose
10
- - covisibility
11
- - mapanything
12
- license: cc-by-nc-4.0
13
- language:
14
- - en
15
- pipeline_tag: image-to-3d
16
  ---
17
 
18
- ## Overview
19
-
20
- MapAnything is a simple, end-to-end trained transformer model that directly regresses the factored metric 3D geometry of a scene given various types of modalities as inputs. A single feed-forward model supports over 12 different 3D reconstruction tasks including multi-image sfm, multi-view stereo, monocular metric depth estimation, registration, depth completion and more.
21
-
22
- This is the Apache 2.0 variant of the model. Latest release on Dec 18th 2025.
23
-
24
- ## Quick Start
25
-
26
- Please refer to our Github Repo: https://github.com/facebookresearch/map-anything
27
-
28
- ## Citation
29
-
30
- If you find our repository useful, please consider giving it a star ⭐ and citing our paper in your work:
31
-
32
- ```bibtex
33
- @inproceedings{keetha2026mapanything,
34
- title={{MapAnything}: Universal Feed-Forward Metric 3D Reconstruction},
35
- author={Keetha, Nikhil and M{\"u}ller, Norman and Sch{\"o}nberger, Johannes and Porzi, Lorenzo and Zhang, Yuchen and Fischer, Tobias and Knapitsch, Arno and Zauss, Duncan and Weber, Ethan and Antunes, Nelson and others},
36
- booktitle={International Conference on 3D Vision (3DV)},
37
- year={2026},
38
- organization={IEEE}
39
- }
40
- ```
 
2
  tags:
3
  - model_hub_mixin
4
  - pytorch_model_hub_mixin
 
 
 
 
 
 
 
 
 
 
 
5
  ---
6
 
7
+ This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
8
+ - Code: [More Information Needed]
9
+ - Paper: [More Information Needed]
10
+ - Docs: [More Information Needed]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
config.json CHANGED
@@ -2,7 +2,7 @@
2
  "encoder_config": {
3
  "data_norm_type": "dinov2",
4
  "encoder_str": "dinov2",
5
- "gradient_checkpointing": true,
6
  "keep_first_n_layers": 24,
7
  "name": "dinov2_giant_24_layers",
8
  "norm_returned_features": false,
 
2
  "encoder_config": {
3
  "data_norm_type": "dinov2",
4
  "encoder_str": "dinov2",
5
+ "gradient_checkpointing": false,
6
  "keep_first_n_layers": 24,
7
  "name": "dinov2_giant_24_layers",
8
  "norm_returned_features": false,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8a213f800a28da955a9885d7b637279fe339a310b722c46a4a4029d6c56e16db
3
  size 4914062480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa06c0fdccefc5048e072c85935d5789b1e36b307f3859033c17f9dcb9fd5201
3
  size 4914062480