File size: 1,541 Bytes
b419c06
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
---
library_name: pytorch
---

![depthanythingv2_logo](resource/DepthAnythingV2.png)

DepthAnythingV2 is a lightweight depth estimation model designed to predict accurate per-pixel depth maps from a single RGB image, optimized for efficiency and versatility across scenes.

Original paper: [DepthAnything V2](https://arxiv.org/abs/2406.09414)

# DepthAnythingV2-Small

This model uses the DepthAnythingV2-Small variant, which balances model size and inference speed while maintaining strong depth estimation accuracy. It is well suited for applications such as AR/VR, robotics, scene reconstruction, and real-time 3D perception on edge devices.

Model Configuration:
- Reference implementation: [Official DepthAnythingV2 source code](https://github.com/DepthAnything/Depth-Anything-V2)
- Original Weight:[Depth-Anything-V2-Small](https://huggingface.co/depth-anything/Depth-Anything-V2-Small/resolve/main/depth_anything_v2_vits.pth?download=true)
- Resolution: 3x224x224
- Support Cooper version:
    - Cooper SDK: [2.5.2]
    - Cooper Foundry: [2.2]

| Model | Device | Model Link |
| :-----: | :-----: | :-----: |
| DepthAnythingV2-Small | N1-655 | [Model_Link](https://huggingface.co/Ambarella/DepthAnythingV2/blob/main/n1-655_depthanything_v2_small.bin) |
| DepthAnythingV2-Small | CV72 | [Model_Link](https://huggingface.co/Ambarella/DepthAnythingV2/blob/main/cv72_depthanything_v2_small.bin) |
| DepthAnythingV2-Small | CV75 | [Model_Link](https://huggingface.co/Ambarella/DepthAnythingV2/blob/main/cv75_depthanything_v2_small.bin) |