DepthAnythingV2 / README.md
cooper_robot
Add release note for v1.1.0
b419c06
---
library_name: pytorch
---
![depthanythingv2_logo](resource/DepthAnythingV2.png)
DepthAnythingV2 is a lightweight depth estimation model designed to predict accurate per-pixel depth maps from a single RGB image, optimized for efficiency and versatility across scenes.
Original paper: [DepthAnything V2](https://arxiv.org/abs/2406.09414)
# DepthAnythingV2-Small
This model uses the DepthAnythingV2-Small variant, which balances model size and inference speed while maintaining strong depth estimation accuracy. It is well suited for applications such as AR/VR, robotics, scene reconstruction, and real-time 3D perception on edge devices.
Model Configuration:
- Reference implementation: [Official DepthAnythingV2 source code](https://github.com/DepthAnything/Depth-Anything-V2)
- Original Weight:[Depth-Anything-V2-Small](https://huggingface.co/depth-anything/Depth-Anything-V2-Small/resolve/main/depth_anything_v2_vits.pth?download=true)
- Resolution: 3x224x224
- Support Cooper version:
- Cooper SDK: [2.5.2]
- Cooper Foundry: [2.2]
| Model | Device | Model Link |
| :-----: | :-----: | :-----: |
| DepthAnythingV2-Small | N1-655 | [Model_Link](https://huggingface.co/Ambarella/DepthAnythingV2/blob/main/n1-655_depthanything_v2_small.bin) |
| DepthAnythingV2-Small | CV72 | [Model_Link](https://huggingface.co/Ambarella/DepthAnythingV2/blob/main/cv72_depthanything_v2_small.bin) |
| DepthAnythingV2-Small | CV75 | [Model_Link](https://huggingface.co/Ambarella/DepthAnythingV2/blob/main/cv75_depthanything_v2_small.bin) |