Update README.md
#2
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -5,15 +5,14 @@ license: mit
|
|
| 5 |
# DPT 3.1 (BEiT backbone)
|
| 6 |
|
| 7 |
DPT (Dense Prediction Transformer) model trained on 1.4 million images for monocular depth estimation. It was introduced in the paper [Vision Transformers for Dense Prediction](https://arxiv.org/abs/2103.13413) by Ranftl et al. (2021) and first released in [this repository](https://github.com/isl-org/DPT).
|
| 8 |
-
DPT uses the [BEiT](https://huggingface.co/docs/transformers/model_doc/beit) model as backbone and adds a neck + head on top for monocular depth estimation.
|
| 9 |
-
|
| 10 |
-

|
| 11 |
|
| 12 |
Disclaimer: The team releasing DPT did not write a model card for this model so this model card has been written by the Hugging Face team.
|
| 13 |
|
| 14 |
## Model description
|
| 15 |
|
| 16 |
-
|
|
|
|
|
|
|
| 17 |
|
| 18 |
## How to use
|
| 19 |
|
|
|
|
| 5 |
# DPT 3.1 (BEiT backbone)
|
| 6 |
|
| 7 |
DPT (Dense Prediction Transformer) model trained on 1.4 million images for monocular depth estimation. It was introduced in the paper [Vision Transformers for Dense Prediction](https://arxiv.org/abs/2103.13413) by Ranftl et al. (2021) and first released in [this repository](https://github.com/isl-org/DPT).
|
|
|
|
|
|
|
|
|
|
| 8 |
|
| 9 |
Disclaimer: The team releasing DPT did not write a model card for this model so this model card has been written by the Hugging Face team.
|
| 10 |
|
| 11 |
## Model description
|
| 12 |
|
| 13 |
+
This DPT model uses the [BEiT](https://huggingface.co/docs/transformers/model_doc/beit) model as backbone and adds a neck + head on top for monocular depth estimation.
|
| 14 |
+
|
| 15 |
+

|
| 16 |
|
| 17 |
## How to use
|
| 18 |
|