Add pipeline tag and links to paper/code
#1
by nielsr HF Staff - opened
README.md
CHANGED
|
@@ -2,15 +2,20 @@
|
|
| 2 |
license: other
|
| 3 |
license_name: nvidia
|
| 4 |
license_link: LICENSE
|
|
|
|
| 5 |
---
|
| 6 |
|
|
|
|
| 7 |
|
|
|
|
|
|
|
|
|
|
| 8 |
|
| 9 |
## Model Overview
|
| 10 |
|
| 11 |
### Description:
|
| 12 |
|
| 13 |
-
VideoMAE model used for training AutoGaze. This model is for research and development only.
|
| 14 |
|
| 15 |
### License/Terms of Use:
|
| 16 |
|
|
@@ -124,3 +129,17 @@ The raw videos are collected from public dataset including Ego4D, 100DoH, Intern
|
|
| 124 |
NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications. When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse. Please report model quality, risk, security vulnerabilities or NVIDIA AI Concerns [here](https://www.nvidia.com/en-us/support/submit-security-vulnerability/). <br>
|
| 125 |
|
| 126 |
Please make sure you have proper rights and permissions for all input image and video content; if image or video includes people, personal health information, or intellectual property, the image or video generated will not blur or maintain proportions of image subjects included. <br>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2 |
license: other
|
| 3 |
license_name: nvidia
|
| 4 |
license_link: LICENSE
|
| 5 |
+
pipeline_tag: video-classification
|
| 6 |
---
|
| 7 |
|
| 8 |
+
# VideoMAE_AutoGaze
|
| 9 |
|
| 10 |
+
[**Attend Before Attention: Efficient and Scalable Video Understanding via Autoregressive Gazing**](https://huggingface.co/papers/2603.12254)
|
| 11 |
+
|
| 12 |
+
[**Project Page**](https://autogaze.github.io/) | [**GitHub**](https://github.com/NVlabs/AutoGaze) | [**Demo**](https://huggingface.co/spaces/bfshi/AutoGaze)
|
| 13 |
|
| 14 |
## Model Overview
|
| 15 |
|
| 16 |
### Description:
|
| 17 |
|
| 18 |
+
VideoMAE model used for training AutoGaze. This model is for research and development only. <br>
|
| 19 |
|
| 20 |
### License/Terms of Use:
|
| 21 |
|
|
|
|
| 129 |
NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications. When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse. Please report model quality, risk, security vulnerabilities or NVIDIA AI Concerns [here](https://www.nvidia.com/en-us/support/submit-security-vulnerability/). <br>
|
| 130 |
|
| 131 |
Please make sure you have proper rights and permissions for all input image and video content; if image or video includes people, personal health information, or intellectual property, the image or video generated will not blur or maintain proportions of image subjects included. <br>
|
| 132 |
+
|
| 133 |
+
## Citation
|
| 134 |
+
|
| 135 |
+
```bibtex
|
| 136 |
+
@misc{shi2026attendattentionefficientscalable,
|
| 137 |
+
title={Attend Before Attention: Efficient and Scalable Video Understanding via Autoregressive Gazing},
|
| 138 |
+
author={Baifeng Shi and Stephanie Fu and Long Lian and Hanrong Ye and David Eigen and Aaron Reite and Boyi Li and Jan Kautz and Song Han and David M. Chan and Pavlo Molchanov and Trevor Darrell and Hongxu Yin},
|
| 139 |
+
year={2026},
|
| 140 |
+
eprint={2603.12254},
|
| 141 |
+
archivePrefix={arXiv},
|
| 142 |
+
primaryClass={cs.CV},
|
| 143 |
+
url={https://arxiv.org/abs/2603.12254},
|
| 144 |
+
}
|
| 145 |
+
```
|