Improve model card for AllTracker: add metadata, project page, and usage
Browse filesThis PR enhances the model card for the AllTracker model by:
- Adding the `pipeline_tag: image-to-image` for better discoverability and categorization on the Hub.
- Including the official project page link for comprehensive context.
- Expanding the model description based on the paper abstract and GitHub README to provide a richer overview.
- Adding clear usage instructions to enable users to quickly run the demo.
- Including the Gradio demo link and BibTeX citation for full information and proper attribution.
README.md
CHANGED
|
@@ -2,8 +2,58 @@
|
|
| 2 |
license: mit
|
| 3 |
tags:
|
| 4 |
- tracking
|
|
|
|
| 5 |
---
|
| 6 |
|
| 7 |
-
|
| 8 |
|
| 9 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2 |
license: mit
|
| 3 |
tags:
|
| 4 |
- tracking
|
| 5 |
+
pipeline_tag: image-to-image
|
| 6 |
---
|
| 7 |
|
| 8 |
+
# AllTracker: Efficient Dense Point Tracking at High Resolution
|
| 9 |
|
| 10 |
+
[[Paper](https://huggingface.co/papers/2506.07310)] | [[Project Page](https://alltracker.github.io/)] | [[Code](https://github.com/aharley/alltracker)] | [[Gradio Demo](https://huggingface.co/spaces/aharley/alltracker)]
|
| 11 |
+
|
| 12 |
+
<img src='https://alltracker.github.io/images/monkey.jpg'>
|
| 13 |
+
|
| 14 |
+
**AllTracker is a point tracking model which is faster and more accurate than other similar models, while also producing dense output at high resolution.**
|
| 15 |
+
|
| 16 |
+
AllTracker estimates long-range point tracks by way of estimating the flow field between a query frame and every other frame of a video. Unlike existing point tracking methods, our approach delivers high-resolution and dense (all-pixel) correspondence fields, which can be visualized as flow maps. Unlike existing optical flow methods, our approach corresponds one frame to hundreds of subsequent frames, rather than just the next frame. We develop a new architecture for this task, blending techniques from existing work in optical flow and point tracking: the model performs iterative inference on low-resolution grids of correspondence estimates, propagating information spatially via 2D convolution layers, and propagating information temporally via pixel-aligned attention layers. The model is fast and parameter-efficient (16 million parameters), and delivers state-of-the-art point tracking accuracy at high resolution (i.e., tracking 768x1024 pixels, on a 40G GPU). A benefit of our design is that we can train jointly on optical flow datasets and point tracking datasets, and we find that doing so is crucial for top performance.
|
| 17 |
+
|
| 18 |
+
## Usage (Running the Demo)
|
| 19 |
+
|
| 20 |
+
First, set up a fresh conda environment for AllTracker:
|
| 21 |
+
|
| 22 |
+
```bash
|
| 23 |
+
conda create -n alltracker python=3.12.8
|
| 24 |
+
conda activate alltracker
|
| 25 |
+
pip install -r requirements.txt
|
| 26 |
+
```
|
| 27 |
+
|
| 28 |
+
Download the sample video:
|
| 29 |
+
```bash
|
| 30 |
+
cd demo_video
|
| 31 |
+
sh download_video.sh
|
| 32 |
+
cd ..
|
| 33 |
+
```
|
| 34 |
+
|
| 35 |
+
Run the demo:
|
| 36 |
+
```bash
|
| 37 |
+
python demo.py --mp4_path ./demo_video/monkey.mp4
|
| 38 |
+
```
|
| 39 |
+
The demo script will automatically download the model weights from [huggingface](https://huggingface.co/aharley/alltracker/tree/main) if needed.
|
| 40 |
+
|
| 41 |
+
For a fancier visualization, giving a side-by-side view of the input and output, try this:
|
| 42 |
+
```bash
|
| 43 |
+
python demo.py --mp4_path ./demo_video/monkey.mp4 --query_frame 32 --conf_thr 0.01 --bkg_opacity 0.0 --rate 2 --hstack --query_frame 16
|
| 44 |
+
```
|
| 45 |
+
|
| 46 |
+
For more detailed information on training and evaluation, please refer to the [official GitHub repository](https://github.com/aharley/alltracker).
|
| 47 |
+
|
| 48 |
+
## Citation
|
| 49 |
+
|
| 50 |
+
If you use this code for your research, please cite:
|
| 51 |
+
|
| 52 |
+
```bibtex
|
| 53 |
+
@inproceedings{harley2025alltracker,
|
| 54 |
+
author = {Adam W. Harley and Yang You and Xinglong Sun and Yang Zheng and Nikhil Raghuraman and Yunqi Gu and Sheldon Liang and Wen-Hsuan Chu and Achal Dave and Pavel Tokmakov and Suya You and Rares Ambrus and Katerina Fragkiadaki and Leonidas J. Guibas},
|
| 55 |
+
title = {All{T}racker: {E}fficient Dense Point Tracking at High Resolution},
|
| 56 |
+
booktitle = {ICCV},
|
| 57 |
+
year = {2025}
|
| 58 |
+
}
|
| 59 |
+
```
|