Improve model card for AllTracker: add metadata, project page, and usage
#2
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -2,8 +2,58 @@
|
|
| 2 |
license: mit
|
| 3 |
tags:
|
| 4 |
- tracking
|
|
|
|
| 5 |
---
|
| 6 |
|
| 7 |
-
|
| 8 |
|
| 9 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2 |
license: mit
|
| 3 |
tags:
|
| 4 |
- tracking
|
| 5 |
+
pipeline_tag: image-to-image
|
| 6 |
---
|
| 7 |
|
| 8 |
+
# AllTracker: Efficient Dense Point Tracking at High Resolution
|
| 9 |
|
| 10 |
+
[[Paper](https://huggingface.co/papers/2506.07310)] | [[Project Page](https://alltracker.github.io/)] | [[Code](https://github.com/aharley/alltracker)] | [[Gradio Demo](https://huggingface.co/spaces/aharley/alltracker)]
|
| 11 |
+
|
| 12 |
+
<img src='https://alltracker.github.io/images/monkey.jpg'>
|
| 13 |
+
|
| 14 |
+
**AllTracker is a point tracking model which is faster and more accurate than other similar models, while also producing dense output at high resolution.**
|
| 15 |
+
|
| 16 |
+
AllTracker estimates long-range point tracks by way of estimating the flow field between a query frame and every other frame of a video. Unlike existing point tracking methods, our approach delivers high-resolution and dense (all-pixel) correspondence fields, which can be visualized as flow maps. Unlike existing optical flow methods, our approach corresponds one frame to hundreds of subsequent frames, rather than just the next frame. We develop a new architecture for this task, blending techniques from existing work in optical flow and point tracking: the model performs iterative inference on low-resolution grids of correspondence estimates, propagating information spatially via 2D convolution layers, and propagating information temporally via pixel-aligned attention layers. The model is fast and parameter-efficient (16 million parameters), and delivers state-of-the-art point tracking accuracy at high resolution (i.e., tracking 768x1024 pixels, on a 40G GPU). A benefit of our design is that we can train jointly on optical flow datasets and point tracking datasets, and we find that doing so is crucial for top performance.
|
| 17 |
+
|
| 18 |
+
## Usage (Running the Demo)
|
| 19 |
+
|
| 20 |
+
First, set up a fresh conda environment for AllTracker:
|
| 21 |
+
|
| 22 |
+
```bash
|
| 23 |
+
conda create -n alltracker python=3.12.8
|
| 24 |
+
conda activate alltracker
|
| 25 |
+
pip install -r requirements.txt
|
| 26 |
+
```
|
| 27 |
+
|
| 28 |
+
Download the sample video:
|
| 29 |
+
```bash
|
| 30 |
+
cd demo_video
|
| 31 |
+
sh download_video.sh
|
| 32 |
+
cd ..
|
| 33 |
+
```
|
| 34 |
+
|
| 35 |
+
Run the demo:
|
| 36 |
+
```bash
|
| 37 |
+
python demo.py --mp4_path ./demo_video/monkey.mp4
|
| 38 |
+
```
|
| 39 |
+
The demo script will automatically download the model weights from [huggingface](https://huggingface.co/aharley/alltracker/tree/main) if needed.
|
| 40 |
+
|
| 41 |
+
For a fancier visualization, giving a side-by-side view of the input and output, try this:
|
| 42 |
+
```bash
|
| 43 |
+
python demo.py --mp4_path ./demo_video/monkey.mp4 --query_frame 32 --conf_thr 0.01 --bkg_opacity 0.0 --rate 2 --hstack --query_frame 16
|
| 44 |
+
```
|
| 45 |
+
|
| 46 |
+
For more detailed information on training and evaluation, please refer to the [official GitHub repository](https://github.com/aharley/alltracker).
|
| 47 |
+
|
| 48 |
+
## Citation
|
| 49 |
+
|
| 50 |
+
If you use this code for your research, please cite:
|
| 51 |
+
|
| 52 |
+
```bibtex
|
| 53 |
+
@inproceedings{harley2025alltracker,
|
| 54 |
+
author = {Adam W. Harley and Yang You and Xinglong Sun and Yang Zheng and Nikhil Raghuraman and Yunqi Gu and Sheldon Liang and Wen-Hsuan Chu and Achal Dave and Pavel Tokmakov and Suya You and Rares Ambrus and Katerina Fragkiadaki and Leonidas J. Guibas},
|
| 55 |
+
title = {All{T}racker: {E}fficient Dense Point Tracking at High Resolution},
|
| 56 |
+
booktitle = {ICCV},
|
| 57 |
+
year = {2025}
|
| 58 |
+
}
|
| 59 |
+
```
|