Improve model card for AllTracker: add metadata, project page, and usage

#2
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +52 -2
README.md CHANGED
@@ -2,8 +2,58 @@
2
  license: mit
3
  tags:
4
  - tracking
 
5
  ---
6
 
7
- This repository contains the model described in [AllTracker: Efficient Dense Point Tracking at High Resolution](https://huggingface.co/papers/2506.07310).
8
 
9
- Code: https://github.com/aharley/alltracker
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  license: mit
3
  tags:
4
  - tracking
5
+ pipeline_tag: image-to-image
6
  ---
7
 
8
+ # AllTracker: Efficient Dense Point Tracking at High Resolution
9
 
10
+ [[Paper](https://huggingface.co/papers/2506.07310)] | [[Project Page](https://alltracker.github.io/)] | [[Code](https://github.com/aharley/alltracker)] | [[Gradio Demo](https://huggingface.co/spaces/aharley/alltracker)]
11
+
12
+ <img src='https://alltracker.github.io/images/monkey.jpg'>
13
+
14
+ **AllTracker is a point tracking model which is faster and more accurate than other similar models, while also producing dense output at high resolution.**
15
+
16
+ AllTracker estimates long-range point tracks by way of estimating the flow field between a query frame and every other frame of a video. Unlike existing point tracking methods, our approach delivers high-resolution and dense (all-pixel) correspondence fields, which can be visualized as flow maps. Unlike existing optical flow methods, our approach corresponds one frame to hundreds of subsequent frames, rather than just the next frame. We develop a new architecture for this task, blending techniques from existing work in optical flow and point tracking: the model performs iterative inference on low-resolution grids of correspondence estimates, propagating information spatially via 2D convolution layers, and propagating information temporally via pixel-aligned attention layers. The model is fast and parameter-efficient (16 million parameters), and delivers state-of-the-art point tracking accuracy at high resolution (i.e., tracking 768x1024 pixels, on a 40G GPU). A benefit of our design is that we can train jointly on optical flow datasets and point tracking datasets, and we find that doing so is crucial for top performance.
17
+
18
+ ## Usage (Running the Demo)
19
+
20
+ First, set up a fresh conda environment for AllTracker:
21
+
22
+ ```bash
23
+ conda create -n alltracker python=3.12.8
24
+ conda activate alltracker
25
+ pip install -r requirements.txt
26
+ ```
27
+
28
+ Download the sample video:
29
+ ```bash
30
+ cd demo_video
31
+ sh download_video.sh
32
+ cd ..
33
+ ```
34
+
35
+ Run the demo:
36
+ ```bash
37
+ python demo.py --mp4_path ./demo_video/monkey.mp4
38
+ ```
39
+ The demo script will automatically download the model weights from [huggingface](https://huggingface.co/aharley/alltracker/tree/main) if needed.
40
+
41
+ For a fancier visualization, giving a side-by-side view of the input and output, try this:
42
+ ```bash
43
+ python demo.py --mp4_path ./demo_video/monkey.mp4 --query_frame 32 --conf_thr 0.01 --bkg_opacity 0.0 --rate 2 --hstack --query_frame 16
44
+ ```
45
+
46
+ For more detailed information on training and evaluation, please refer to the [official GitHub repository](https://github.com/aharley/alltracker).
47
+
48
+ ## Citation
49
+
50
+ If you use this code for your research, please cite:
51
+
52
+ ```bibtex
53
+ @inproceedings{harley2025alltracker,
54
+ author = {Adam W. Harley and Yang You and Xinglong Sun and Yang Zheng and Nikhil Raghuraman and Yunqi Gu and Sheldon Liang and Wen-Hsuan Chu and Achal Dave and Pavel Tokmakov and Suya You and Rares Ambrus and Katerina Fragkiadaki and Leonidas J. Guibas},
55
+ title = {All{T}racker: {E}fficient Dense Point Tracking at High Resolution},
56
+ booktitle = {ICCV},
57
+ year = {2025}
58
+ }
59
+ ```