vision
tracking
yangyi02 commited on
Commit
9db51a7
Β·
verified Β·
1 Parent(s): 5a62c1b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -9
README.md CHANGED
@@ -13,16 +13,10 @@ This repository contains the checkpoints of several point tracking models develo
13
 
14
  ## Included Models
15
 
16
- - **TAPIR** – A fast and accurate point tracker for continuous point trajectories in space-time.
17
 
18
- 🌐 **Project page**: [https://deepmind-tapir.github.io/](https://deepmind-tapir.github.io/)
19
-
20
- - **BootsTAPIR** – A bootstrapped variant of TAPIR that improves robustness and stability across long videos via self-supervised refinement.
21
-
22
- 🌐 **Project page**: [https://bootstap.github.io/](https://bootstap.github.io/)
23
 
24
- - **TAPNext** – A new generative approach that frames point tracking as next-token prediction, enabling semi-dense, accurate, and temporally coherent tracking across challenging videos, including those presented in the paper [**TAPNext: Tracking Any Point (TAP) as Next Token Prediction**](https://huggingface.co/papers/2504.05579).
25
-
26
- 🌐 **Project page**: [https://tap-next.github.io/](https://tap-next.github.io/)
27
 
28
  These models provide state-of-the-art performance for tracking arbitrary points in videos and support research and applications in robotics, perception, and video generation.
 
13
 
14
  ## Included Models
15
 
16
+ [**TAPIR**](https://deepmind-tapir.github.io/) – A fast and accurate point tracker for continuous point trajectories in space-time.
17
 
18
+ [**BootsTAPIR**](https://bootstap.github.io/) – A bootstrapped variant of TAPIR that improves robustness and stability across long videos via self-supervised refinement.
 
 
 
 
19
 
20
+ [**TAPNext**](https://tap-next.github.io/) – A new generative approach that frames point tracking as next-token prediction, enabling semi-dense, accurate, and temporally coherent tracking across challenging videos, including those presented in the paper [**TAPNext: Tracking Any Point (TAP) as Next Token Prediction**](https://huggingface.co/papers/2504.05579).
 
 
21
 
22
  These models provide state-of-the-art performance for tracking arbitrary points in videos and support research and applications in robotics, perception, and video generation.