Nima Boscarino
commited on
Commit
·
ebf9aef
1
Parent(s):
6bbf0db
Update README.md
Browse files
README.md
CHANGED
|
@@ -1,3 +1,32 @@
|
|
| 1 |
---
|
|
|
|
|
|
|
|
|
|
| 2 |
license: mit
|
|
|
|
|
|
|
| 3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
+
tags:
|
| 3 |
+
- pixel-tracking
|
| 4 |
+
- computer-vision
|
| 5 |
license: mit
|
| 6 |
+
library: pytorch
|
| 7 |
+
inference: false
|
| 8 |
---
|
| 9 |
+
|
| 10 |
+
# PIPS: Particle Video Revisited: Tracking Through Occlusions Using Point Trajectories
|
| 11 |
+
|
| 12 |
+
* Model Authors: Adam W Harley and Zhaoyuan Fang and Katerina Fragkiadaki
|
| 13 |
+
* Paper: Particle Video Revisited: Tracking Through Occlusions Using Point Trajectories (ECCV 2022 - https://arxiv.org/abs/2204.04153
|
| 14 |
+
* Code Repo: https://github.com/aharley/pips
|
| 15 |
+
* Project Homepage: https://particle-video-revisited.github.io
|
| 16 |
+
|
| 17 |
+
From the paper abstract:
|
| 18 |
+
|
| 19 |
+
> [...] we revisit Sand and Teller's "particle video" approach, and study pixel tracking as a long-range motion estimation problem, where every pixel is described with a trajectory that locates it in multiple future frames. We re-build this classic approach using components that drive the current state-of-the-art in flow and object tracking, such as dense cost maps, iterative optimization, and learned appearance updates. We train our models using long-range amodal point trajectories mined from existing optical flow data that we synthetically augment with multi-frame occlusions.
|
| 20 |
+
|
| 21 |
+

|
| 22 |
+
|
| 23 |
+
# Citation
|
| 24 |
+
|
| 25 |
+
```
|
| 26 |
+
@inproceedings{harley2022particle,
|
| 27 |
+
title={Particle Video Revisited: Tracking Through Occlusions Using Point Trajectories},
|
| 28 |
+
author={Adam W Harley and Zhaoyuan Fang and Katerina Fragkiadaki},
|
| 29 |
+
booktitle={ECCV},
|
| 30 |
+
year={2022}
|
| 31 |
+
}
|
| 32 |
+
```
|