Darknsu commited on
Commit
9de6411
·
verified ·
1 Parent(s): 5e0414e

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +91 -0
README.md ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ # HAT: History-Augmented Anchor Transformer for Online Temporal Action Localization (ECCV 2024)
3
+ ### Sakib Reza, Yuexi Zhang, Mohsen Moghaddam, Octavia Camps
4
+ #### Northeastern University, Boston, United States
5
+ {reza.s,zhang.yuex,mohsen,o.camps}@northeastern.edu
6
+
7
+ ## [Arxiv Preprint](https://arxiv.org/abs/2408.06437)
8
+
9
+
10
+ ## Updates
11
+ - Aug 22, 2024 - EGTEA pre-extracted features and config files for other datasets added
12
+ - Aug 14, 2024 - Arxiv preprint added
13
+ - July 7, 2024 - initial code release
14
+
15
+ ## Installation
16
+
17
+ ### Prerequisites
18
+ - Ubuntu 20.04
19
+ - Python 3.10.9
20
+ - CUDA 12.0
21
+
22
+ ### Requirements
23
+ - pytorch==2.0.0
24
+ - numpy==1.23.5
25
+ - h5py==3.9.0
26
+ - ...
27
+
28
+ To install all required libraries, execute the pip command below.
29
+ ```
30
+ pip install -r requirement.txt
31
+ ```
32
+
33
+ ## Training
34
+
35
+ ### Input Features
36
+ The Kinetics I3D pre-trained feature of EGTEA dataset can be downloaded from [GDrive link](https://drive.google.com/drive/folders/1Zj1B2UZnjPgLrylhKOfu7m_9rkQFa14T?usp=sharing).
37
+ Files should be located in 'data/'.
38
+ You can get other features from the following links -
39
+ - [EPIC-Kitchen 100](https://github.com/happyharrycn/actionformer_release)
40
+ - [THUMOS'14](https://github.com/YHKimGithub/OAT-OSN/)
41
+ - [MUSES](https://songbai.site/muses/)
42
+
43
+ ### Config Files
44
+ The configuration files for EGTEA are already provided in the repository. For other datasets, they can be downloaded from [GDrive link](https://drive.google.com/drive/folders/19__GnM2HZCCDshED9kadsLNAI9XBvrFd?usp=sharing).
45
+
46
+ ### Training Model
47
+ To train the main HAT model, execute the command below.
48
+ ```
49
+ python main.py --mode=train --split=[split #]*
50
+
51
+ ```
52
+ ```
53
+ !python main.py --mode=train --batch_size=256 --epoch=1
54
+ ```
55
+ *If the dataset has any splits (e.g., EGTEA has 4 splits)
56
+
57
+ To train the post-processing network (OSN), execute the commands below.
58
+ ```
59
+ python supnet.py --mode=make --inference_subset=train --split=[split #]
60
+ python supnet.py --mode=make --inference_subset=test --split=[split #]
61
+ python supnet.py --mode=train --split=[split #]
62
+ ```
63
+
64
+
65
+ ## Testing
66
+ To test HAT, execute the command below.
67
+ ```
68
+ python main.py --mode=test --split=[split #]
69
+ ```
70
+
71
+ ```
72
+ !python main.py --mode=test --batch_size=256 --epoch=1
73
+
74
+ ```
75
+
76
+ ## Citing HAT
77
+ Please cite our paper in your publications if it helps your research:
78
+
79
+ ```BibTeX
80
+ @inproceedings{reza2022history,
81
+ title={HAT: History-Augmented Anchor Transformer for Online Temporal Action Localization},
82
+ author={Reza, Sakib and Zhang, Yuexi and Moghaddam, Mohsen and Camps, Octavia},
83
+ booktitle={European Conference on Computer Vision},
84
+ pages={XXX--XXX},
85
+ year={2024},
86
+ organization={Springer}
87
+ }
88
+ ```
89
+
90
+ ## Acknowledgment
91
+ This repository is created based on the repository of the baseline work [OAT-OSN](https://github.com/YHKimGithub/OAT-OSN/).