Darknsu commited on
Commit
d1fce63
·
verified ·
1 Parent(s): 9de6411

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -91
README.md CHANGED
@@ -1,91 +1,6 @@
1
-
2
- # HAT: History-Augmented Anchor Transformer for Online Temporal Action Localization (ECCV 2024)
3
- ### Sakib Reza, Yuexi Zhang, Mohsen Moghaddam, Octavia Camps
4
- #### Northeastern University, Boston, United States
5
- {reza.s,zhang.yuex,mohsen,o.camps}@northeastern.edu
6
-
7
- ## [Arxiv Preprint](https://arxiv.org/abs/2408.06437)
8
-
9
-
10
- ## Updates
11
- - Aug 22, 2024 - EGTEA pre-extracted features and config files for other datasets added
12
- - Aug 14, 2024 - Arxiv preprint added
13
- - July 7, 2024 - initial code release
14
-
15
- ## Installation
16
-
17
- ### Prerequisites
18
- - Ubuntu 20.04
19
- - Python 3.10.9
20
- - CUDA 12.0
21
-
22
- ### Requirements
23
- - pytorch==2.0.0
24
- - numpy==1.23.5
25
- - h5py==3.9.0
26
- - ...
27
-
28
- To install all required libraries, execute the pip command below.
29
- ```
30
- pip install -r requirement.txt
31
- ```
32
-
33
- ## Training
34
-
35
- ### Input Features
36
- The Kinetics I3D pre-trained feature of EGTEA dataset can be downloaded from [GDrive link](https://drive.google.com/drive/folders/1Zj1B2UZnjPgLrylhKOfu7m_9rkQFa14T?usp=sharing).
37
- Files should be located in 'data/'.
38
- You can get other features from the following links -
39
- - [EPIC-Kitchen 100](https://github.com/happyharrycn/actionformer_release)
40
- - [THUMOS'14](https://github.com/YHKimGithub/OAT-OSN/)
41
- - [MUSES](https://songbai.site/muses/)
42
-
43
- ### Config Files
44
- The configuration files for EGTEA are already provided in the repository. For other datasets, they can be downloaded from [GDrive link](https://drive.google.com/drive/folders/19__GnM2HZCCDshED9kadsLNAI9XBvrFd?usp=sharing).
45
-
46
- ### Training Model
47
- To train the main HAT model, execute the command below.
48
- ```
49
- python main.py --mode=train --split=[split #]*
50
-
51
- ```
52
- ```
53
- !python main.py --mode=train --batch_size=256 --epoch=1
54
- ```
55
- *If the dataset has any splits (e.g., EGTEA has 4 splits)
56
-
57
- To train the post-processing network (OSN), execute the commands below.
58
- ```
59
- python supnet.py --mode=make --inference_subset=train --split=[split #]
60
- python supnet.py --mode=make --inference_subset=test --split=[split #]
61
- python supnet.py --mode=train --split=[split #]
62
- ```
63
-
64
-
65
- ## Testing
66
- To test HAT, execute the command below.
67
- ```
68
- python main.py --mode=test --split=[split #]
69
- ```
70
-
71
- ```
72
- !python main.py --mode=test --batch_size=256 --epoch=1
73
-
74
- ```
75
-
76
- ## Citing HAT
77
- Please cite our paper in your publications if it helps your research:
78
-
79
- ```BibTeX
80
- @inproceedings{reza2022history,
81
- title={HAT: History-Augmented Anchor Transformer for Online Temporal Action Localization},
82
- author={Reza, Sakib and Zhang, Yuexi and Moghaddam, Mohsen and Camps, Octavia},
83
- booktitle={European Conference on Computer Vision},
84
- pages={XXX--XXX},
85
- year={2024},
86
- organization={Springer}
87
- }
88
- ```
89
-
90
- ## Acknowledgment
91
- This repository is created based on the repository of the baseline work [OAT-OSN](https://github.com/YHKimGithub/OAT-OSN/).
 
1
+ FROM python:3.9
2
+ WORKDIR /app
3
+ COPY . .
4
+ RUN apt-get update && apt-get install -y ffmpeg
5
+ RUN pip install -r requirements.txt
6
+ CMD ["python", "main.py", "--gradio", "True"]