constanceCM's picture
upload
8aa674c
# Installation
**Step 1.** Install PyTorch=2.0.1, Python=3.10.12
```
conda create -n opentad python=3.10.12
source activate opentad
conda install pytorch=2.0.1 torchvision=0.15.2 pytorch-cuda=11.8 -c pytorch -c nvidia
```
**Step 2.** Install mmaction2 for end-to-end training
```
pip install openmim
mim install mmcv==2.0.1
mim install mmaction2==1.1.0
```
**Step 3.** Install OpenTAD
```
git clone git@github.com:sming256/OpenTAD.git
cd OpenTAD
pip install -r requirements.txt
```
The code is tested with Python 3.10.12, PyTorch 2.0.1, CUDA 11.8, and gcc 11.3.0 on Ubuntu 20.04, other versions might also work.
**Step 4.** Prepare the annotation and data.
| Dataset | Description |
| :--------------------------------------------------------- | :-------------------------------------------------------------------------------------------- |
| [ActivityNet](/tools/prepare_data/activitynet/README.md) | A Large-Scale Video Benchmark for Human Activity Understanding with 19,994 videos. |
| [THUMOS14](/tools/prepare_data/thumos/README.md) | Consists of 413 videos with temporal annotations. |
| [EPIC-KITCHENS](/tools/prepare_data/epic/README.md) | Large-scale dataset in first-person (egocentric) vision. Latest version is EPIC-KITCHENS-100. |
| [Ego4D-MQ](/tools/prepare_data/ego4d/README.md) | Ego4D is the world's largest egocentric video dataset. MQ refers to its moment query task. |
| [HACS](/tools/prepare_data/hacs/README.md) | The same action taxonomy with ActivityNet, but consists of around 50K videos. |
| [FineAction](/tools/prepare_data/fineaction/README.md) | Contains 103K temporal instances of 106 action categories, annotated in 17K untrimmed videos. |
| [Multi-THUMOS](/tools/prepare_data/multi-thumos/README.md) | Dense, multilabel action annotations of THUMOS14. |
| [Charades](/tools/prepare_data/charades/README.md) | Contains dense-labeled 9,848 annotated videos of daily activities. |
PS: If you meet `FileNotFoundError: [Errno 2] No such file or directory: 'xxx/missing_files.txt'`
- It means you may need to generate a `missing_files.txt`, which should record the missing features compared to all the videos in the annotation files. You can use `python tools/prepare_data/generate_missing_list.py annotation.json feature_folder` to generate the txt file.
- eg. `python tools/prepare_data/generate_missing_list.py data/fineaction/annotations/annotations_gt.json data/fineaction/features/fineaction_mae_g`
- In the provided feature from this codebase, we have already included this txt in the zip file.