| # Installation | |
| **Step 1.** Install PyTorch=2.0.1, Python=3.10.12 | |
| ``` | |
| conda create -n opentad python=3.10.12 | |
| source activate opentad | |
| conda install pytorch=2.0.1 torchvision=0.15.2 pytorch-cuda=11.8 -c pytorch -c nvidia | |
| ``` | |
| **Step 2.** Install mmaction2 for end-to-end training | |
| ``` | |
| pip install openmim | |
| mim install mmcv==2.0.1 | |
| mim install mmaction2==1.1.0 | |
| ``` | |
| **Step 3.** Install OpenTAD | |
| ``` | |
| git clone git@github.com:sming256/OpenTAD.git | |
| cd OpenTAD | |
| pip install -r requirements.txt | |
| ``` | |
| The code is tested with Python 3.10.12, PyTorch 2.0.1, CUDA 11.8, and gcc 11.3.0 on Ubuntu 20.04, other versions might also work. | |
| **Step 4.** Prepare the annotation and data. | |
| | Dataset | Description | | |
| | :--------------------------------------------------------- | :-------------------------------------------------------------------------------------------- | | |
| | [ActivityNet](/tools/prepare_data/activitynet/README.md) | A Large-Scale Video Benchmark for Human Activity Understanding with 19,994 videos. | | |
| | [THUMOS14](/tools/prepare_data/thumos/README.md) | Consists of 413 videos with temporal annotations. | | |
| | [EPIC-KITCHENS](/tools/prepare_data/epic/README.md) | Large-scale dataset in first-person (egocentric) vision. Latest version is EPIC-KITCHENS-100. | | |
| | [Ego4D-MQ](/tools/prepare_data/ego4d/README.md) | Ego4D is the world's largest egocentric video dataset. MQ refers to its moment query task. | | |
| | [HACS](/tools/prepare_data/hacs/README.md) | The same action taxonomy with ActivityNet, but consists of around 50K videos. | | |
| | [FineAction](/tools/prepare_data/fineaction/README.md) | Contains 103K temporal instances of 106 action categories, annotated in 17K untrimmed videos. | | |
| | [Multi-THUMOS](/tools/prepare_data/multi-thumos/README.md) | Dense, multilabel action annotations of THUMOS14. | | |
| | [Charades](/tools/prepare_data/charades/README.md) | Contains dense-labeled 9,848 annotated videos of daily activities. | | |
| PS: If you meet `FileNotFoundError: [Errno 2] No such file or directory: 'xxx/missing_files.txt'` | |
| - It means you may need to generate a `missing_files.txt`, which should record the missing features compared to all the videos in the annotation files. You can use `python tools/prepare_data/generate_missing_list.py annotation.json feature_folder` to generate the txt file. | |
| - eg. `python tools/prepare_data/generate_missing_list.py data/fineaction/annotations/annotations_gt.json data/fineaction/features/fineaction_mae_g` | |
| - In the provided feature from this codebase, we have already included this txt in the zip file. |