| # Deployment | |
| We provide support to some popular deployment tools. This part is built upon the implementation of [YOLOX Deployment](https://github.com/Megvii-BaseDetection/YOLOX/tree/main/demo) and [the adaptation by ByteTrack](https://github.com/ifzhang/ByteTrack/tree/main/deploy). | |
| ## ONNX support | |
| 1. convert the pytorch model to onnx checkpoints, we provide an example here. | |
| ```python | |
| # In pratice you may want smaller model for faster inference. | |
| python deploy/scripts/export_onnx.py --output-name ocsort.onnx -f exps/example/mot/yolox_x_mix_det.py -c pretrained/bytetrack_x_mot17.pth.tar | |
| ``` | |
| 2. run on the provided model video by | |
| ```shell | |
| cd $OCSORT_HOME/deploy/ONNXRuntime | |
| python onnx_inference.py | |
| ``` | |
| ## TensorRT support (Python) | |
| 1. Follow [TensorRT Installation Guide](https://docs.nvidia.com/deeplearning/tensorrt/install-guide/index.html) and [torch2trt](https://github.com/NVIDIA-AI-IOT/torch2trt) to install TensorRT (Version 7 recommended) and torch2trt. | |
| 2. Convert Model | |
| ```python | |
| # you have to download checkpoint bytetrack_s_mot17.pth.tar from model zoo of ByteTrack | |
| python3 deploy/scripts/trt.py -f exps/example/mot/yolox_s_mix_det.py -c pretrained/bytetrack_s_mot17.pth.tar | |
| ``` | |
| 3. Run on a demo video | |
| ```python | |
| python3 tools/demo_track.py video -f exps/example/mot/yolox_s_mix_det.py --trt --save_result | |
| ``` | |
| *Note: We haven't validated the C++ support for TensorRT yet, please refer to [ByteTrack guidance](https://github.com/ifzhang/ByteTrack/tree/main/deploy/TensorRT/cpp) for adaptation for now.* | |
| ## ncnn support | |
| Please follow the [guidelines](https://github.com/ifzhang/ByteTrack/tree/main/deploy/ncnn/cpp) from ByteTrack to deploy by support from ncnn. |