| # 4DHumans: Reconstructing and Tracking Humans with Transformers |
| Code repository for the paper: |
| **Humans in 4D: Reconstructing and Tracking Humans with Transformers** |
| [Shubham Goel](https://people.eecs.berkeley.edu/~shubham-goel/), [Georgios Pavlakos](https://geopavlakos.github.io/), [Jathushan Rajasegaran](http://people.eecs.berkeley.edu/~jathushan/), [Angjoo Kanazawa](https://people.eecs.berkeley.edu/~kanazawa/)<sup>\*</sup>, [Jitendra Malik](http://people.eecs.berkeley.edu/~malik/)<sup>\*</sup> |
| arXiv preprint 2023 |
| [[paper]()] [[project page](https://shubham-goel.github.io/4dhumans/)] [[hugging faces space]()] |
|
|
|  |
|
|
| ## Download dependencies |
| Our demo code depends on [detectron2](https://github.com/facebookresearch/detectron2) to detect humans. |
| To automatically download this dependency, clone this repo using `--recursive`, or run `git submodule update --init` if you've already cloned the repository. You should see the detectron2 source code at `vendor/detectron2`. |
| ```bash |
| git clone https://github.com/shubham-goel/4D-Humans.git --recursive |
| # OR |
| git clone https://github.com/shubham-goel/4D-Humans.git |
| cd 4D-Humans |
| git submodule update --init |
| ``` |
|
|
| ## Installation |
| We recommend creating a clean [conda](https://docs.conda.io/) environment and installing all dependencies, as follows: |
| ```bash |
| conda env create -f environment.yml |
| ``` |
|
|
| After the installation is complete you can activate the conda environment by running: |
| ``` |
| conda activate 4D-humans |
| ``` |
|
|
| ## Download checkpoints and SMPL models |
| To download the checkpoints and SMPL models, run |
| ```bash |
| ./fetch_data.sh |
| ``` |
|
|
| ## Run demo on images |
| You may now run our demo to 3D reconstruct humans in images using the following command, which will run ViTDet and HMR2.0 on all images in the specified `--img_folder` and save renderings of the reconstructions in `--out_folder`. You can also use the `--side_view` flag to additionally render the side view of the reconstructed mesh. `--batch_size` batches the images together for faster processing. |
| ```bash |
| python demo.py \ |
| --img_folder example_data/images \ |
| --out_folder demo_out \ |
| --batch_size=48 --side_view |
| ``` |
|
|
| ## Run demo on videos |
| Coming soon. |
|
|
| ## Training and evaluation |
| Cmoing soon. |
|
|
| ## Acknowledgements |
| Parts of the code are taken or adapted from the following repos: |
| - [ProHMR](https://github.com/nkolot/ProHMR) |
| - [SPIN](https://github.com/nkolot/SPIN) |
| - [SMPLify-X](https://github.com/vchoutas/smplify-x) |
| - [HMR](https://github.com/akanazawa/hmr) |
| - [ViTPose](https://github.com/ViTAE-Transformer/ViTPose) |
| - [Detectron2](https://github.com/facebookresearch/detectron2) |
|
|
| Additionally, we thank [StabilityAI](https://stability.ai/) for a generous compute grant that enabled this work. |
|
|
| ## Citing |
| If you find this code useful for your research, please consider citing the following paper: |
|
|
| ``` |
| @article{4DHUMANS, |
| title={Humans in 4{D}: Reconstructing and Tracking Humans with Transformers}, |
| author={Goel, Shubham and Pavlakos, Georgios and Rajasegaran, Jathushan and Kanazawa, Angjoo and Malik, Jitendra}, |
| journal={arXiv preprint}, |
| year={2023} |
| } |
| ``` |
|
|