Spaces:
Runtime error
Runtime error
| # 4DHumans: Reconstructing and Tracking Humans with Transformers | |
| Code repository for the paper: | |
| **Humans in 4D: Reconstructing and Tracking Humans with Transformers** | |
| [Shubham Goel](https://people.eecs.berkeley.edu/~shubham-goel/), [Georgios Pavlakos](https://geopavlakos.github.io/), [Jathushan Rajasegaran](http://people.eecs.berkeley.edu/~jathushan/), [Angjoo Kanazawa](https://people.eecs.berkeley.edu/~kanazawa/)<sup>\*</sup>, [Jitendra Malik](http://people.eecs.berkeley.edu/~malik/)<sup>\*</sup> | |
| arXiv preprint 2023 | |
| [[paper]()] [[project page](https://shubham-goel.github.io/4dhumans/)] [[hugging faces space]()] | |
|  | |
| ## Download dependencies | |
| Our demo code depends on [detectron2](https://github.com/facebookresearch/detectron2) to detect humans. | |
| To automatically download this dependency, clone this repo using `--recursive`, or run `git submodule update --init` if you've already cloned the repository. You should see the detectron2 source code at `vendor/detectron2`. | |
| ```bash | |
| git clone https://github.com/shubham-goel/4D-Humans.git --recursive | |
| # OR | |
| git clone https://github.com/shubham-goel/4D-Humans.git | |
| cd 4D-Humans | |
| git submodule update --init | |
| ``` | |
| ## Installation | |
| We recommend creating a clean [conda](https://docs.conda.io/) environment and installing all dependencies, as follows: | |
| ```bash | |
| conda env create -f environment.yml | |
| ``` | |
| After the installation is complete you can activate the conda environment by running: | |
| ``` | |
| conda activate 4D-humans | |
| ``` | |
| ## Download checkpoints and SMPL models | |
| To download the checkpoints and SMPL models, run | |
| ```bash | |
| ./fetch_data.sh | |
| ``` | |
| ## Run demo on images | |
| You may now run our demo to 3D reconstruct humans in images using the following command, which will run ViTDet and HMR2.0 on all images in the specified `--img_folder` and save renderings of the reconstructions in `--out_folder`. You can also use the `--side_view` flag to additionally render the side view of the reconstructed mesh. `--batch_size` batches the images together for faster processing. | |
| ```bash | |
| python demo.py \ | |
| --img_folder example_data/images \ | |
| --out_folder demo_out \ | |
| --batch_size=48 --side_view | |
| ``` | |
| ## Run demo on videos | |
| Coming soon. | |
| ## Training and evaluation | |
| Cmoing soon. | |
| ## Acknowledgements | |
| Parts of the code are taken or adapted from the following repos: | |
| - [ProHMR](https://github.com/nkolot/ProHMR) | |
| - [SPIN](https://github.com/nkolot/SPIN) | |
| - [SMPLify-X](https://github.com/vchoutas/smplify-x) | |
| - [HMR](https://github.com/akanazawa/hmr) | |
| - [ViTPose](https://github.com/ViTAE-Transformer/ViTPose) | |
| - [Detectron2](https://github.com/facebookresearch/detectron2) | |
| Additionally, we thank [StabilityAI](https://stability.ai/) for a generous compute grant that enabled this work. | |
| ## Citing | |
| If you find this code useful for your research, please consider citing the following paper: | |
| ``` | |
| @article{4DHUMANS, | |
| title={Humans in 4{D}: Reconstructing and Tracking Humans with Transformers}, | |
| author={Goel, Shubham and Pavlakos, Georgios and Rajasegaran, Jathushan and Kanazawa, Angjoo and Malik, Jitendra}, | |
| journal={arXiv preprint}, | |
| year={2023} | |
| } | |
| ``` | |