PyTorch implementation of CloudWalk’s recent paper DenseBody.
Note: For most recent updates, please check out the dev
branch.
Update on 20190613 A toy dataset has been released to facilitate the reproduction of this project. checkout PREPS.md
for details.
Update on 20190826 A pre-trained model (Encoder/Decoder) has been released to facilitate the reproduction of this project.
Here is the reproduction result (left: input image; middle: ground truth UV position map; right: estimated UV position map)
PREPS.md
for details.data_utils/UV_map_generator.py
for more details.Please follow the instructions PREPS.md
to prepare your training dataset and UV maps. Then run train.sh
or nohup_train.sh
to begin training.
To train with your own UV map, checkout UV_MAPS.md
for detailed instructions.
To explore different network architectures, checkout NETWORKS.md
for detailed instructions.
Lingbo Yang(Lotayou): The owner and maintainer of this repo.
Raj Advani(radvani): Provide several hand-crafted UV maps and many constructive feedbacks.
Please consider citing the following paper if you find this project useful.
DenseBody: Directly Regressing Dense 3D Human Pose and Shape From a Single Color Image
The network training part is inspired by BicycleGAN