AMASS is a large database of human motion unifying different optical marker-based motion capture datasets by representing them within a common framework and parameterization. AMASS is readily useful for animation, visualization, and generating training data for deep learning.
Here we provide tools and tutorials to use AMASS in your research projects. More specifically:
- Following the recommended splits of data by AMASS, we provide three non-overlapping train/validation/test splits.
- AMASS uses an extended version of SMPL+H with DMPLs. Here we show how to load different components and visualize a body model with AMASS data.
- AMASS is also compatible with SMPL and SMPL-X body models. We show how to use the body data from AMASS to animate these models.
Requirements
- Python 3.7
- PyTorch 1.7.1
- Human Body Prior
- Pyrender for visualizations
Clone this repo and run the following from the root folder:
python install -r requirements.txt
python setup.py develop
AMASS uses MoSh++ pipeline to fit SMPL+H body model to human optical marker based motion capture (mocap) data. In the paper we use SMPL+H with extended shape space, i.e. 16 betas, and 8 DMPLs. Please download models and place them them in body_models folder of this repository after you obtained the code from GitHub.
We release tools and Jupyter notebooks to demonstrate how to use AMASS to animate SMPL+H body model.
Furthermore, as promised in the supplementary material of the paper, we release code to produce synthetic mocap using DFaust registrations.
Please refer to tutorials for further details.
Please cite the following paper if you use this code directly or indirectly in your research/projects:
@inproceedings{AMASS:2019,
title={AMASS: Archive of Motion Capture as Surface Shapes},
author={Mahmood, Naureen and Ghorbani, Nima and F. Troje, Nikolaus and Pons-Moll, Gerard and Black, Michael J.},
booktitle = {The IEEE International Conference on Computer Vision (ICCV)},
year={2019},
month = {Oct},
url = {https://amass.is.tue.mpg.de},
month_numeric = {10}
}
Software Copyright License for non-commercial scientific research purposes. Please read carefully the terms and conditions and any accompanying documentation before you download and/or use the AMASS dataset, and software, (the "Model & Software"). By downloading and/or using the Model & Software (including downloading, cloning, installing, and any other use of this GitHub repository), you acknowledge that you have read these terms and conditions, understand them, and agree to be bound by them. If you do not agree with these terms and conditions, you must not download and/or use the Model & Software. Any infringement of the terms of this agreement will automatically terminate your rights under this License.
The code in this repository is developed by Nima Ghorbani.
If you have any questions you can contact us at [email protected].
For commercial licensing, please contact [email protected]
To find out about the latest developments stay tuned to AMASS twitter.
The research community needs more human motion data. If you have interesting marker based motion capture data, and you are willing to share it for research purposes, then we will label and clean your mocap and MoSh it for you and add it to the AMASS dataset, naturally citing you as the original owner of the marker data. For this purposes feel free to contact [email protected].