Need help with amass?
Click the “chat” button below for chat support from the developer who created it, or find similar developers for support.

About the developer

nghorbani
249 Stars 40 Forks Other 35 Commits 8 Opened issues

Description

Data preparation and loader for AMASS

Services available

!
?

Need anything else?

Contributors list

# 246,804
Python
Jupyter...
vae
action-...
10 commits

AMASS: Archive of Motion Capture as Surface Shapes

alt text

AMASS is a large database of human motion unifying different optical marker-based motion capture datasets by representing them within a common framework and parameterization. AMASS is readily useful for animation, visualization, and generating training data for deep learning.

Here we provide tools and tutorials to use AMASS in your research projects. More specifically: - Following the recommended splits of data by AMASS, we provide three non-overlapping train/validation/test splits. - AMASS uses an extended version of SMPL+H with DMPLs. Here we show how to load different components and visualize a body model with AMASS data. - AMASS is also compatible with SMPL and SMPL-X body models. We show how to use the body data from AMASS to animate these models.

Table of Contents

Installation

Requirements - Python 3.7 - PyTorch 1.7.1 - Human Body Prior - Pyrender for visualizations

Clone this repo and run the following from the root folder:

bash
python install -r requirements.txt
python setup.py develop

Body Models

AMASS uses MoSh++ pipeline to fit SMPL+H body model to human optical marker based motion capture (mocap) data. In the paper we use SMPL+H with extended shape space, i.e. 16 betas, and 8 DMPLs. Please download models and place them them in body_models folder of this repository after you obtained the code from GitHub.

Tutorials

We release tools and Jupyter notebooks to demonstrate how to use AMASS to animate SMPL+H body model.

Furthermore, as promised in the supplementary material of the paper, we release code to produce synthetic mocap using DFaust registrations.

Please refer to tutorials for further details.

Citation

Please cite the following paper if you use this code directly or indirectly in your research/projects:

@inproceedings{AMASS:2019,
  title={AMASS: Archive of Motion Capture as Surface Shapes},
  author={Mahmood, Naureen and Ghorbani, Nima and F. Troje, Nikolaus and Pons-Moll, Gerard and Black, Michael J.},
  booktitle = {The IEEE International Conference on Computer Vision (ICCV)},
  year={2019},
  month = {Oct},
  url = {https://amass.is.tue.mpg.de},
  month_numeric = {10}
}

License

Software Copyright License for non-commercial scientific research purposes. Please read carefully the terms and conditions and any accompanying documentation before you download and/or use the AMASS dataset, and software, (the "Model & Software"). By downloading and/or using the Model & Software (including downloading, cloning, installing, and any other use of this GitHub repository), you acknowledge that you have read these terms and conditions, understand them, and agree to be bound by them. If you do not agree with these terms and conditions, you must not download and/or use the Model & Software. Any infringement of the terms of this agreement will automatically terminate your rights under this License.

## Contact The code in this repository is developed by Nima Ghorbani.

If you have any questions you can contact us at [email protected].

For commercial licensing, please contact [email protected]

To find out about the latest developments stay tuned to AMASS twitter.

Contribute to AMASS

The research community needs more human motion data. If you have interesting marker based motion capture data, and you are willing to share it for research purposes, then we will label and clean your mocap and MoSh it for you and add it to the AMASS dataset, naturally citing you as the original owner of the marker data. For this purposes feel free to contact [email protected].

We use cookies. If you continue to browse the site, you agree to the use of cookies. For more information on our use of cookies please see our Privacy Policy.