Tensorflow Implementation of ECCV'18 paper: Multimodal Human Motion Synthesis

Overview

MT-VAE for Multimodal Human Motion Synthesis

This is the code for ECCV 2018 paper MT-VAE: Learning Motion Transformations to Generate Multimodal Human Dynamics by Xinchen Yan, Akash Rastogi, Ruben Villegas, Kalyan Sunkavalli, Eli Shechtman, Sunil Hadap, Ersin Yumer, Honglak Lee.

Please follow the instructions to run the code.

Requirements

MT-VAE requires or works with

  • Mac OS X or Linux
  • NVIDIA GPU

Installing Dependency

Data Preprocessing

bash prep_human36m_joints.sh
  • Disclaimer: Please check the license of Human3.6M dataset if you download this preprocessed version.

Training (MT-VAE)

  • If you want to train the MT-VAE human motion generator, please run the following script (usually it takes 1 day with a single Titan GPU).
bash demo_human36m_trainMTVAE.sh
  • Alternatively, you can download the pre-trained MT-VAE model, please run the following script.
bash prep_human36m_model.sh

Motion Synthesis Using Pre-trained MT-VAE Model

  • Please run the following command to generate multiple diverse human motion given initial motion.
bash demo_human36m_inferMTVAE.sh

Motion Analogy-making Using Pre-trained MT-VAE Model

  • Please run the following command to execute motion analogy-making.
bash demo_human36m_analogyMTVAE.sh

Hierchical Video Synthesis Using Pre-trained Image Generation Model

CUDA_VISIBLE_DEVICE=0 python h36m_hierach_gensample.py
  • Disclaimer: Please double check the license in that repository and cite HierchVid paper when use.

Citation

If you find this useful, please cite our work as follows:

@inproceedings{yan2018mt,
  title={MT-VAE: Learning Motion Transformations to Generate Multimodal Human Dynamics},
  author={Yan, Xinchen and Rastogi, Akash and Villegas, Ruben and Sunkavalli, Kalyan and Shechtman, Eli and Hadap, Sunil and Yumer, Ersin and Lee, Honglak},
  booktitle={European Conference on Computer Vision},
  pages={276--293},
  year={2018},
  organization={Springer}
}

Acknowledgements

We would like to thank the amazing developers and the open-sourcing community. Our implementation has especially been benefited from the following excellent repositories:

You might also like...
EasyMocap is an open-source toolbox for markerless human motion capture from RGB videos.
EasyMocap is an open-source toolbox for markerless human motion capture from RGB videos.

EasyMocap is an open-source toolbox for markerless human motion capture from RGB videos. In this project, we provide the basic code for fitt

Synthesizing Long-Term 3D Human Motion and Interaction in 3D in CVPR2021

Long-term-Motion-in-3D-Scenes This is an implementation of the CVPR'21 paper "Synthesizing Long-Term 3D Human Motion and Interaction in 3D". Please ch

[WACV 2020] Reducing Footskate in Human Motion Reconstruction with Ground Contact Constraints

Reducing Footskate in Human Motion Reconstruction with Ground Contact Constraints Official implementation for Reducing Footskate in Human Motion Recon

A TensorFlow implementation of Neural Program Synthesis from Diverse Demonstration Videos

ViZDoom http://vizdoom.cs.put.edu.pl ViZDoom allows developing AI bots that play Doom using only the visual information (the screen buffer). It is pri

Liquid Warping GAN with Attention: A Unified Framework for Human Image Synthesis
Liquid Warping GAN with Attention: A Unified Framework for Human Image Synthesis

Liquid Warping GAN with Attention: A Unified Framework for Human Image Synthesis, including human motion imitation, appearance transfer, and novel view synthesis. Currently the paper is under review of IEEE TPAMI. It is an extension of our previous ICCV project impersonator, and it has a more powerful ability in generalization and produces higher-resolution results (512 x 512, 1024 x 1024) than the previous ICCV version.

Code and datasets for the paper
Code and datasets for the paper "Combining Events and Frames using Recurrent Asynchronous Multimodal Networks for Monocular Depth Prediction" (RA-L, 2021)

Combining Events and Frames using Recurrent Asynchronous Multimodal Networks for Monocular Depth Prediction This is the code for the paper Combining E

Preprocessed Datasets for our Multimodal NER paper

Unified Multimodal Transformer (UMT) for Multimodal Named Entity Recognition (MNER) Two MNER Datasets and Codes for our ACL'2020 paper: Improving Mult

The code repository for EMNLP 2021 paper
The code repository for EMNLP 2021 paper "Vision Guided Generative Pre-trained Language Models for Multimodal Abstractive Summarization".

Vision Guided Generative Pre-trained Language Models for Multimodal Abstractive Summarization [Paper] accepted at the EMNLP 2021: Vision Guided Genera

 the official code for ICRA 2021 Paper:
the official code for ICRA 2021 Paper: "Multimodal Scale Consistency and Awareness for Monocular Self-Supervised Depth Estimation"

G2S This is the official code for ICRA 2021 Paper: Multimodal Scale Consistency and Awareness for Monocular Self-Supervised Depth Estimation by Hemang

Comments
  • No corresponding file when to test

    No corresponding file when to test

    Hi, I met some problems when to execute your code. After downloading preprocessed dataset and pre-trained models, when I tried to generate multiple diverse human motion, I met IO error and there is no corresponding file in the preprocessed dataset.Do I need to download the whole Human3.6M dataset from the official website? Thanks!

    opened by shiwei6523 2
  • data processing steps

    data processing steps

    Hi, I am trying to replicate the code for the paper by myself to practice, I am thinking of extending your model. I am trying to understand the data pre processing that you've done. I would like it if you could may be explain the structure of it to me please ?

    Regards, Prashanth.

    opened by peacekurella 0
  • About h36m_input.py

    About h36m_input.py

    Hi, When I execute python h36m_gensample.py --model_version=MTVAE, happens an error from
    h36m_input.py .
    Error:FileNotFoundError: [Errno 2] No such file or directory: 'workspace/Human3.6M/annot_pts/770/vid_770_t000.csv' I am confused about the file annot_pts/770/vid_770_t000.csv. How can I generate these files? Could you help me? Thanks.😍

    opened by chenhaomingbob 0
  • Code for evaluation metrics

    Code for evaluation metrics

    Hi,

    Would you please let me know where can I find the code for S-MSE, R-MSE and Test CLL? In fact, I'm wondering how the MSE errors were computed? which time-step is reported? How joint errors were aggregated?

    Thanks.

    opened by sadegh-aa 0
Owner
Xinchen Yan
Computer Research Scientist; Prev @ Uber ATG Toronto, Google Brain, Google X Robotics, Adobe Research; Ph.D. @ Umich
Xinchen Yan
This repository contains the official implementation code of the paper Improving Multimodal Fusion with Hierarchical Mutual Information Maximization for Multimodal Sentiment Analysis, accepted at EMNLP 2021.

MultiModal-InfoMax This repository contains the official implementation code of the paper Improving Multimodal Fusion with Hierarchical Mutual Informa

Deep Cognition and Language Research (DeCLaRe) Lab 89 Dec 26, 2022
Exploring Versatile Prior for Human Motion via Motion Frequency Guidance (3DV2021)

Exploring Versatile Prior for Human Motion via Motion Frequency Guidance This is the codebase for video-based human motion reconstruction in human-mot

Jiachen Xu 5 Jul 14, 2022
This repository contains the code for the paper "Hierarchical Motion Understanding via Motion Programs"

Hierarchical Motion Understanding via Motion Programs (CVPR 2021) This repository contains the official implementation of: Hierarchical Motion Underst

Sumith Kulal 40 Dec 5, 2022
Official Pytorch implementation of the paper "MotionCLIP: Exposing Human Motion Generation to CLIP Space"

MotionCLIP Official Pytorch implementation of the paper "MotionCLIP: Exposing Human Motion Generation to CLIP Space". Please visit our webpage for mor

Guy Tevet 173 Dec 26, 2022
Code for ICCV 2021 paper "HuMoR: 3D Human Motion Model for Robust Pose Estimation"

Code for ICCV 2021 paper "HuMoR: 3D Human Motion Model for Robust Pose Estimation"

Davis Rempe 367 Dec 24, 2022
Implementation of "Generalizable Neural Performer: Learning Robust Radiance Fields for Human Novel View Synthesis"

Generalizable Neural Performer: Learning Robust Radiance Fields for Human Novel View Synthesis Abstract: This work targets at using a general deep lea

null 163 Dec 14, 2022
Tensorflow implementation of the paper "HumanGPS: Geodesic PreServing Feature for Dense Human Correspondences", CVPR 2021.

HumanGPS: Geodesic PreServing Feature for Dense Human Correspondences Tensorflow implementation of the paper "HumanGPS: Geodesic PreServing Feature fo

Google Interns 50 Dec 21, 2022
The official TensorFlow implementation of the paper Action Transformer: A Self-Attention Model for Short-Time Pose-Based Human Action Recognition

Action Transformer A Self-Attention Model for Short-Time Human Action Recognition This repository contains the official TensorFlow implementation of t

PIC4SeRCentre 20 Jan 3, 2023
PaddleRobotics is an open-source algorithm library for robots based on Paddle, including open-source parts such as human-robot interaction, complex motion control, environment perception, SLAM positioning, and navigation.

简体中文 | English PaddleRobotics paddleRobotics是基于paddle的机器人开源算法库集,包括人机交互、复杂运动控制、环境感知、slam定位导航等开源算法部分。 人机交互 主动多模交互技术TFVT-HRI 主动多模交互技术是通过视觉、语音、触摸传感器等输入机器人

null 185 Dec 26, 2022