Pose-to-Body

also see Video Synthesis

Types

A. Train a uniform generator for different videos. Reference appearance frame and pose/motion are input as condition during inference.
B. Train a generator on frames of same video. The appearance is encoded in the model. Only pose/motion input is required during inference.
Generally, type B seems give nicer result with less artifact, but need to re-train a new model for each traget person. Used in motion re-target papers.

  • vid2vid
  • Deep Video-Based Performance Cloning
  • Everybody Dance Now

PoseWarp

Synthesizing images of humans in unseen poses (CVPR 2018)
CVPR 2018 Oral | Keras
Source Image Segmentation + Spatial Transformation + Foreground Synthesis + Background Synthesis

Pose Guided Human Video Generation

Pose Guided Human Video Generation (ECCV 2018)

Deep Video-Based Performance Cloning

Deep Video-Based Performance Cloning (Eurographics 2019)

Progressive Pose Attention Transfer for Person Image Generation

Progressive Pose Attention Transfer for Person Image Generation (CVPR 2019)
PyTorch 0.3.1 or 1.0

Everybody Dance Now

EDN (ICCV 2019) skeleton-to-rendering part https://carolineec.github.io/everybody_dance_now/images/teaser.png