Everybody Dance Now
Caroline Chan
and
Shiry Ginosar
and
Tinghui Zhou
and
Alexei A. Efros
arXiv e-Print archive - 2018 via Local arXiv
Keywords:
cs.GR, cs.CV
First published: 2018/08/22 (6 years ago) Abstract: This paper presents a simple method for "do as I do" motion transfer: given a
source video of a person dancing we can transfer that performance to a novel
(amateur) target after only a few minutes of the target subject performing
standard moves. We pose this problem as a per-frame image-to-image translation
with spatio-temporal smoothing. Using pose detections as an intermediate
representation between source and target, we learn a mapping from pose images
to a target subject's appearance. We adapt this setup for temporally coherent
video generation including realistic face synthesis. Our video demo can be
found at https://youtu.be/PCBTZh41Ris .