ABSTRACT
Child characters are commonly seen in leading roles in top-selling video games. Previous studies have shown that child motions are perceptually and stylistically different from those of adults. Creating motion for these characters by motion capturing children is uniquely challenging because of confusion, lack of patience and regulations. Retargeting adult motion, which is much easier to record, onto child skeletons, does not capture the stylistic differences. In this paper, we propose that style translation is an effective way to transform adult motion capture data to the style of child motion. Our method is based on CycleGAN, which allows training on a relatively small number of sequences of child and adult motions that do not even need to be temporally aligned. Our adult2child network converts short sequences of motions called motion words from one domain to the other. The network was trained using a motion capture database collected by our team containing 23 locomotion and exercise motions. We conducted a perception study to evaluate the success of style translation algorithms, including our algorithm and recently presented style translation neural networks. Results show that the translated adult motions are recognized as child motions significantly more often than adult motions.
Supplemental Material
- Kfir Aberman, Yijia Weng, Dani Lischinski, Daniel Cohen-Or, and Baoquan Chen. 2020. Unpaired Motion Style Transfer from Video to Animation. ACM Trans. Graph. 39, 4 (July 2020).Google ScholarDigital Library
- Aishat Aloba. 2019. Tailoring Motion Recognition Systems to Children’s Motions. In 2019 International Conference on Multimodal Interaction (Suzhou, China) (ICMI ’19). ACM, NY, USA, 457–462.Google ScholarDigital Library
- Aishat Aloba, Gianne Flores, Julia Woodward, Alex Shaw, Amanda Castonguay, Isabella Cuba, Yuzhu Dong, Eakta Jain, and Lisa Anthony. 2018. Kinder-Gator: The UF Kinect Database of Child and Adult Motion. In Proceedings of the 39th Annual European Association for Computer Graphics Conference: Short Papers (Delft, The Netherlands) (EG). Eurographics Association, 13–16.Google Scholar
- Kenji Amaya, Armin Bruderlin, and Tom Calvert. 1996. Emotion from Motion. In Proceedings of the Conference on Graphics Interface ’96 (Toronto, Ontario, Canada) (GI ’96). Canadian Info. Proc. Society, CAN, 222–229.Google ScholarDigital Library
- Andreas Aristidou, Daniel Cohen-Or, Jessica K. Hodgins, Yiorgos Chrysanthou, and Ariel Shamir. 2018b. Deep Motifs and Motion Signatures. ACM Trans. Graph. 37, 6, Article 187 (Nov. 2018), 13 pages.Google ScholarDigital Library
- Andreas Aristidou, Daniel Cohen-Or, Jessica K. Hodgins, and Ariel Shamir. 2018a. Self-similarity Analysis for Motion Capture Cleaning. Comput. Graph. Forum 37, 2 (May 2018), 297–309.Google ScholarCross Ref
- Andreas Aristidou, Qiong Zeng, Efstathios Stavrakis, Kangkang Yin, Daniel Cohen-Or, Yiorgos Chrysanthou, and Baoquan Chen. 2017. Emotion Control of Unstructured Dance Movements. In Proceedings of the ACM SIGGRAPH / Eurographics Symposium on Computer Animation (Los Angeles, California) (SCA ’17). ACM, NY, USA, Article 9, 10 pages.Google ScholarDigital Library
- Emad Barsoum, John Kender, and Zicheng Liu. 2018. HP-GAN: Probabilistic 3D Human Motion Prediction via GAN. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (Salt Lake City, UT, USA). IEEE Computer Society, 1418–1427.Google ScholarCross Ref
- Matthew Brand and Aaron Hertzmann. 2000. Style Machines. In Proceedings of the 27th Annual Conference on Computer Graphics and Interactive Techniques(SIGGRAPH ’00). ACM Press/Addison-Wesley Publishing Co., USA, 183–192.Google ScholarDigital Library
- Kwang-Jin Choi and Hyeong-Seok Ko. 1999. On-Line Motion Retargetting. In Proceedings of the 7th Pacific Conference on Computer Graphics and Applications(PG ’99). IEEE Computer Society, USA, 32.Google ScholarDigital Library
- CMU. 2020. Carnegie Mellon University MoCap Database: http://mocap.cs.cmu.edu/. [Online; Retrieved July, 2020].Google Scholar
- Yuzhu Dong, Aishat Aloba, Lisa Anthony, and Eakta Jain. 2018. Style Translation to Create Child-like Motion. In Proceedings of the 39th Annual European Association for Computer Graphics Conference: Posters (Delft, The Netherlands) (EG ’18). Eurographics Association, Goslar, DEU, 31–32.Google ScholarDigital Library
- Yuzhu Dong, Aishat Aloba, Sachin Paryani, Lisa Anthony, Neha Rana, and Eakta Jain. 2017. Adult2Child: Dynamic Scaling Laws to Create Child-like Motion. In Proceedings of the Tenth International Conference on Motion in Games (Barcelona, Spain) (MIG ’17). ACM, NY, USA, Article 13.Google ScholarDigital Library
- Han Du, Erik Herrmann, Janis Sprenger, Noshaba Cheema, Somayeh Hosseini, Klaus Fischer, and Philipp Slusallek. 2019a. Stylistic Locomotion Modeling with Conditional Variational Autoencoder. In 40th Annual Conference of the European Association for Computer Graphics, Eurographics 2019(Genoa, Italy), Paolo Cignoni and Eder Miguel (Eds.). The Eurographics Association, 9–12.Google Scholar
- Han Du, Erik Herrmann, Janis Sprenger, Klaus Fischer, and Philipp Slusallek. 2019b. Stylistic Locomotion Modeling and Synthesis Using Variational Generative Models. In Motion, Interaction and Games (Newcastle upon Tyne, United Kingdom) (MIG ’19). ACM, NY, USA, Article 32, 10 pages.Google Scholar
- Leon A. Gatys, Alexander S. Ecker, and Matthias Bethge. 2016. Image Style Transfer Using Convolutional Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition(CVPR’16). 2414–2423.Google ScholarCross Ref
- Michael Gleicher. 1998. Retargetting Motion to New Characters. In Proceedings of the 25th Annual Conference on Computer Graphics and Interactive Techniques(SIGGRAPH ’98). ACM, NY, USA, 33–42.Google ScholarDigital Library
- Gutemberg Guerra-Filho and Arnab Biswas. 2012. The Human Motion Database: A Cognitive and Parametric Sampling of Human Motion. Image Vision Comput. 30, 3 (March 2012), 251–261.Google ScholarDigital Library
- Chris Hecker, Bernd Raabe, Ryan W. Enslow, John DeWeese, Jordan Maynard, and Kees van Prooijen. 2008. Real-Time Motion Retargeting to Highly Varied User-Created Morphologies. ACM Trans. Graph. 27, 3 (Aug. 2008), 1–11.Google ScholarDigital Library
- Jessica K. Hodgins and Nancy S. Pollard. 1997. Adapting Simulated Behaviors for New Characters. In Proceedings of the 24th Annual Conference on Computer Graphics and Interactive Techniques(SIGGRAPH ’97). ACM Press/Addison-Wesley Publishing Co., USA, 153–162.Google Scholar
- Daniel Holden, Taku Komura, and Jun Saito. 2017. Phase-Functioned Neural Networks for Character Control. ACM Trans. Graph. 36, 4, Article 42 (July 2017).Google ScholarDigital Library
- Daniel Holden, Jun Saito, and Taku Komura. 2016. A Deep Learning Framework for Character Motion Synthesis and Editing. ACM Trans. Graph. 35, 4, Article 138 (July 2016).Google ScholarDigital Library
- Daniel Holden, Jun Saito, Taku Komura, and Thomas Joyce. 2015. Learning Motion Manifolds with Convolutional Autoencoders. In SIGGRAPH Asia 2015 Technical Briefs (Kobe, Japan) (SA ’15). ACM, NY, USA, Article 18.Google Scholar
- Marijana Hraski, Željko Hraski, and Ivan Prskalo. 2015. Comparison of standing long jump technique performed by subjects from different age groups. Baltic Journal of Sport and Health Sciences 98, 3 (2015), 2.Google ScholarCross Ref
- Eugene Hsu, Kari Pulli, and Jovan Popović. 2005. Style Translation for Human Motion. ACM Trans. Graph. 24, 3 (July 2005), 1082–1089.Google ScholarDigital Library
- Donald F. Huelke. 1998. An overview of anatomical considerations of infants and children in the adult world of automobile safety design. Annual Proceedings / Association for the Advancement of Automotive Medicine 42(1998), 93–113.Google Scholar
- Leslie Ikemoto, Okan Arikan, and David Forsyth. 2009. Generalizing Motion Edits with Gaussian Processes. ACM Trans. Graph. 28, 1, Article 1 (Feb. 2009).Google ScholarDigital Library
- Phillip Isola, Jun-Yan Zhu, Tinghui Zhou, and Alexei A Efros. 2017. Image-to-Image Translation with Conditional Adversarial Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition(CVPR’17). 1125–1134.Google ScholarCross Ref
- Eakta Jain, Lisa Anthony, Aishat Aloba, Amanda Castonguay, Isabella Cuba, Alex Shaw, and Julia Woodward. 2016. Is the Motion of a Child Perceivably Different from the Motion of an Adult?ACM Trans. Appl. Percept. 13, 4, Article 22 (July 2016).Google Scholar
- Joseph Johnson. 2018. Hours children spend gaming weekly in the UK from 2013 to 2017, by age group.https://www.statista.com/statistics/274434/time-spent-gaming-weekly-among-children-in-the-uk-by-age [Online; Retrieved July, 2020].Google Scholar
- Wanli Ma, Shihong Xia, Jessica K. Hodgins, Xiao Yang, Chunpeng Li, and Zhaoqi Wang. 2010. Modeling Style and Variation in Human Motion. In Proceedings of the 2010 ACM SIGGRAPH/Eurographics Symposium on Computer Animation (Madrid, Spain) (SCA ’10). Eurographics Association, 21–30.Google ScholarDigital Library
- Xudong Mao, Qing Li, Haoran Xie, Raymond YK Lau, Zhen Wang, and Stephen Paul Smolley. 2017. Least squares generative adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision(ICCV’17). 2794–2802.Google ScholarCross Ref
- Ian Mason, Sebastian Starke, He Zhang, Hakan Bilen, and Taku Komura. 2018. Few-shot Learning of Homogeneous Human Locomotion Styles. Comput. Graph. Forum 37, 7 (2018), 143–153.Google ScholarCross Ref
- Alberto Menache. 2000. Understanding motion capture for computer animation and video games. Morgan kaufmann.Google Scholar
- Philip R. Nader, Robert H. Bradley, Renate M. Houts, Susan L. McRitchie, and Marion O’Brien. 2008. Moderate-to-vigorous physical activity from ages 9 to 15 years. JAMA 300, 3 (2008), 295–305.Google ScholarCross Ref
- OSU. 2020. Ohio State University MoCap Database https://accad.osu.edu/research/motion-lab/mocap-system-and-data. [Online; Retrieved July, 2020].Google Scholar
- Jean Piaget. 2015. The Grasp of Consciousness (Psychology Revivals): Action and Concept in the Young Child. Psychology Press.Google Scholar
- Marc H. Raibert and Jessica K. Hodgins. 1991. Animation of Dynamic Legged Locomotion. In Proceedings of the 18th Annual Conference on Computer Graphics and Interactive Techniques(SIGGRAPH ’91). ACM, NY, USA, 349–358.Google ScholarDigital Library
- Ari Shapiro, Yong Cao, and Petros Faloutsos. 2006. Style Components. In Proceedings of Graphics Interface 2006 (Quebec, Canada) (GI ’06). Canadian Info. Proc. Society, CAN, 33–39.Google Scholar
- Mingyi Shi, Kfir Aberman, Andreas Aristidou, Taku Komura, Dani Lischinski, Daniel Cohen-Or, and Baoquan Chen. 2020. MotioNet: 3D Human Motion Reconstruction from Monocular Video with Skeleton Consistency. ACM Trans. Graph. (June 2020).Google Scholar
- Harrison Jesse Smith, Chen Cao, Michael Neff, and Yingying Wang. 2019. Efficient Neural Networks for Real-Time Motion Style Transfer. Proc. ACM Comput. Graph. Interact. Tech. 2, 2, Article 13 (July 2019).Google ScholarDigital Library
- Graham W. Taylor and Geoffrey E. Hinton. 2009. Factored Conditional Restricted Boltzmann Machines for Modeling Motion Style. In Proceedings of the 26th Annual International Conference on Machine Learning (Montreal, Quebec, Canada) (ICML ’09). ACM, NY, USA, 1025–1032.Google ScholarDigital Library
- Munetoshi Unuma, Ken Anjyo, and Ryozo Takeuchi. 1995. Fourier Principles for Emotion-Based Human Figure Animation. In Proceedings of the 22nd Annual Conference on Computer Graphics and Interactive Techniques(SIGGRAPH ’95). ACM, NY, USA, 91–96.Google ScholarDigital Library
- Jack M. Wang, David J. Fleet, and Aaron Hertzmann. 2007. Multifactor Gaussian Process Models for Style-content Separation. In Proceedings of the 24th International Conference on Machine Learning (Corvalis, Oregon, USA) (ICML ’07). ACM, NY, USA, 975–982.Google ScholarDigital Library
- Qi Wang, Thierry Artières, Mickael Chen, and Ludovic Denoyer. 2020. Adversarial learning for modeling human motion. Vis. Comp. 36(2020), 141–160.Google ScholarDigital Library
- Andrew Witkin and Zoran Popović. 1995. Motion Warping. In Proceedings of the 22nd Annual Conference on Computer Graphics and Interactive Techniques(SIGGRAPH ’95). ACM, NY, USA, 105–108.Google ScholarDigital Library
- Shihong Xia, Congyi Wang, Jinxiang Chai, and Jessica Hodgins. 2015. Realtime Style Transfer for Unlabeled Heterogeneous Human Motion. ACM Trans. Graph. 34, 4, Article 119 (July 2015).Google ScholarDigital Library
- M. Ersin Yumer and Niloy J. Mitra. 2016. Spectral Style Transfer for Human Motion between Independent Actions. ACM Trans. Graph. 35, 4, Article 137 (July 2016).Google ScholarDigital Library
- Jun-Yan Zhu, Taesung Park, Phillip Isola, and Alexei A Efros. 2017. Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks. In Proceedings of the IEEE International Conference on Computer Vision(ICCV’17). 2242–2251.Google ScholarCross Ref
Recommendations
Unpaired motion style transfer from video to animation
Transferring the motion style from one animation clip to another, while preserving the motion content of the latter, has been a long-standing problem in character animation. Most existing data-driven approaches are supervised and rely on paired data, ...
Motion Puzzle: Arbitrary Motion Style Transfer by Body Part
This article presents Motion Puzzle, a novel motion style transfer network that advances the state-of-the-art in several important respects. The Motion Puzzle is the first that can control the motion style of individual body parts, allowing for local ...
Efficient Neural Networks for Real-time Motion Style Transfer
Style is an intrinsic, inescapable part of human motion. It complements the content of motion to convey meaning, mood, and personality. Existing state-of-the-art motion style methods require large quantities of example data and intensive computational ...
Comments