ABSTRACT
In this paper we address the problem of automatically deriving vocabularies of motion modules from human motion data, taking advantage of the underlying spatio-temporal structure in motion. We approach this problem with a data-driven methodology for modularizing a motion stream (or time-series of human motion) into a vocabulary of parameterized primitive motion modules and a set of meta-level behaviors characterizing extended combinations of the primitives. Central to this methodology is the discovery of spatio-temporal structure in a motion stream. We estimate this structure by extending an existing nonlinear dimension reduction technique, Isomap, to handle motion data with spatial and temporal dependencies. The motion vocabularies derived by our methodology provide a substrate of autonomous behavior and can be used in a variety of applications. We demonstrate the utility of derived vocabularies for the application of synthesizing new humanoid motion that is structurally similar to the original demonstrated motion.
- O. Arikan and D. A. Forsyth. Interactive motion generation from examples. ACM Transactions on Graphics (TOG), 21(3):483--490, 2002. Google ScholarDigital Library
- D. C. Bentivegna and C. G. Atkeson. Learning from observation using primitives. In IEEE International Conference on Robotics and Automation, pages 1988--1993, Seoul, Korea, May 2001.Google ScholarCross Ref
- D. C. Bentivegna, A. Ude, C. G. Atkeson, and G. Cheng. Humanoid robot learning and game playing using pc-based vision. In IEEE/RSJ International Conference on Intelligent Robots and Systems, volume 3, pages 2449--2454, Lausanne, Switzerland, October 2002.Google ScholarCross Ref
- M. Brand and A. Hertzmann. Style machines. In Proceedings of ACM SIGGRAPH 2000, Computer Graphics Proceedings, Annual Conference Series, pages 183--192. ACM Press / ACM SIGGRAPH / Addison Wesley Longman, July 2000. ISBN 1-58113-208-5. Google ScholarDigital Library
- C.-W. Chu, O. C. Jenkins, and M. J. Mataric. Markerless kinematic model and motion capture from volume sequences. In To appear in the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2003), Madison, Wisconsin, USA, June 2003.Google Scholar
- J. D. Cohen, M. C. Lin, D. Manocha, and M. K. Ponamgi. I-COLLIDE: An interactive and exact collision detection system for large-scale environments. In Proceedings of the 1995 symposium on Interactive 3D graphics, pages 189--196, 218, 1995. Google ScholarDigital Library
- A. Fod, M. Mataric, and O. Jenkins. Automated derivation of primitives for movement classification. Autonomous Robots, 12(1):39--54, January 2002. Google ScholarDigital Library
- A. J. Ijspeert, J. Nakanishi, and S. Schaal. Trajectory formation for imitation with nonlinear dynamical systems. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS2001), pages 752--757, Maui, Hawaii, USA, 2001.Google ScholarCross Ref
- M. Inc. http://www.measurand.com.Google Scholar
- M. Kallmann, J.-S. Monzani, A. Caicedo, and D. Thalmann. Ace: A platform for real time simulation of virtual human agents. In 11th Eurographics Workshop on Animation and Simulation, Interlaken, Switzerland, August 2000.Google ScholarCross Ref
- L. Kovar, M. Gleicher, and F. Pighin. Motion graphs. ACM Transactions on Graphics (TOG), 21(3):473--482, 2002. Google ScholarDigital Library
- J. Lee, J. Chai, P. S. A. Reitsma, J. K. Hodgins, and N. S. Pollard. Interactive control of avatars animated with human motion data. ACM Transactions on Graphics (TOG), 21(3):491--500, 2002. Google ScholarDigital Library
- Y. Li, T. Wang, and H.-Y. Shum. Motion texture: a two-level statistical model for character motion synthesis. ACM Transactions on Graphics (TOG), 21(3):465--472, 2002. Google ScholarDigital Library
- M. J. Mataric. Sensory-motor primitives as a basis for imitation: Linking perception to action and biology to robotics. In C. Nehaniv and K. Dautenhahn, editors, Imitation in Animals and Artifacts, pages 392--422. MIT Press, 2002. Google ScholarDigital Library
- I. Mikic, M. Trivedi, E. Hunter, and P. Cosman. Articulated body posture estimation from multi-camera voxel data. In IEEE International Conference on Computer Vision and Pattern Recognition, pages 455--460, Kauai, HI, USA, December 2001.Google ScholarCross Ref
- J. Rickel, S. Marsella, J. Gratch, R. Hill, D. Traum, and W. Swartout. Toward a new generation of virtual humans for interactive experiences. IEEE Intelligent Systems, 17(4):32--38, July/August 2002. Google ScholarDigital Library
- C. Rose, M. F. Cohen, and B. Bodenheimer. Verbs and adverbs: Multidimensional motion interpolation. IEEE Computer Graphics & Applications, 18(5):32--40, September - October 1998. ISSN 0272-1716. Google ScholarDigital Library
- S. T. Roweis and L. K. Saul. Nonlinear dimensionality reduction by locally linear embedding. Science, 290(5500):2323--2326, 2000.Google ScholarCross Ref
- B. Scholkopf, A. J. Smola, and K.-R. Muller. Nonlinear component analysis as a kernel eigenvalue problem. Neural Computation, 10(5):1299--1319, 1998. Google ScholarDigital Library
- D. Shepard. A two-dimensional interpolation function for irregularly-spaced data. In Proceedings of the ACM national conference, pages 517--524. ACM Press, 1968. Google ScholarDigital Library
- J. B. Tenenbaum, V. de Silva, and J. C. Langford. A global geometric framework for nonlinear dimensionality reduction. Science, 290(5500):2319--2323, 2000.Google ScholarCross Ref
- D. M. Wolpert and M. Kawato. Multiple paired forward and inverse models for motor control. Neural Networks, 11(7-8):1317--1329, 1998. Google ScholarDigital Library
Index Terms
- Automated derivation of behavior vocabularies for autonomous humanoid motion
Recommendations
Dynamic walking and whole-body motion planning for humanoid robots: an integrated approach
This paper presents a general method for planning collision-free whole-body walking motions for humanoid robots. First, we present a randomized algorithm for constrained motion planning, that is used to generate collision-free statically balanced paths ...
Sensory integration with articulated motion on a humanoid robot
This paper describes the integration of articulated motion with auditory and visual sensory information that enables a humanoid robot to achieve certain reflex actions that mimic those of people. Reflexes such as reach-and-grasp behavior enables the ...
Automated human motion segmentation via motion regularities
Analysis and reuse of human motion capture (mocap) data play an important role in animation, games and medical rehabilitation. In various mocap-based animation techniques, motion segmentation is regarded as one of the fundamental functions. Many ...
Comments