Abstract
Accelerometer-based gesture control is studied as a supplementary or an alternative interaction modality. Gesture commands freely trainable by the user can be used for controlling external devices with handheld wireless sensor unit. Two user studies are presented. The first study concerns finding gestures for controlling a design environment (Smart Design Studio), TV, VCR, and lighting. The results indicate that different people usually prefer different gestures for the same task, and hence it should be possible to personalise them. The second user study concerns evaluating the usefulness of the gesture modality compared to other interaction modalities for controlling a design environment. The other modalities were speech, RFID-based physical tangible objects, laser-tracked pen, and PDA stylus. The results suggest that gestures are a natural modality for certain tasks, and can augment other modalities. Gesture commands were found to be natural, especially for commands with spatial association in design environment control.
Similar content being viewed by others
References
Starner T, Auxier J, Ashbrook D, Gandy M (2000) The gesture pendant: a self-illuminating, wearable, infrared computer vision system for home automation control and medical monitoring. In: proceedings of the fourth international symposium on wearable computers, ISWC 2000, pp 87–95
Rekimoto J (2001) GestureWrist and gesturepad : unobtrusive wearable interaction devices. In: proceedings of the fifth international symposium on wearable computers, ISWC 2001, pp 21–31
Sawada H, Hashimoto S (2000) Gesture recognition using an accelerometer sensor and its application to musical performance control. Electron Commun Jpn Part 3, pp 9–17
Hoffman F, Heyer P, Hommel G (1997) Velocity profile based recognition of dynamic gestures with discrete hidden markov models. Proceedings of gesture workshop ‘97, Springer, Berlin Heidelberg, Newyork
Tsukada K, Yasumura M (2002) Ubi-finger: gesture input device for mobile use. In: Proceedings of APCHI 2002, Vol. 1, pp 388-400
Wilson A, Shafer S (2003) Between u and i: XWand: UI for intelligent spaces. In: Proceedings of the conference on human factors in computing systems, CHI 2003, April 2003, pp 545–552
Flanagan J, Mäntyjärvi J, Korpiaho K, Tikanmäki J (2002) Recognizing movements of a handheld device using symbolic representation and coding of sensor signals. In: Proceedings of the first international conference on mobile and ubiquitous multimedia, pp 104–112
Mäntylä V-M, Mäntyjärvi J, Seppänen T, Tuulari E (2000) Hand gesture recognition of a mobile device user. In: Proceedings of the international IEEE conference on multimedia and expo, pp 281–284
Theodoridis S, Koutroumbas K (1999) Pattern recognition. Academic press, London
Pirhonen A, Brewster S, Holgiun C (2002) Gestural and audio metaphors as a means of control for mobile devices. CHI 2002, April 2002, pp 291–298
Tuulari E, Ylisaukko-oja A (2002) SoapBox: a platform for ubiquitous computing research and applications. In: First international conference, Pervasive 2002, pp 26-28
Rabiner L (1998) Tutorial on hidden markov models and selected applications in speech recognition. In: Proceedings of the IEEE, Vol. 77, No. 2
Kay S (2000) Can detectability be improved by adding noise? IEEE Signal Process Lett, 7(1):8–10
Ailisto H, Plomp J, Pohjanheimo L, Strömmer E (2003) A physical selection paradigm for ubiquitous computing. In: 1st European symposium on ambient intelligence (EUSAI 2003). ambient intelligence, Lecture Notes in Computer Science Vol. 2875. Aarts, Emile et al (eds) Springer, Berlin Heidelberg, Newyork, pp 372–383
Peltola J, Plomp J, Seppänen T (1999) A dictionary-adaptive speech driven user interface for distributed multimedia platform. In: Euromicro Workshop on multimedia and telecommunications, Milan, Italy
Kallio S, Kela J, Mäntyjärvi J (2003) Online gesture recognition system for mobile interaction. IEEE International Conference on Systems, Man & Cybernetics, Vol. 3, Washington DC, USA, pp 2070–2076
Mäntyjärvi J (2003) Sensor-based context recognition for mobile applications. VTT Publications 511
Iacucci G, Kela J, Pehkonen P (2004) Computational support to record and re-experience visits. Personal and ubiquitous computing, Vol 8 No 2, Springer, Berlin Heidelberg, Newyork, pp 100–109
Gersho A, Gray RM (1991) Vector Quantization and signal compression. Kluwer, Dordrecht
Yoon HS (2001) Hand gesture recognition using combined features of location, angle and velocity. Pattern Recogn 34:491–501
Mäntylä V-M (2001) Discrete hidden markov models with application to isolated user-dependent hand gesture recognition. VTT Publications 449
Mäntyjärvi J, Kela J, Korpipää P, Kallio S (2004) Enabling fast and effortless customisation in accelerometer based gesture interaction. In: Proceedings of the third international conference on mobile and ubiquitous multimedia, ACM, pp 25–31
Acknowledgements
We gratefully acknowledge research funding from the National Technology Agency of Finland (Tekes) and the Italian Ministry of Education, University and Research (MIUR). We would also like to thank our partners in the ITEA Ambience project.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Kela, J., Korpipää, P., Mäntyjärvi, J. et al. Accelerometer-based gesture control for a design environment. Pers Ubiquit Comput 10, 285–299 (2006). https://doi.org/10.1007/s00779-005-0033-8
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00779-005-0033-8