Abstract
In this paper, we describe an algorithm which can naturally communicate with human and robot for Human-Robot Interaction by utilizing vision. We propose a state transition model using attentive features for gesture recognition. This method defines the recognition procedure as five different states; NULL, OBJECT, POSE, Local Gesture and Global Gesture. We first infer the situation of the system by estimating the transition of the state model and then apply different recognition algorithms according to the system state for robust recognition. And we propose Active Plane Model (APM) that can represent 3D and 2D information of gesture simultaneously. This method is constructing a gesture space by analyzing the statistical information of training images with PCA and the symbolized images are recognized with HMM as one of model gestures. Therefore, proposed algorithm can be used for real world application efficiently such as controlling intelligent home appliance and humanoid robot.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Liu, X., Fujimura, K.: Hand gesture Recognition using depth data. In: Proceeding 6th IEEE International Conference on AFGR, 2004 (May 2004)
Huang, F.J., Chen, T.: Tracking of Multiple Faces for Human-Computer Interfaces and Virtual Environments. In: IEEE Intl. Conf. on Multimedia and Expo. (July 2000)
Li, B., Holstein, H.: Articulated point pattern matching in optical motion capture systems. In: 7th International Conference on QinggangMeng Control, Automation, Robotics and Vision, ICARCV 2002, 2-5 December, vol. 1 (2002)
Su, Y., Allen, C.R., Geng, D., Burn, D., Brechany, U., Bell, G.D., Rowland, R.: 3-D motion system (datagloves): application for Parkinsons disease. IEEE Transactions on Instrumentation and Measurement 52(3) (2003)
Davis, J.: Recognizing Movement using Motion Histograms, MIT Media Lab. Technical Report No. 487, March (1999)
Cutler, R., Turk, M.: View-based Interpretation of Real-time Optical Flow for Gesture Recognition. In: Third IEEE International Conf. on Automatic Face and Gesture Recognition (1998)
Yoon, H.-S., Min, B.-W., Soh, J., Bae, Y.-I., Yang, H.S.: Human computer interface for gesture-based editing system. Image Analysis and Processing, 969–974 (1999)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2006 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Hong, SJ., Setiawan, N.A., Kim, SG., Lee, CW. (2006). Gesture Recognition Based on Context Awareness for Human-Robot Interaction. In: Pan, Z., Cheok, A., Haller, M., Lau, R.W.H., Saito, H., Liang, R. (eds) Advances in Artificial Reality and Tele-Existence. ICAT 2006. Lecture Notes in Computer Science, vol 4282. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11941354_1
Download citation
DOI: https://doi.org/10.1007/11941354_1
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-49776-9
Online ISBN: 978-3-540-49779-0
eBook Packages: Computer ScienceComputer Science (R0)