3D Hand Motion Evaluation Using HMM

Document Type: Research Paper

Authors

1 School of Science Engineering, Sharif University of Technology, International Campus, Kish Island, Iran

2 School of Electrical Engineering, Sharif University of Technology, Tehran, Iran

Abstract

Gesture and motion recognition are needed for a variety of applications. The use of human hand motions as a natural interface tool has motivated researchers to conduct research in the modeling, analysis and recognition of various hand movements. In particular, human-computer intelligent interaction has been a focus of research in vision-based gesture recognition. In this work, we introduce a 3-D hand model recognition method that offers flexible and elaborate representation of hand motion. We used landmark points on the tips and joints of the fingers and calculated the 3-D coordinates of these points through a stereo vision system followed by a Hidden Markov Model (HMM) to recognize hand motions. Experimentally, in an effort to evaluate the formation of hand gestures similar to those used in rehabilitation sessions, we studied three evolving motions. Given the natural hand features and uncontrolled environment, we were able to classify and differentiate unnatural slowness or rapidness in the performance of such motions, ranging from 45% to 93%.  

Keywords


[1] Mehrdad Sangi, Mehran Jahed, “A Fast 3D Hand Model Reconstruction by Stereo Vision System,” IEEE 2010.

[2] X. Zabulis, H. Baltzakis, A. Argyros, “Vision-based Hand Gesture Recognition for Human-Computer Interaction,” Institute of Computer Science Foundation for Research and Technology - Hellas (FORTH) Heraklion, Crete, Greece.

[3] Pragati Garg, Naveen Aggarwal and Sanjeev Sofat, “Vision Based Hand Gesture Recognition,” World Academy of Science, Engineering and Technology 49 2009.

[4] Ying Wu and Thomas S. Huang, ”Hand Modeling, Analysis, and Recognition,” IEEE Signal Processing Magazine,2001.

[5] Ali Erol, George Bebis, Mircea Nicolescu, “A Review on VisionBased Full DOF Hand Motion Estimation,” IEEEHuman Motion Proceedings, 2000.

[6] Nilanjan Sarkar , “Human-Robot Interaction,” Itech Education and Publishing, Vienna, Austria, 2007.

[7] Nilanjan Sarkar , “Human-Robot Interaction,” Itech Education and Publishing, Vienna, Austria, 2007.

[8] Available in: http://en.wikipedia.org/wiki/Kmeans_clustering.

[9] Dana Elena Ilea and Paul F. Whelan, “Color Image Segmentation Using a Spatial K-Means Clustering Algorithm,” IMVIP 2006 - 10th International Machine Vision and Image Processing Conference, 30 August - 1 September 2006, Dublin, Ireland.

[10] Tse-Wei Chen,Yi-Ling Chen, Shao-Yi Chien, “Fast image segmentation based on K-Means clustering with histograms in HSV color space,” Multimedia Signal Processing, 2008 IEEE 10th Workshop on, 8-10 Oct. 2008.

[11] “Definition of CIELab Color Space,” available in: http://www.optelvision.com/documents/optel-vision-sexplanation-on-cielab-color-space.pdf

[12] Nathaniel J. Short, “ 3-D Point Cloud Generation from Rigid and Flexible Stereo Vision Systems,” Thesis submitted to the faculty of the Virginia Polytechnic Institute and State University in partial fulfillment of the requirements for the degree of MASTER OF SCIENCE in Computer Engineering 2009.

[13] Gary Bradski and Adrian Kaehler, “Learning OpenCV,” O’REILLY,2008.

[14] Ricardo A. Arango S, “Learning Action Primitives From 3D Stereo Vision Measurements,” A thesis submitted to the University of Aalborg for the degree of Master of Engineering in Computer Vision and Graphics 2010.

[15] J. Lou, Q. Liu, T. Tan, and W. Hu, “Semantic interpretation of object activities in a surveillance system,” In Proc. International Conference on Pattern Recognition, pages III: 777–780, 2002.

[16] N.M. Oliver, B. Rosario, and A.P. Pentland, “A Bayesian computer vision system for modeling human interactions,” IEEE Trans. Pattern Analysis and Machine Intelligence, 22(8):831–843, August 2000.

[17] Available in: http://en.wikipedia.org/wiki/Epipolar_geometry.

[18] Roger Y. TSAI , “A Versatile Camera Calibration Technique for High-Accuracy 3D Machine Vision Metrology Using Off the-shelf TV Cameras and Lenses,” IEEE Journal of Robotics and Automation, VOL. RA-3, NO. 4, AUGUST 1987.

[19] Ying Wu, John Y.Lin, Thomas S.Huang, “ Capturing Natural Hand Articulation,” IEEE International Conference on Computer Vision, Canada, 2001.

[20] Makoto Kato, Yen-Wei Chen and Gang Xu, “Articulated Hand Tracking by PCA-ICA Approach,” International Conference on Automatic Face and Gesture Recognition, IEEE 2006.

[21] Brent Martin, “INSTANCE-BASED LEARNING: Nearest Neighbor with Generalization,” Department of Computer Science University of Waikato Hamilton, New Zealand, March 1995.

[22] Gernot A. Fink, “Markov Models for Pattern Recognition From Theory to Applications,” Springer-Verlag Berlin Heidelberg 2008.

[23] Mahmoud Elmezain, Ayoub Al-Hamadi, J¨org Appenrodt, Bernd Michaelis, “A Hidden Markov Model-Based Continuous Gesture Recognition System for Hand Motion Trajectory,” IEEE 2008.

[24] Lawrence R. Rabiner, “A Tutorial on Hidden Markov Models and Selected Applications in Speech Recognition,” Proceedings of the IEEE, Vol. 77, No. 2, pp. 257-286, February, 1989.