By Topic

A real-time approach to the lip-motion extraction in video sequence

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$31 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

5 Author(s)
Jian-Ming Zhang ; Sch. of Comput., Jiangsu Univ., Zhenjiang, China ; Hong Tao ; Wang Liang-Min ; Zhan Yong-Zhao
more authors

A three-stage method to extract the visual pronunciation feature of lip movements is presented in this paper. Firstly, an approach using the "Red Exclusion + Fisher Transformation'' to enhance the chromatic images in video sequences is presented, then an algorithm for segmenting the enhanced gray images with adaptive thresholding is proposed to get the box of the lip regions. Secondly, the authors classify the lip sub-images in the obtained boxes according to the visual-pronunciation features, and two formulae are presented to normalize the dimensions and the gray values of these sub-images, then a method based on SVD is used to extract features from the normalized images. Finally, the matching template based on Mahalanobis distance is applied to recognize lipshapes. The experimental results show that the features extracted by this method have the advantages of the lower dimension, more information and applicable in natural conditions over the available methods.

Published in:

Systems, Man and Cybernetics, 2004 IEEE International Conference on  (Volume:7 )

Date of Conference:

10-13 Oct. 2004