By Topic

3-D motion estimation of human head for model-based image coding

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$33 $33
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

2 Author(s)
T. Fukuhara ; Commun. Syst. Lab., Kanagawa, Japan ; T. Murakami

Model-based image coding applied to interpersonal communication achieves very low bit-rate image transmission. To accomplish it, accurate three-dimensional (3-D) motion estimation of a speaker is necessary. A new method of 3-D motion estimation is presented, consisting of two steps. In the first, facial contours and feature points of a speaker are extracted using filtering and snake algorithms. Five feature points on a speaker's facial image are tracked between consecutive picture frames, which gives 2-D motion vectors of the feature points. Then, in the second step, the 3-D motion of a speaker's head is estimated using a three-layered neural network model, after training with many possible motion patterns of the human head using an existing 3-D general shape model. Experimental results show that the method not only achieves good results but is also more robust than existing methods.<>

Published in:

IEE Proceedings I - Communications, Speech and Vision  (Volume:140 ,  Issue: 1 )