Lips tracking biometrics for speaker recognition
by Waleed H. Abdulla, Paul W.T. Yu, Paul Calverly
International Journal of Biometrics (IJBM), Vol. 1, No. 3, 2009

Abstract: A novel approach to extract successful biometrics from mouth visual images is presented in this paper. Visual features are extracted from a sequence of images of speakers' lips while speaking. These features consist of the shape and intensity of pixels around the edge of the lips as well as their dynamics. The features are extracted by using particle filters technique to track the movements of the lips. The lips tracker shows adequate accuracy and ability to maintain lock in different speaking scenarios. Speaker models based on these features are built using Gaussian Mixture Models (GMM) trained through the Expectation-Maximisation (EM) algorithm. Satisfactory results are obtained for text-independent speaker recognition carried out on a video database of 35 individuals. A recognition rate of 82.8% for speaker identification and equal error rate of 18% for speaker verification are achieved using this technique.

Online publication date: Mon, 30-Mar-2009

The full text of this article is only available to individual subscribers or to users at subscribing institutions.

 
Existing subscribers:
Go to Inderscience Online Journals to access the Full Text of this article.

Pay per view:
If you are not a subscriber and you just want to read the full contents of this article, buy online access here.

Complimentary Subscribers, Editors or Members of the Editorial Board of the International Journal of Biometrics (IJBM):
Login with your Inderscience username and password:

    Username:        Password:         

Forgotten your password?


Want to subscribe?
A subscription gives you complete access to all articles in the current issue, as well as to all articles in the previous three years (where applicable). See our Orders page to subscribe.

If you still need assistance, please email subs@inderscience.com