Publication: Discriminative LIP-motion features for biometric speaker identification
Program
KU Authors
Co-Authors
Advisor
Publication Date
2004
Language
English
Type
Conference proceeding
Journal Title
Journal ISSN
Volume Title
Abstract
This paper addresses the selection of best lip motion features for biometric open-set speaker identification. The best features are those that result in the highest discrimination of individual speakers in a population. We first detect the face region in each video frame. The lip region for each frame is then segmented following registration of successive face regions by global motion compensation. The initial lip feature vector is composed of the 2D-DCT coefficients of the optical flow vectors within the lip region at each frame. The discriminant analysis is composed of two stages. At the first stage, the most discriminative features are selected from the full set of DCT coefficients of a single lip motion frame by using a probabilistic measure that maximizes the ratio of intra-class and inter-class probabilities. At the second stage, the resulting discriminative feature vectors are interpolated and concatenated for each time instant within a neighborhood, and further analyzed by LDA to reduce dimension, this time taking into account temporal discrimination information. Experimental results of the HMM-based speaker identification system are included to demonstrate the performance.
Description
Source:
Proceedings - International Conference on Image Processing, ICIP
Publisher:
IEEE
Keywords:
Subject
Electrical electronics engineering, Computer engineering