Publication:
Discriminative LIP-motion features for biometric speaker identification

Placeholder

Program

KU Authors

Co-Authors

Advisor

Publication Date

2004

Language

English

Type

Conference proceeding

Journal Title

Journal ISSN

Volume Title

Abstract

This paper addresses the selection of best lip motion features for biometric open-set speaker identification. The best features are those that result in the highest discrimination of individual speakers in a population. We first detect the face region in each video frame. The lip region for each frame is then segmented following registration of successive face regions by global motion compensation. The initial lip feature vector is composed of the 2D-DCT coefficients of the optical flow vectors within the lip region at each frame. The discriminant analysis is composed of two stages. At the first stage, the most discriminative features are selected from the full set of DCT coefficients of a single lip motion frame by using a probabilistic measure that maximizes the ratio of intra-class and inter-class probabilities. At the second stage, the resulting discriminative feature vectors are interpolated and concatenated for each time instant within a neighborhood, and further analyzed by LDA to reduce dimension, this time taking into account temporal discrimination information. Experimental results of the HMM-based speaker identification system are included to demonstrate the performance.

Description

Source:

Proceedings - International Conference on Image Processing, ICIP

Publisher:

IEEE

Keywords:

Subject

Electrical electronics engineering, Computer engineering

Citation

Endorsement

Review

Supplemented By

Referenced By

Copy Rights Note

0

Views

0

Downloads

View PlumX Details