Publication:
Estimation and analysis of facial animation parameter patterns

Placeholder

Program

KU Authors

Co-Authors

N/A

Advisor

Publication Date

Language

English

Journal Title

Journal ISSN

Volume Title

Abstract

We propose a framework for estimation and analysis of temporal facial expression patterns of a speaker. ne proposed system aims to learn personalized elementary dynamic facial expression patterns for a particular speaker. We use head-and-shoulder stereo video sequences to track lip, eye, eyebrow, and eyelid motion of a speaker in 3D. MPEG-4 Facial Definition Parameters (FDPs) are used as the feature set, and temporal facial expression patterns are represented by the MPEG-4 Facial Animation Parameters (FAPs). We perform Hidden Markov Model (HMM) based unsupervised temporal segmentation of upper and lower facial expression features separately to determine recurrent elementary facial expression patterns for a particular speaker. These facial expression patterns coded by FAP sequences, which may not be tied with prespecified emotions, can be used for personalized emotion estimation and synthesis of a speaker. Experimental results are presented.

Source:

2007 Ieee International Conference On Image Processing, Vols 1-7

Publisher:

Ieee

Keywords:

Subject

Engineering, Electrical and electronic engineering, Imaging science, Photographic technology

Citation

Endorsement

Review

Supplemented By

Referenced By

Copyrights Note

0

Views

0

Downloads

View PlumX Details