Publication:
Prosody-driven head-gesture animation

dc.contributor.coauthorErdem, A. T.
dc.contributor.coauthorErdem, C.
dc.contributor.coauthorÖzkan, M.
dc.contributor.departmentN/A
dc.contributor.departmentDepartment of Computer Engineering
dc.contributor.departmentDepartment of Computer Engineering
dc.contributor.departmentDepartment of Electrical and Electronics Engineering
dc.contributor.kuauthorSargın, Mehmet Emre
dc.contributor.kuauthorErzin, Engin
dc.contributor.kuauthorYemez, Yücel
dc.contributor.kuauthorTekalp, Ahmet Murat
dc.contributor.kuprofileMaster Student
dc.contributor.kuprofileFaculty Member
dc.contributor.kuprofileFaculty Member
dc.contributor.kuprofileFaculty Member
dc.contributor.otherDepartment of Computer Engineering
dc.contributor.otherDepartment of Electrical and Electronics Engineering
dc.contributor.schoolcollegeinstituteGraduate School of Sciences and Engineering
dc.contributor.schoolcollegeinstituteCollege of Engineering
dc.contributor.schoolcollegeinstituteCollege of Engineering
dc.contributor.schoolcollegeinstituteCollege of Engineering
dc.contributor.yokidN/A
dc.contributor.yokid34503
dc.contributor.yokid107907
dc.contributor.yokid26207
dc.date.accessioned2024-11-09T23:22:33Z
dc.date.issued2007
dc.description.abstractWe present a new framework for joint analysis of head gesture and speech prosody patterns of a speaker towards automatic realistic synthesis of head gestures from speech prosody. The proposed two-stage analysis aims to "learn" both elementary prosody and head gesture patterns for a particular speaker, as well as the correlations between these head gesture and prosody patterns from a training video sequence. The resulting audio-visual mapping model is then employed to synthesize natural head gestures from arbitrary input test speech given a head model for the speaker. Objective and subjective evaluations indicate that the proposed synthesis by analysis scheme provides natural looking head gestures for the speaker with any input test speech.
dc.description.indexedbyWoS
dc.description.indexedbyScopus
dc.description.openaccessNO
dc.description.publisherscopeInternational
dc.description.sponsorshipEuropean FP6 Network of Excellence SIMILAR
dc.description.sponsorshipTUBITAK[EEEAG-106E201]
dc.description.sponsorshipCOST2102 action The authors would like to thank Momentum Inc. for making the talking head avatar available and for their collaboration to build the MVGL-MASAL gesture-speech database. This work has been supported by the European FP6 Network of Excellence SIMILAR (http://ww.similar.cc)and by TUBITAKunder project EEEAG-106E201 and COST2102 action). The work of M.E. Sargin was done while he worked at Koc University
dc.identifier.doiN/A
dc.identifier.issn1520-6149
dc.identifier.quartileN/A
dc.identifier.scopus2-s2.0-34547506366
dc.identifier.urihttps://hdl.handle.net/20.500.14288/11074
dc.identifier.wos248908100170
dc.keywordsMan-machine systems
dc.keywordsMultimedia systems
dc.keywordsGesture and prosody analysis
dc.keywordsGesture synthesis
dc.languageEnglish
dc.publisherInstitute of Electrical and Electronics Engineers (IEEE)
dc.source2007 IEEE International Conference on Acoustics, Speech, and Signal Processing, Vol Ii, Pts 1-3
dc.subjectAcoustics
dc.subjectComputer science
dc.titleProsody-driven head-gesture animation
dc.typeConference proceeding
dspace.entity.typePublication
local.contributor.authoridN/A
local.contributor.authorid0000-0002-2715-2368
local.contributor.authorid0000-0002-7515-3138
local.contributor.authorid0000-0003-1465-8121
local.contributor.kuauthorSargın, Mehmet Emre
local.contributor.kuauthorErzin, Engin
local.contributor.kuauthorYemez, Yücel
local.contributor.kuauthorTekalp, Ahmet Murat
relation.isOrgUnitOfPublication89352e43-bf09-4ef4-82f6-6f9d0174ebae
relation.isOrgUnitOfPublication21598063-a7c5-420d-91ba-0cc9b2db0ea0
relation.isOrgUnitOfPublication.latestForDiscovery21598063-a7c5-420d-91ba-0cc9b2db0ea0

Files