Publication:
An audio-driven dancing avatar

dc.contributor.coauthorBalci, Koray
dc.contributor.coauthorKizoglu, Idil
dc.contributor.coauthorAkarun, Lale
dc.contributor.coauthorCanton-Ferrer, Cristian
dc.contributor.coauthorTilmanne, Joelle
dc.contributor.coauthorBozkurt, Elif
dc.contributor.coauthorErdem, A. Tanju
dc.contributor.departmentDepartment of Computer Engineering
dc.contributor.departmentN/A
dc.contributor.departmentN/A
dc.contributor.departmentDepartment of Computer Engineering
dc.contributor.departmentDepartment of Electrical and Electronics Engineering
dc.contributor.kuauthorYemez, Yücel
dc.contributor.kuauthorOfli, Ferda
dc.contributor.kuauthorDemir, Yasemin
dc.contributor.kuauthorErzin, Engin
dc.contributor.kuauthorTekalp, Ahmet Murat
dc.contributor.kuprofileFaculty Member
dc.contributor.kuprofilePhD Student
dc.contributor.kuprofileMaster Student
dc.contributor.kuprofileFaculty Member
dc.contributor.kuprofileFaculty Member
dc.contributor.otherDepartment of Computer Engineering
dc.contributor.otherDepartment of Electrical and Electronics Engineering
dc.contributor.schoolcollegeinstituteCollege of Engineering
dc.contributor.schoolcollegeinstituteGraduate School of Sciences and Engineering
dc.contributor.schoolcollegeinstituteGraduate School of Sciences and Engineering
dc.contributor.schoolcollegeinstituteCollege of Engineering
dc.contributor.schoolcollegeinstituteCollege of Engineering
dc.contributor.yokid107907
dc.contributor.yokidN/A
dc.contributor.yokidN/A
dc.contributor.yokid34503
dc.contributor.yokid26207
dc.date.accessioned2024-11-10T00:12:26Z
dc.date.issued2008
dc.description.abstractWe present a framework for training and synthesis of an audio-driven dancing avatar. The avatar is trained for a given musical genre using the multicamera video recordings of a dance performance. The video is analyzed to capture the time-varying posture of the dancer's body whereas the musical audio signal is processed to extract the beat information. We consider two different marker-based schemes for the motion capture problem. The first scheme uses 3D joint positions to represent the body motion whereas the second uses joint angles. Body movements of the dancer are characterized by a set of recurring semantic motion patterns, i.e., dance figures. Each dance figure is modeled in a supervised manner with a set of HMM (Hidden Markov Model) structures and the associated beat frequency. In the synthesis phase, an audio signal of unknown musical type is first classified, within a time interval, into one of the genres that have been learnt in the analysis phase, based on mel frequency cepstral coefficients (MFCC). The motion parameters of the corresponding dance figures are then synthesized via the trained HMM structures in synchrony with the audio signal based on the estimated tempo information. Finally, the generated motion parameters, either the joint angles or the 3D joint positions of the body, are animated along with the musical audio using two different animation tools that we have developed. Experimental results demonstrate the effectiveness of the proposed framework.
dc.description.indexedbyWoS
dc.description.issue2
dc.description.openaccessNO
dc.description.publisherscopeInternational
dc.description.sponsorshipEuropean Sixth Framework Programme Network of Excellence SIMILAR
dc.description.sponsorshipScientific and Technological Research Council of Turkey (TUBITAK) [EEEAG-106E201, 2102]
dc.description.sponsorshipTurkish Academy of Sciences (TUBA) A part of this work is developed during the ENTERFACE'07 Summer Workshop on Multi-Modal Interfaces<SUP>2</SUP> and it has been supported by the European Sixth Framework Programme Network of Excellence SIMILAR,<SUP>3</SUP> by the Scientific and Technological Research Council of Turkey (TUBITAK)<SUP>4</SUP> under project EEEAG-106E201 and COST Action: 2102.<SUP>5</SUP> A. Murat Tekalp acknowledges support from Turkish Academy of Sciences (TUBA).<SUP>6</SUP>
dc.description.volume2
dc.identifier.doi10.1007/s12193-008-0009-x
dc.identifier.eissn1783-8738
dc.identifier.issn1783-7677
dc.identifier.quartileQ3
dc.identifier.urihttp://dx.doi.org/10.1007/s12193-008-0009-x
dc.identifier.urihttps://hdl.handle.net/20.500.14288/17659
dc.identifier.wos208536800003
dc.keywordsMulticamera motion capture
dc.keywordsAudio-driven body motion synthesis
dc.keywordsDancing avatar animation performance
dc.languageEnglish
dc.publisherSpringer
dc.sourceJournal on Multimodal User Interfaces
dc.subjectComputer Science
dc.subjectArtificial intelligence
dc.subjectCybernetics
dc.titleAn audio-driven dancing avatar
dc.typeJournal Article
dspace.entity.typePublication
local.contributor.authorid0000-0002-7515-3138
local.contributor.authorid0000-0003-3918-3230
local.contributor.authoridN/A
local.contributor.authorid0000-0002-2715-2368
local.contributor.authorid0000-0003-1465-8121
local.contributor.kuauthorYemez, Yücel
local.contributor.kuauthorOfli, Ferda
local.contributor.kuauthorDemir, Yasemin
local.contributor.kuauthorErzin, Engin
local.contributor.kuauthorTekalp, Ahmet Murat
relation.isOrgUnitOfPublication89352e43-bf09-4ef4-82f6-6f9d0174ebae
relation.isOrgUnitOfPublication21598063-a7c5-420d-91ba-0cc9b2db0ea0
relation.isOrgUnitOfPublication.latestForDiscovery21598063-a7c5-420d-91ba-0cc9b2db0ea0

Files