Publication: Affective synthesis and animation of arm gestures from speech prosody
dc.contributor.department | Department of Computer Engineering | |
dc.contributor.department | Graduate School of Sciences and Engineering | |
dc.contributor.kuauthor | Bozkurt, Elif | |
dc.contributor.kuauthor | Erzin, Engin | |
dc.contributor.kuauthor | Yemez, Yücel | |
dc.contributor.schoolcollegeinstitute | College of Engineering | |
dc.contributor.schoolcollegeinstitute | GRADUATE SCHOOL OF SCIENCES AND ENGINEERING | |
dc.date.accessioned | 2024-11-09T23:05:35Z | |
dc.date.issued | 2020 | |
dc.description.abstract | In human-to-human communication, speech signals carry rich emotional cues that are further emphasized by affect-expressive gestures. In this regard, automatic synthesis and animation of gestures accompanying affective verbal communication can help to create more naturalistic virtual agents in human-computer interaction systems. Speech-driven gesture synthesis can map emotional cues of the speech signal to affect-expressive gestures by modeling complex variability and timing relationships of speech and gesture. In this paper, we investigate the use of continuous affect attributes, which are activation, valence and dominance, for speech-driven affective synthesis and animation of arm gestures. To this effect, we present a statistical framework based on hidden semi-Markov models (HSMM), where states are gestures and observations are speech-prosody and continuous affect attributes. The proposed framework is evaluated considering four distinct HSMM structures which differ by their emission distributions. Evaluations are performed over the USC CreativeIT database in a speaker-independent setup. Among the four statistical structures, the conditional structure, which models observation distributions as prosody given affect, achieves the best performance under both objective and subjective evaluations. | |
dc.description.indexedby | WOS | |
dc.description.indexedby | Scopus | |
dc.description.openaccess | NO | |
dc.description.sponsoredbyTubitakEu | N/A | |
dc.description.sponsorship | TUBITAK [113E102] This work was supported by TUBITAK under grant number 113E102. | |
dc.description.volume | 119 | |
dc.identifier.doi | 10.1016/j.specom.2020.02.005 | |
dc.identifier.eissn | 1872-7182 | |
dc.identifier.issn | 0167-6393 | |
dc.identifier.scopus | 2-s2.0-85080918361 | |
dc.identifier.uri | https://doi.org/10.1016/j.specom.2020.02.005 | |
dc.identifier.uri | https://hdl.handle.net/20.500.14288/8837 | |
dc.identifier.wos | 531017100001 | |
dc.keywords | Prosody analysis | |
dc.keywords | Gesture segmentation | |
dc.keywords | Arm gesture animation | |
dc.keywords | Unit selection | |
dc.keywords | Hidden semi-markov models | |
dc.keywords | Speech-driven affective gesture synthesis | |
dc.keywords | Body movement | |
dc.keywords | Expression | |
dc.keywords | Dynamics | |
dc.keywords | Emotion | |
dc.language.iso | eng | |
dc.publisher | Elsevier | |
dc.relation.ispartof | Speech Communication | |
dc.subject | Acoustics | |
dc.subject | Computer science | |
dc.title | Affective synthesis and animation of arm gestures from speech prosody | |
dc.type | Journal Article | |
dspace.entity.type | Publication | |
local.contributor.kuauthor | Bozkurt, Elif | |
local.contributor.kuauthor | Erzin, Engin | |
local.contributor.kuauthor | Yemez, Yücel | |
local.publication.orgunit1 | GRADUATE SCHOOL OF SCIENCES AND ENGINEERING | |
local.publication.orgunit1 | College of Engineering | |
local.publication.orgunit2 | Department of Computer Engineering | |
local.publication.orgunit2 | Graduate School of Sciences and Engineering | |
relation.isOrgUnitOfPublication | 89352e43-bf09-4ef4-82f6-6f9d0174ebae | |
relation.isOrgUnitOfPublication | 3fc31c89-e803-4eb1-af6b-6258bc42c3d8 | |
relation.isOrgUnitOfPublication.latestForDiscovery | 89352e43-bf09-4ef4-82f6-6f9d0174ebae | |
relation.isParentOrgUnitOfPublication | 8e756b23-2d4a-4ce8-b1b3-62c794a8c164 | |
relation.isParentOrgUnitOfPublication | 434c9663-2b11-4e66-9399-c863e2ebae43 | |
relation.isParentOrgUnitOfPublication.latestForDiscovery | 8e756b23-2d4a-4ce8-b1b3-62c794a8c164 |