Publication: Cross-subject continuous emotion recognition using speech and body motion in dyadic interactions
dc.contributor.coauthor | N/A | |
dc.contributor.department | N/A | |
dc.contributor.department | Department of Computer Engineering | |
dc.contributor.kuauthor | Fatima, Syeda Narjis | |
dc.contributor.kuauthor | Erzin, Engin | |
dc.contributor.kuprofile | PhD Student | |
dc.contributor.kuprofile | Faculty Member | |
dc.contributor.other | Department of Computer Engineering | |
dc.contributor.schoolcollegeinstitute | Graduate School of Sciences and Engineering | |
dc.contributor.schoolcollegeinstitute | College of Engineering | |
dc.contributor.yokid | N/A | |
dc.contributor.yokid | 34503 | |
dc.date.accessioned | 2024-11-09T23:40:04Z | |
dc.date.issued | 2017 | |
dc.description.abstract | Dyadic interactions encapsulate rich emotional exchange between interlocutors suggesting a multimodal, cross-speaker and cross-dimensional continuous emotion dependency. This study explores the dynamic inter-attribute emotional dependency at the cross-subject level with implications to continuous emotion recognition based on speech and body motion cues. We propose a novel two-stage Gaussian Mixture Model mapping framework for the continuous emotion recognition problem. In the first stage, we perform continuous emotion recognition (CER) of both speakers from speech and body motion modalities to estimate activation, valence and dominance (AVD) attributes. In the second stage, we improve the first stage estimates by performing CER of the selected speaker using her/his speech and body motion modalities as well as using the estimated affective attribute(s) of the other speaker. Our experimental evaluations indicate that the second stage, cross-subject continuous emotion recognition (CSCER), provides complementary information to recognize the affective state, and delivers promising improvements for the continuous emotion recognition problem. | |
dc.description.indexedby | WoS | |
dc.description.indexedby | Scopus | |
dc.description.openaccess | NO | |
dc.description.publisherscope | International | |
dc.description.sponsoredbyTubitakEu | N/A | |
dc.identifier.doi | 10.21437/Interspeech.2017-1413 | |
dc.identifier.isbn | 978-1-5108-4876-4 | |
dc.identifier.issn | 2308-457X | |
dc.identifier.quartile | N/A | |
dc.identifier.scopus | 2-s2.0-85039154284 | |
dc.identifier.uri | http://dx.doi.org/10.21437/Interspeech.2017-1413 | |
dc.identifier.uri | https://hdl.handle.net/20.500.14288/13229 | |
dc.identifier.wos | 457505000357 | |
dc.keywords | Continuous emotion recognition | |
dc.keywords | Dyadic emotion estimator | |
dc.keywords | Side emotional information | |
dc.keywords | Cross-subject continuous emotion recognition (CSCER) | |
dc.keywords | Gaussian mixture regression | |
dc.keywords | Activation | |
dc.keywords | Valence | |
dc.keywords | Dominance FACE | |
dc.language | English | |
dc.publisher | International Speech Communication Association ( ISCA) | |
dc.source | 18th Annual Conference of the International-Speech-Communication-Association (INTERSPEECH 2017) | |
dc.subject | Computer science | |
dc.subject | Artificial intelligence | |
dc.subject | Engineering | |
dc.subject | Electrical electronic engineering | |
dc.title | Cross-subject continuous emotion recognition using speech and body motion in dyadic interactions | |
dc.type | Conference proceeding | |
dspace.entity.type | Publication | |
local.contributor.authorid | N/A | |
local.contributor.authorid | 0000-0002-2715-2368 | |
local.contributor.kuauthor | Fatima, Syeda Narjis | |
local.contributor.kuauthor | Erzin, Engin | |
relation.isOrgUnitOfPublication | 89352e43-bf09-4ef4-82f6-6f9d0174ebae | |
relation.isOrgUnitOfPublication.latestForDiscovery | 89352e43-bf09-4ef4-82f6-6f9d0174ebae |