Publication:
Use of affect context in dyadic interactions for continuous emotion recognition

dc.contributor.coauthorErzin, Engin
dc.contributor.departmentN/A
dc.contributor.kuauthorFatima, Syeda Narjis
dc.contributor.kuprofilePhD Student
dc.contributor.schoolcollegeinstituteGraduate School of Sciences and Engineering
dc.contributor.yokidN/A
dc.date.accessioned2024-11-09T23:14:03Z
dc.date.issued2021
dc.description.abstractEmotional dependencies play a crucial role in understanding complexities of dyadic interactions. Recent studies have shown that affect recognition tasks can benefit by the incorporation of a particular interaction's context, however, the investigation of affect context in dyadic settings using neural network frameworks remains a complex and open problem. In this paper, we formulate the concept of dyadic affect context (DAC) and propose convolutional neural network (CNN) based architectures to model and incorporate DAC to improve continuous emotion recognition (CER) in dyadic scenarios. We begin by defining a CNN architecture for single-subject CER-based on speech and body motion data. We then introduce dyadic CER as a two-stage regression framework. Specifically, we propose two dyadic CNN architectures where cross-speaker affect contribution to the CER task is achieved by: (i) the fusion of cross-subject affect (FoA) or (ii) the fusion of cross-subject feature maps (FoM). Based on the preceding dyadic models, we finally propose a new Convolutional LSTM (ConvLSTM) model for the dyadic CER. ConvLSTM architecture captures local spectro-temporal correlations in speech and body motion as well as the long-term affect inter-dependencies between subjects. Our multimodal analysis demonstrates that modeling and incorporation of the DAC in the proposed CER models provide significant performance improvements on the USC CreativeIT database and the achieved results compare favorably to the state-of-the-art.
dc.description.indexedbyWoS
dc.description.indexedbyScopus
dc.description.openaccessNO
dc.description.publisherscopeInternational
dc.description.volume132
dc.identifier.doi10.1016/j.specom.2021.05.010
dc.identifier.eissn1872-7182
dc.identifier.issn0167-6393
dc.identifier.quartileQ2
dc.identifier.scopus2-s2.0-85107675361
dc.identifier.urihttp://dx.doi.org/10.1016/j.specom.2021.05.010
dc.identifier.urihttps://hdl.handle.net/20.500.14288/10079
dc.identifier.wos680415400007
dc.keywordsDyadic interactions
dc.keywordsContinuous emotion recognition (CER)
dc.keywordsDyadic affect context (DAC)
dc.keywordsCNN
dc.languageEnglish
dc.publisherElsevier
dc.sourceSpeech Communication
dc.subjectAcoustics
dc.subjectComputer science
dc.titleUse of affect context in dyadic interactions for continuous emotion recognition
dc.typeJournal Article
dspace.entity.typePublication
local.contributor.authoridN/A
local.contributor.kuauthorFatima, Syeda Narjis

Files