Publication: Formant position based weighted spectral features for emotion recognition
dc.contributor.coauthor | Erdem, Çiğdem Eroğlu | |
dc.contributor.coauthor | Erdem, Arif Tanju | |
dc.contributor.department | Department of Computer Engineering | |
dc.contributor.department | Graduate School of Sciences and Engineering | |
dc.contributor.kuauthor | Bozkurt, Elif | |
dc.contributor.kuauthor | Erzin, Engin | |
dc.contributor.schoolcollegeinstitute | College of Engineering | |
dc.contributor.schoolcollegeinstitute | GRADUATE SCHOOL OF SCIENCES AND ENGINEERING | |
dc.date.accessioned | 2024-11-09T23:39:22Z | |
dc.date.issued | 2011 | |
dc.description.abstract | In this paper, we propose novel spectrally weighted mel-frequency cepstral coefficient (WMFCC) features for emotion recognition from speech. The idea is based on the fact that formant locations carry emotion-related information, and therefore critical spectral bands around formant locations can be emphasized during the calculation of MFCC features. The spectral weighting is derived from the normalized inverse harmonic mean function of the line spectral frequency (LSF) features, which are known to be localized around formant frequencies. The above approach can be considered as an early data fusion of spectral content and formant location information. We also investigate methods for late decision fusion of unimodal classifiers. We evaluate the proposed WMFCC features together with the standard spectral and prosody features using HMM based classifiers on the spontaneous FAU Aibo emotional speech corpus. The results show that unimodal classifiers with the WMFCC features perform significantly better than the classifiers with standard spectral features. Late decision fusion of classifiers provide further significant performance improvements. (C) 2011 Elsevier B.V. All rights reserved. | |
dc.description.indexedby | WOS | |
dc.description.indexedby | Scopus | |
dc.description.issue | 45208 | |
dc.description.openaccess | NO | |
dc.description.publisherscope | International | |
dc.description.sponsoredbyTubitakEu | N/A | |
dc.description.sponsorship | Turkish Scientific and Technical Research Council (TUBITAK) [106E201, COST2102, 110E056] This work was supported in part by the Turkish Scientific and Technical Research Council (TUBITAK) under projects 106E201 (COST2102 action) and 110E056. The authors would like to acknowledge and thank the anonymous referees for their valuable comments that have significantly improved the quality of the paper. | |
dc.description.volume | 53 | |
dc.identifier.doi | 10.1016/j.specom.2011.04.003 | |
dc.identifier.eissn | 1872-7182 | |
dc.identifier.issn | 0167-6393 | |
dc.identifier.quartile | Q2 | |
dc.identifier.scopus | 2-s2.0-79960848203 | |
dc.identifier.uri | https://doi.org/10.1016/j.specom.2011.04.003 | |
dc.identifier.uri | https://hdl.handle.net/20.500.14288/13099 | |
dc.identifier.wos | 294104000010 | |
dc.keywords | Emotion recognition | |
dc.keywords | Emotional speech classification | |
dc.keywords | Spectral features | |
dc.keywords | Formant frequency | |
dc.keywords | Line spectral frequency | |
dc.keywords | Decision fusion | |
dc.language.iso | eng | |
dc.publisher | Elsevier | |
dc.relation.ispartof | Speech Communication | |
dc.subject | Acoustics | |
dc.subject | Computer science | |
dc.title | Formant position based weighted spectral features for emotion recognition | |
dc.type | Journal Article | |
dspace.entity.type | Publication | |
local.contributor.kuauthor | Bozkurt, Elif | |
local.contributor.kuauthor | Erzin, Engin | |
local.publication.orgunit1 | GRADUATE SCHOOL OF SCIENCES AND ENGINEERING | |
local.publication.orgunit1 | College of Engineering | |
local.publication.orgunit2 | Department of Computer Engineering | |
local.publication.orgunit2 | Graduate School of Sciences and Engineering | |
relation.isOrgUnitOfPublication | 89352e43-bf09-4ef4-82f6-6f9d0174ebae | |
relation.isOrgUnitOfPublication | 3fc31c89-e803-4eb1-af6b-6258bc42c3d8 | |
relation.isOrgUnitOfPublication.latestForDiscovery | 89352e43-bf09-4ef4-82f6-6f9d0174ebae | |
relation.isParentOrgUnitOfPublication | 8e756b23-2d4a-4ce8-b1b3-62c794a8c164 | |
relation.isParentOrgUnitOfPublication | 434c9663-2b11-4e66-9399-c863e2ebae43 | |
relation.isParentOrgUnitOfPublication.latestForDiscovery | 8e756b23-2d4a-4ce8-b1b3-62c794a8c164 |