Publication: Improving automatic emotion recognition from speech signals
dc.contributor.coauthor | Erdem, Çiǧdem Eroǧlu | |
dc.contributor.coauthor | Erdem, A. Tanju | |
dc.contributor.department | Department of Computer Engineering | |
dc.contributor.department | Graduate School of Sciences and Engineering | |
dc.contributor.kuauthor | Bozkurt, Elif | |
dc.contributor.kuauthor | Erzin, Engin | |
dc.contributor.schoolcollegeinstitute | College of Engineering | |
dc.contributor.schoolcollegeinstitute | GRADUATE SCHOOL OF SCIENCES AND ENGINEERING | |
dc.date.accessioned | 2024-11-09T23:46:45Z | |
dc.date.issued | 2009 | |
dc.description.abstract | We present a speech signal driven emotion recognition system. Our system is trained and tested with the INTERSPEECH 2009 Emotion Challenge corpus, which includes spontaneous and emotionally rich recordings. The challenge includes classifier and feature sub-challenges with five-class and two-class classification problems. We investigate prosody related, spectral and HMM-based features for the evaluation of emotion recognition with Gaussian mixture model (GMM) based classifiers. Spectral features consist of mel-scale cepstral coefficients (MFCC), line spectral frequency (LSF) features and their derivatives, whereas prosody-related features consist of mean normalized values of pitch, first derivative of pitch and intensity. Unsupervised training of HMM structures are employed to define prosody related temporal features for the emotion recognition problem. We also investigate data fusion of different features and decision fusion of different classifiers, which are not well studied for emotion recognition framework. Experimental results of automatic emotion recognition with the INTERSPEECH 2009 Emotion Challenge corpus are presented. Copyright © 2009 ISCA. | |
dc.description.indexedby | Scopus | |
dc.description.indexedby | WOS | |
dc.description.openaccess | YES | |
dc.description.publisherscope | International | |
dc.description.sponsoredbyTubitakEu | N/A | |
dc.identifier.issn | 1990-9772 | |
dc.identifier.link | https://www.scopus.com/inward/record.uri?eid=2-s2.0-70450177656andpartnerID=40andmd5=9a09582190febf6e1f7d59bc66e68432 | |
dc.identifier.quartile | N/A | |
dc.identifier.scopus | 2-s2.0-70450177656 | |
dc.identifier.uri | https://hdl.handle.net/20.500.14288/14007 | |
dc.identifier.wos | 276842800076 | |
dc.keywords | Emotion recognition | |
dc.keywords | Prosody modeling Cepstral coefficients | |
dc.keywords | Decision fusion | |
dc.keywords | Emotion recognition | |
dc.keywords | First derivative | |
dc.keywords | Gaussian Mixture Model | |
dc.keywords | Line spectral frequencies | |
dc.keywords | Normalized values | |
dc.keywords | Prosody modeling | |
dc.keywords | Spectral feature | |
dc.keywords | Speech signals | |
dc.keywords | Temporal features | |
dc.keywords | Two-class classification problems | |
dc.keywords | Unsupervised training | |
dc.keywords | Classifiers | |
dc.keywords | Face recognition | |
dc.keywords | Hidden Markov models | |
dc.keywords | Learning systems | |
dc.keywords | Speech communication | |
dc.keywords | Speech recognition | |
dc.language.iso | eng | |
dc.publisher | INTERSPEECH | |
dc.relation.ispartof | Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH | |
dc.subject | Computer engineering | |
dc.title | Improving automatic emotion recognition from speech signals | |
dc.type | Conference Proceeding | |
dspace.entity.type | Publication | |
local.contributor.kuauthor | Erzin, Engin | |
local.contributor.kuauthor | Bozkurt, Elif | |
local.publication.orgunit1 | College of Engineering | |
local.publication.orgunit1 | GRADUATE SCHOOL OF SCIENCES AND ENGINEERING | |
local.publication.orgunit2 | Department of Computer Engineering | |
local.publication.orgunit2 | Graduate School of Sciences and Engineering | |
relation.isOrgUnitOfPublication | 89352e43-bf09-4ef4-82f6-6f9d0174ebae | |
relation.isOrgUnitOfPublication | 3fc31c89-e803-4eb1-af6b-6258bc42c3d8 | |
relation.isOrgUnitOfPublication.latestForDiscovery | 89352e43-bf09-4ef4-82f6-6f9d0174ebae | |
relation.isParentOrgUnitOfPublication | 8e756b23-2d4a-4ce8-b1b3-62c794a8c164 | |
relation.isParentOrgUnitOfPublication | 434c9663-2b11-4e66-9399-c863e2ebae43 | |
relation.isParentOrgUnitOfPublication.latestForDiscovery | 8e756b23-2d4a-4ce8-b1b3-62c794a8c164 |