Publication:
Improving automatic emotion recognition from speech signals

dc.contributor.coauthorErdem, Çiǧdem Eroǧlu
dc.contributor.coauthorErdem, A. Tanju
dc.contributor.departmentDepartment of Computer Engineering
dc.contributor.departmentGraduate School of Sciences and Engineering
dc.contributor.kuauthorBozkurt, Elif
dc.contributor.kuauthorErzin, Engin
dc.contributor.schoolcollegeinstituteCollege of Engineering
dc.contributor.schoolcollegeinstituteGRADUATE SCHOOL OF SCIENCES AND ENGINEERING
dc.date.accessioned2024-11-09T23:46:45Z
dc.date.issued2009
dc.description.abstractWe present a speech signal driven emotion recognition system. Our system is trained and tested with the INTERSPEECH 2009 Emotion Challenge corpus, which includes spontaneous and emotionally rich recordings. The challenge includes classifier and feature sub-challenges with five-class and two-class classification problems. We investigate prosody related, spectral and HMM-based features for the evaluation of emotion recognition with Gaussian mixture model (GMM) based classifiers. Spectral features consist of mel-scale cepstral coefficients (MFCC), line spectral frequency (LSF) features and their derivatives, whereas prosody-related features consist of mean normalized values of pitch, first derivative of pitch and intensity. Unsupervised training of HMM structures are employed to define prosody related temporal features for the emotion recognition problem. We also investigate data fusion of different features and decision fusion of different classifiers, which are not well studied for emotion recognition framework. Experimental results of automatic emotion recognition with the INTERSPEECH 2009 Emotion Challenge corpus are presented. Copyright © 2009 ISCA.
dc.description.indexedbyScopus
dc.description.indexedbyWOS
dc.description.openaccessYES
dc.description.publisherscopeInternational
dc.description.sponsoredbyTubitakEuN/A
dc.identifier.issn1990-9772
dc.identifier.linkhttps://www.scopus.com/inward/record.uri?eid=2-s2.0-70450177656andpartnerID=40andmd5=9a09582190febf6e1f7d59bc66e68432
dc.identifier.quartileN/A
dc.identifier.scopus2-s2.0-70450177656
dc.identifier.urihttps://hdl.handle.net/20.500.14288/14007
dc.identifier.wos276842800076
dc.keywordsEmotion recognition
dc.keywordsProsody modeling Cepstral coefficients
dc.keywordsDecision fusion
dc.keywordsEmotion recognition
dc.keywordsFirst derivative
dc.keywordsGaussian Mixture Model
dc.keywordsLine spectral frequencies
dc.keywordsNormalized values
dc.keywordsProsody modeling
dc.keywordsSpectral feature
dc.keywordsSpeech signals
dc.keywordsTemporal features
dc.keywordsTwo-class classification problems
dc.keywordsUnsupervised training
dc.keywordsClassifiers
dc.keywordsFace recognition
dc.keywordsHidden Markov models
dc.keywordsLearning systems
dc.keywordsSpeech communication
dc.keywordsSpeech recognition
dc.language.isoeng
dc.publisherINTERSPEECH
dc.relation.ispartofProceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH
dc.subjectComputer engineering
dc.titleImproving automatic emotion recognition from speech signals
dc.typeConference Proceeding
dspace.entity.typePublication
local.contributor.kuauthorErzin, Engin
local.contributor.kuauthorBozkurt, Elif
local.publication.orgunit1College of Engineering
local.publication.orgunit1GRADUATE SCHOOL OF SCIENCES AND ENGINEERING
local.publication.orgunit2Department of Computer Engineering
local.publication.orgunit2Graduate School of Sciences and Engineering
relation.isOrgUnitOfPublication89352e43-bf09-4ef4-82f6-6f9d0174ebae
relation.isOrgUnitOfPublication3fc31c89-e803-4eb1-af6b-6258bc42c3d8
relation.isOrgUnitOfPublication.latestForDiscovery89352e43-bf09-4ef4-82f6-6f9d0174ebae
relation.isParentOrgUnitOfPublication8e756b23-2d4a-4ce8-b1b3-62c794a8c164
relation.isParentOrgUnitOfPublication434c9663-2b11-4e66-9399-c863e2ebae43
relation.isParentOrgUnitOfPublication.latestForDiscovery8e756b23-2d4a-4ce8-b1b3-62c794a8c164

Files