Show simple item record

dc.contributor.authorBozkurt, E.
dc.contributor.authorErzin, E.
dc.contributor.authorEroğlu Erdem, Ç.
dc.contributor.authorErdem, Tanju
dc.date.accessioned2016-02-11T06:46:20Z
dc.date.available2016-02-11T06:46:20Z
dc.date.issued2009
dc.identifier.isbn978-1-61567-692-7
dc.identifier.urihttp://hdl.handle.net/10679/2007
dc.description.abstractWe present a speech signal driven emotion recognition system. Our system is trained and tested with the INTERSPEECH 2009 Emotion Challenge corpus, which includes spontaneous and emotionally rich recordings. The challenge includes classifier and feature sub-challenges with five-class and two-class classification problems. We investigate prosody related, spectral and HMM-based features for the evaluation of emotion recognition with Gaussian mixture model (GMM) based classifiers. Spectral features consist of mel-scale cepstral coefficients (MFCC), line spectral frequency (LSF) features and their derivatives, whereas prosody-related features consist of mean normalized values of pitch, first derivative of pitch and intensity. Unsupervised training of HMM structures are employed to define prosody related temporal features for the emotion recognition problem. We also investigate data fusion of different features and decision fusion of different classifiers, which are not well studied for emotion recognition framework. Experimental results of automatic emotion recognition with the INTERSPEECH 2009 Emotion Challenge corpus are presented.
dc.description.sponsorshipTÜBİTAK
dc.language.isoengen_US
dc.publisherInternational Speech Communications Association
dc.relationinfo:turkey/grantAgreement/TUBITAK/106E201
dc.relationinfo:turkey/grantAgreement/TUBITAK/3070796
dc.relation.ispartof10th Annual Conference Of The International Speech Communication Association 2009 (INTERSPEECH 2009)
dc.rightsrestrictedAccess
dc.titleImproving automatic emotion recognition from speech signalsen_US
dc.typeConference paperen_US
dc.publicationstatuspublisheden_US
dc.contributor.departmentÖzyeğin University
dc.contributor.authorID(ORCID 0000-0002-8841-1642 & YÖK ID 45777) Erdem, Tanju
dc.contributor.ozuauthorErdem, Tanju
dc.identifier.startpage312
dc.identifier.endpage315
dc.identifier.wosWOS:000276842800076
dc.subject.keywordsEmotion recognition
dc.subject.keywordsProsody modeling
dc.identifier.scopusSCOPUS:2-s2.0-70450177656
dc.contributor.authorMale1
dc.relation.publicationcategoryConference Paper - International - Institutional Academic Staff


Files in this item

FilesSizeFormatView

There are no files associated with this item.

This item appears in the following Collection(s)

Show simple item record


Share this page