Emotion recognition using multi-parameter speech feature classification

Speech emotion recognition is basically extraction and identification of emotion from a speech signal. Speech data, corresponding to various emotions as happiness, sadness and anger, was recorded from 30 subjects. A local database called Amritaemo was created with 300 samples of speech waveforms cor...

Full description

Saved in:
Bibliographic Details
Published in2015 International Conference on Computers, Communications, and Systems (ICCCS) pp. 217 - 222
Main Authors Poorna, S. S., Jeevitha, C. Y., Nair, Shyama Jayan, Santhosh, Sini, Nair, G. J.
Format Conference Proceeding
LanguageEnglish
Published IEEE 01.11.2015
Subjects
Online AccessGet full text

Cover

Loading…
Abstract Speech emotion recognition is basically extraction and identification of emotion from a speech signal. Speech data, corresponding to various emotions as happiness, sadness and anger, was recorded from 30 subjects. A local database called Amritaemo was created with 300 samples of speech waveforms corresponding to each emotion. Based on the prosodic features: energy contour and pitch contour, and spectral features: cepstral coefficients, quefrency coefficients and formant frequencies, the speech data was classified into respective emotions. The supervised learning method was used for training and testing, and the two algorithms used were Hybrid Rule based K-mean clustering and multiclass Support Vector Machine (SVM) algorithms. The results of the study showed that, for optimized set of features, Hybrid-rule based K mean clustering gave better performance compared to Multi class SVM.
AbstractList Speech emotion recognition is basically extraction and identification of emotion from a speech signal. Speech data, corresponding to various emotions as happiness, sadness and anger, was recorded from 30 subjects. A local database called Amritaemo was created with 300 samples of speech waveforms corresponding to each emotion. Based on the prosodic features: energy contour and pitch contour, and spectral features: cepstral coefficients, quefrency coefficients and formant frequencies, the speech data was classified into respective emotions. The supervised learning method was used for training and testing, and the two algorithms used were Hybrid Rule based K-mean clustering and multiclass Support Vector Machine (SVM) algorithms. The results of the study showed that, for optimized set of features, Hybrid-rule based K mean clustering gave better performance compared to Multi class SVM.
Author Nair, G. J.
Jeevitha, C. Y.
Santhosh, Sini
Poorna, S. S.
Nair, Shyama Jayan
Author_xml – sequence: 1
  givenname: S. S.
  surname: Poorna
  fullname: Poorna, S. S.
  email: poorna.surendran@gmail.com
  organization: Dept. of ECE, Amrita Vishwa Vidyapeetham Univ., Kollam, India
– sequence: 2
  givenname: C. Y.
  surname: Jeevitha
  fullname: Jeevitha, C. Y.
  organization: Dept. of ECE, Amrita Vishwa Vidyapeetham Univ., Kollam, India
– sequence: 3
  givenname: Shyama Jayan
  surname: Nair
  fullname: Nair, Shyama Jayan
  organization: Dept. of ECE, Amrita Vishwa Vidyapeetham Univ., Kollam, India
– sequence: 4
  givenname: Sini
  surname: Santhosh
  fullname: Santhosh, Sini
  organization: Dept. of ECE, Amrita Vishwa Vidyapeetham Univ., Kollam, India
– sequence: 5
  givenname: G. J.
  surname: Nair
  fullname: Nair, G. J.
  email: gjnair@am.amrita.edu
  organization: Dept. of ECE, Amrita Vishwa Vidyapeetham Univ., Kollam, India
BookMark eNotj8tOwzAURI0ECyj9Adj4BxL8TrxEUSmVirqg--rGXBdLiRPZzoK_59XVjI40R5o7ch2niIQ8cFZzzuxT1x3e3mvBuK4bbYRl6oqsbdNyZRppf1B7S7abcSphijShm84x_PUlh3im4zKUUM2QYMSCieYZ0X1Sj1CWhNQNkHPwwcHv5p7ceBgyri-5IseXzbF7rfaH7a573lfBslJZ8Ci5abUHyS1I4ZT0ygvJefvRWyaENqAd6AaNb4xC1Ghc33sN2mon5Io8_msDIp7mFEZIX6fLO_kNE7hKpw
ContentType Conference Proceeding
DBID 6IE
6IL
CBEJK
RIE
RIL
DOI 10.1109/CCOMS.2015.7562904
DatabaseName IEEE Electronic Library (IEL) Conference Proceedings
IEEE Proceedings Order Plan All Online (POP All Online) 1998-present by volume
IEEE Xplore All Conference Proceedings
IEEE Electronic Library (IEL)
IEEE Proceedings Order Plans (POP All) 1998-Present
DatabaseTitleList
Database_xml – sequence: 1
  dbid: RIE
  name: IEEE Electronic Library (IEL)
  url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/
  sourceTypes: Publisher
DeliveryMethod fulltext_linktorsrc
EISBN 9781467397568
1467397563
1467397547
9781467397544
EndPage 222
ExternalDocumentID 7562904
Genre orig-research
GroupedDBID 6IE
6IL
CBEJK
RIE
RIL
ID FETCH-LOGICAL-i90t-9afe31685fa319a32c43f4f23118db902256a5ca57e6f764ee5e6cbbf5a595c23
IEDL.DBID RIE
IngestDate Thu Jun 29 18:37:09 EDT 2023
IsPeerReviewed false
IsScholarly false
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-i90t-9afe31685fa319a32c43f4f23118db902256a5ca57e6f764ee5e6cbbf5a595c23
PageCount 6
ParticipantIDs ieee_primary_7562904
PublicationCentury 2000
PublicationDate 2015-Nov.
PublicationDateYYYYMMDD 2015-11-01
PublicationDate_xml – month: 11
  year: 2015
  text: 2015-Nov.
PublicationDecade 2010
PublicationTitle 2015 International Conference on Computers, Communications, and Systems (ICCCS)
PublicationTitleAbbrev CCOMS
PublicationYear 2015
Publisher IEEE
Publisher_xml – name: IEEE
Score 1.7032996
Snippet Speech emotion recognition is basically extraction and identification of emotion from a speech signal. Speech data, corresponding to various emotions as...
SourceID ieee
SourceType Publisher
StartPage 217
SubjectTerms Cepstral analysis
cepstrum
Emotion
Emotion recognition
Feature extraction
hybrid rule based K-mean clustering
pitch
prosody
quefrency
Speech
Speech recognition
Support vector machines
SVM
Two dimensional displays
Title Emotion recognition using multi-parameter speech feature classification
URI https://ieeexplore.ieee.org/document/7562904
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3LSgMxFA21K1cqrfgmC5dmOuY1k3VpLUJVsEJ3JY8bFbEtdWbj15tk2oriwl0IgUlyGc5Ncs65CF16yaXQkhKqy5zw0uYkhFkTw7wuHaU50_FFd3wnR0_8diqmLXS11cIAQCKfQRab6S3fLWwdr8p6RQBrFc0_d8LBrdFqbXQwuer1-_fjx0jWEtl64I-KKQkwhntovPlUwxN5y-rKZPbzlwvjf-eyj7rf0jz8sAWdA9SCeQfdDJpiPHhLBwrtyGh_xokwSKLB93skvuCPJYB9wR6SoSe2MXmObKEUoC6aDAeT_oisKySQV5VXRGkPsfCU8Dr8SZpRy5nnPqRs16UzKsCzkFpYLQqQvpAcQIC0xnihhRKWskPUni_mcIQw80bpwoXDoNTc2kIp6kI2yFxpXGkhP0aduAezZeOBMVsv_-Tv7lO0G-PQaPbOULta1XAewLsyFylqX7WynVs
linkProvider IEEE
linkToHtml http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3PT8IwFH4heNCTGjD-dgePdox17dYzAVEZmogJN9Ifr2qIQHRc_OttN8BoPHhrmibt-tJ8r-v3fQ_g0vKEM8ljEsssIkmmI-LCLImiVmYmjiMq_YtuPuT9p-R2zMY1uNpoYRCxJJ9h6JvlW76Z66X_VdZKHVgLb_655XCftSu11loJE4lWp3OfP3q6FgtXQ3_UTCkho7cL-XqyiikyDZeFCvXnLx_G_65mD5rf4rzgYQM7-1DDWQOuu1U5nmBDCHJtz2l_DkrKIPEW32-e-hJ8LBD1S2CxtPQMtE-fPV-oDFETRr3uqNMnqxoJ5FVEBRHSoi89xax0Z0nSWCfUJtYlbe3MKOEAmnHJtGQpcpvyBJEh10pZJplgOqYHUJ_NZ3gIAbVKyNS46yCXidapELFx-SA1mTKZxugIGn4PJovKBWOy-vzjv7svYLs_ygeTwc3w7gR2fEwqBd8p1Iv3JZ45KC_UeRnBLxm3oKQ
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=2015+International+Conference+on+Computers%2C+Communications%2C+and+Systems+%28ICCCS%29&rft.atitle=Emotion+recognition+using+multi-parameter+speech+feature+classification&rft.au=Poorna%2C+S.+S.&rft.au=Jeevitha%2C+C.+Y.&rft.au=Nair%2C+Shyama+Jayan&rft.au=Santhosh%2C+Sini&rft.date=2015-11-01&rft.pub=IEEE&rft.spage=217&rft.epage=222&rft_id=info:doi/10.1109%2FCCOMS.2015.7562904&rft.externalDocID=7562904