Emotion recognition using multi-parameter speech feature classification
Speech emotion recognition is basically extraction and identification of emotion from a speech signal. Speech data, corresponding to various emotions as happiness, sadness and anger, was recorded from 30 subjects. A local database called Amritaemo was created with 300 samples of speech waveforms cor...
Saved in:
Published in | 2015 International Conference on Computers, Communications, and Systems (ICCCS) pp. 217 - 222 |
---|---|
Main Authors | , , , , |
Format | Conference Proceeding |
Language | English |
Published |
IEEE
01.11.2015
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Abstract | Speech emotion recognition is basically extraction and identification of emotion from a speech signal. Speech data, corresponding to various emotions as happiness, sadness and anger, was recorded from 30 subjects. A local database called Amritaemo was created with 300 samples of speech waveforms corresponding to each emotion. Based on the prosodic features: energy contour and pitch contour, and spectral features: cepstral coefficients, quefrency coefficients and formant frequencies, the speech data was classified into respective emotions. The supervised learning method was used for training and testing, and the two algorithms used were Hybrid Rule based K-mean clustering and multiclass Support Vector Machine (SVM) algorithms. The results of the study showed that, for optimized set of features, Hybrid-rule based K mean clustering gave better performance compared to Multi class SVM. |
---|---|
AbstractList | Speech emotion recognition is basically extraction and identification of emotion from a speech signal. Speech data, corresponding to various emotions as happiness, sadness and anger, was recorded from 30 subjects. A local database called Amritaemo was created with 300 samples of speech waveforms corresponding to each emotion. Based on the prosodic features: energy contour and pitch contour, and spectral features: cepstral coefficients, quefrency coefficients and formant frequencies, the speech data was classified into respective emotions. The supervised learning method was used for training and testing, and the two algorithms used were Hybrid Rule based K-mean clustering and multiclass Support Vector Machine (SVM) algorithms. The results of the study showed that, for optimized set of features, Hybrid-rule based K mean clustering gave better performance compared to Multi class SVM. |
Author | Nair, G. J. Jeevitha, C. Y. Santhosh, Sini Poorna, S. S. Nair, Shyama Jayan |
Author_xml | – sequence: 1 givenname: S. S. surname: Poorna fullname: Poorna, S. S. email: poorna.surendran@gmail.com organization: Dept. of ECE, Amrita Vishwa Vidyapeetham Univ., Kollam, India – sequence: 2 givenname: C. Y. surname: Jeevitha fullname: Jeevitha, C. Y. organization: Dept. of ECE, Amrita Vishwa Vidyapeetham Univ., Kollam, India – sequence: 3 givenname: Shyama Jayan surname: Nair fullname: Nair, Shyama Jayan organization: Dept. of ECE, Amrita Vishwa Vidyapeetham Univ., Kollam, India – sequence: 4 givenname: Sini surname: Santhosh fullname: Santhosh, Sini organization: Dept. of ECE, Amrita Vishwa Vidyapeetham Univ., Kollam, India – sequence: 5 givenname: G. J. surname: Nair fullname: Nair, G. J. email: gjnair@am.amrita.edu organization: Dept. of ECE, Amrita Vishwa Vidyapeetham Univ., Kollam, India |
BookMark | eNotj8tOwzAURI0ECyj9Adj4BxL8TrxEUSmVirqg--rGXBdLiRPZzoK_59XVjI40R5o7ch2niIQ8cFZzzuxT1x3e3mvBuK4bbYRl6oqsbdNyZRppf1B7S7abcSphijShm84x_PUlh3im4zKUUM2QYMSCieYZ0X1Sj1CWhNQNkHPwwcHv5p7ceBgyri-5IseXzbF7rfaH7a573lfBslJZ8Ci5abUHyS1I4ZT0ygvJefvRWyaENqAd6AaNb4xC1Ghc33sN2mon5Io8_msDIp7mFEZIX6fLO_kNE7hKpw |
ContentType | Conference Proceeding |
DBID | 6IE 6IL CBEJK RIE RIL |
DOI | 10.1109/CCOMS.2015.7562904 |
DatabaseName | IEEE Electronic Library (IEL) Conference Proceedings IEEE Proceedings Order Plan All Online (POP All Online) 1998-present by volume IEEE Xplore All Conference Proceedings IEEE Electronic Library (IEL) IEEE Proceedings Order Plans (POP All) 1998-Present |
DatabaseTitleList | |
Database_xml | – sequence: 1 dbid: RIE name: IEEE Electronic Library (IEL) url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/ sourceTypes: Publisher |
DeliveryMethod | fulltext_linktorsrc |
EISBN | 9781467397568 1467397563 1467397547 9781467397544 |
EndPage | 222 |
ExternalDocumentID | 7562904 |
Genre | orig-research |
GroupedDBID | 6IE 6IL CBEJK RIE RIL |
ID | FETCH-LOGICAL-i90t-9afe31685fa319a32c43f4f23118db902256a5ca57e6f764ee5e6cbbf5a595c23 |
IEDL.DBID | RIE |
IngestDate | Thu Jun 29 18:37:09 EDT 2023 |
IsPeerReviewed | false |
IsScholarly | false |
Language | English |
LinkModel | DirectLink |
MergedId | FETCHMERGED-LOGICAL-i90t-9afe31685fa319a32c43f4f23118db902256a5ca57e6f764ee5e6cbbf5a595c23 |
PageCount | 6 |
ParticipantIDs | ieee_primary_7562904 |
PublicationCentury | 2000 |
PublicationDate | 2015-Nov. |
PublicationDateYYYYMMDD | 2015-11-01 |
PublicationDate_xml | – month: 11 year: 2015 text: 2015-Nov. |
PublicationDecade | 2010 |
PublicationTitle | 2015 International Conference on Computers, Communications, and Systems (ICCCS) |
PublicationTitleAbbrev | CCOMS |
PublicationYear | 2015 |
Publisher | IEEE |
Publisher_xml | – name: IEEE |
Score | 1.7032996 |
Snippet | Speech emotion recognition is basically extraction and identification of emotion from a speech signal. Speech data, corresponding to various emotions as... |
SourceID | ieee |
SourceType | Publisher |
StartPage | 217 |
SubjectTerms | Cepstral analysis cepstrum Emotion Emotion recognition Feature extraction hybrid rule based K-mean clustering pitch prosody quefrency Speech Speech recognition Support vector machines SVM Two dimensional displays |
Title | Emotion recognition using multi-parameter speech feature classification |
URI | https://ieeexplore.ieee.org/document/7562904 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3LSgMxFA21K1cqrfgmC5dmOuY1k3VpLUJVsEJ3JY8bFbEtdWbj15tk2oriwl0IgUlyGc5Ncs65CF16yaXQkhKqy5zw0uYkhFkTw7wuHaU50_FFd3wnR0_8diqmLXS11cIAQCKfQRab6S3fLWwdr8p6RQBrFc0_d8LBrdFqbXQwuer1-_fjx0jWEtl64I-KKQkwhntovPlUwxN5y-rKZPbzlwvjf-eyj7rf0jz8sAWdA9SCeQfdDJpiPHhLBwrtyGh_xokwSKLB93skvuCPJYB9wR6SoSe2MXmObKEUoC6aDAeT_oisKySQV5VXRGkPsfCU8Dr8SZpRy5nnPqRs16UzKsCzkFpYLQqQvpAcQIC0xnihhRKWskPUni_mcIQw80bpwoXDoNTc2kIp6kI2yFxpXGkhP0aduAezZeOBMVsv_-Tv7lO0G-PQaPbOULta1XAewLsyFylqX7WynVs |
linkProvider | IEEE |
linkToHtml | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3PT8IwFH4heNCTGjD-dgePdox17dYzAVEZmogJN9Ifr2qIQHRc_OttN8BoPHhrmibt-tJ8r-v3fQ_g0vKEM8ljEsssIkmmI-LCLImiVmYmjiMq_YtuPuT9p-R2zMY1uNpoYRCxJJ9h6JvlW76Z66X_VdZKHVgLb_655XCftSu11loJE4lWp3OfP3q6FgtXQ3_UTCkho7cL-XqyiikyDZeFCvXnLx_G_65mD5rf4rzgYQM7-1DDWQOuu1U5nmBDCHJtz2l_DkrKIPEW32-e-hJ8LBD1S2CxtPQMtE-fPV-oDFETRr3uqNMnqxoJ5FVEBRHSoi89xax0Z0nSWCfUJtYlbe3MKOEAmnHJtGQpcpvyBJEh10pZJplgOqYHUJ_NZ3gIAbVKyNS46yCXidapELFx-SA1mTKZxugIGn4PJovKBWOy-vzjv7svYLs_ygeTwc3w7gR2fEwqBd8p1Iv3JZ45KC_UeRnBLxm3oKQ |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=2015+International+Conference+on+Computers%2C+Communications%2C+and+Systems+%28ICCCS%29&rft.atitle=Emotion+recognition+using+multi-parameter+speech+feature+classification&rft.au=Poorna%2C+S.+S.&rft.au=Jeevitha%2C+C.+Y.&rft.au=Nair%2C+Shyama+Jayan&rft.au=Santhosh%2C+Sini&rft.date=2015-11-01&rft.pub=IEEE&rft.spage=217&rft.epage=222&rft_id=info:doi/10.1109%2FCCOMS.2015.7562904&rft.externalDocID=7562904 |