Artificial neural networks & discrete Wavelet transform enabled healthcare model for stress and emotion assessment using speech signal recognition

Stress and emotion assessment play a pivotal role in healthcare, aiding in the early detection and prevention of various diseases. In this research paper, we propose a novel healthcare model that harnesses the power of artificial neural networks (ANN) and discrete wavelet transform (DWT) to assess s...

Full description

Saved in:
Bibliographic Details
Published inAIP conference proceedings Vol. 3072; no. 1
Main Authors Vashishth, Tarun Kumar, Sharma, Vikas, Sharma, Kewal Krishan, Chaudhary, Sachin, Kumar, Bhupendra, Panwar, Rajneesh
Format Journal Article Conference Proceeding
LanguageEnglish
Published Melville American Institute of Physics 19.03.2024
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Stress and emotion assessment play a pivotal role in healthcare, aiding in the early detection and prevention of various diseases. In this research paper, we propose a novel healthcare model that harnesses the power of artificial neural networks (ANN) and discrete wavelet transform (DWT) to assess stress and emotion through speech signal recognition. The proposed model extracts essential features from speech signals using DWT and employs ANN for the classification of stress and emotional states. Both ANN and DWT are established signal processing techniques that find applications in diverse domains. We present a comprehensive analysis of our healthcare model, which effectively analyzes speech signals to identify the stress and emotional states of individuals. The model’s performance is rigorously evaluated using a publicly available dataset comprising speech signals recorded from subjects experiencing varying stress and emotional states. Experimental results demonstrate that the proposed model achieves high accuracy and surpasses existing state-of-the-art techniques. By employing recurrent neural networks (RNN), we visualize speech’s phonetic characteristics, such as Mel-frequency cepstral coefficients (MFCCs), transforming them into sequences of phonemes or words. RNNs, designed to handle sequential data, form the basis of our speech recognition using neural networks, contributing to the model’s effectiveness and accuracy.
Bibliography:ObjectType-Conference Proceeding-1
SourceType-Conference Papers & Proceedings-1
content type line 21
ISSN:0094-243X
1551-7616
DOI:10.1063/5.0198725