Stress detection based on TEO and MFCC speech features using Convolutional Neural Networks (CNN)
The effect of stress on mental and physical health is very concerning making it a fascinating and socially valuable field of study nowadays. Although a number of stress markers have been deployed, there are still issues involved with using these kinds of approaches. By developing a speech-based...
Saved in:
Main Authors: | , , , , , , |
---|---|
Format: | Conference or Workshop Item |
Language: | English |
Published: |
2022
|
Subjects: | |
Online Access: | http://irep.iium.edu.my/102404/1/Stress%20Detection%20Based%20on%20Teo%20and%20MFCC%20Speech%20Features%20Using%20Convolutional%20Neural%20Networks%20%28CNN%29.pdf http://irep.iium.edu.my/102404/ |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Universiti Islam Antarabangsa Malaysia |
Language: | English |
Summary: | The effect of stress on mental and physical health is
very concerning making it a fascinating and socially valuable field
of study nowadays. Although a number of stress markers have
been deployed, there are still issues involved with using these kinds
of approaches. By developing a speech-based stress detection
system, it could solve the problems faced by other currently
available methods of detecting stress since it is a non-invasive and
contactless approach. In this work, a fusion of Teager Energy
Operator (TEO) and Mel Frequency Cepstral Coefficients
(MFCC) namely Teager-MFCC (T-MFCC) are proposed as the
speech features to be extracted from speech signals in recognizing
stressed emotions. Since stressed emotions affect the nonlinear
components of speech, TEO is applied to reflect the instantaneous
energy of the components. Convolutional Neural Network (CNN)
classifier is used with the proposed T- MFCC features on the
Ryerson Audio-Visual Database of Emotional Speech and Song
(RAVDESS) corpus. The proposed method (T-MFCC) had shown
a better performance with classification accuracies of 95.83% and
95.37% for male and female speakers respectively compared to the
MFCC feature extraction technique which achieves 84.26% (male)
and 93.98% (female) classification accuracies. |
---|