Back to Search
Start Over
Speech emotion recognition using MFCC-based entropy feature.
- Source :
- Signal, Image & Video Processing; Feb2024, Vol. 18 Issue 1, p153-161, 9p
- Publication Year :
- 2024
-
Abstract
- The prime objective of speech emotion recognition is to accurately recognize the emotion from the speech signal. It is a challenging task to accomplish. Speech emotion recognition (SER) has many applications, including medicine, online marketing, strengthening human–computer interaction (HCI), online education, and many more. Hence, it has been a topic of interest for many researchers for last three decades. The researchers used different methodologies to improve the classification accuracy of emotions. In this study, we tried to improve emotion classification accuracy using mel-frequency cepstral coefficient (MFCC)-based entropy features. First, we extracted the MFCC coefficient matrix from every speech of the EMO-DB, RAVDESS and SAVEE datasets, and then we calculated the proposed features: statistical mean ( MFCC mean ), MFCC-based approximate entropy ( MFCC AE ), and MFCC-based spectral entropy ( MFCC SE ), from the MFCC coefficient matrix of every utterance. The performance of the proposed features is accessed using the DNN classifier. We achieved a classification accuracy of 87.48%, 75.9%, and 79.64% using the combination of MFCC mean and MFCC SE features and obtained classification accuracies of 85.61%, 77.54%, and 76.26% using the combination of MFCC mean , MFCC AE , and MFCC SE features for the EMO-DB, RAVDESS, and SAVEE datasets, respectively. [ABSTRACT FROM AUTHOR]
Details
- Language :
- English
- ISSN :
- 18631703
- Volume :
- 18
- Issue :
- 1
- Database :
- Complementary Index
- Journal :
- Signal, Image & Video Processing
- Publication Type :
- Academic Journal
- Accession number :
- 175023504
- Full Text :
- https://doi.org/10.1007/s11760-023-02716-7