Journal article Open Access
Hafsa Qazi; Baij Nath Kaushik
Automatic speech emotion recognition is a very necessary activity for effective human-computer interaction. This paper is motivated by using spectrograms as inputs to the hybrid deep convolutional LSTM for speech emotion recognition. In this study, we trained our proposed model using four convolutional layers for high-level feature extraction from input spectrograms, LSTM layer for accumulating long-term dependencies and finally two dense layers. Experimental results on the SAVEE database shows promising performance. Our proposed model is highly capable as it obtained an accuracy of 94.26%.
Name | Size | |
---|---|---|
E1027069520.pdf
md5:44bfd44793f531deb0b1eba97f3c9289 |
611.2 kB | Download |
Views | 61 |
Downloads | 40 |
Data volume | 24.4 MB |
Unique views | 60 |
Unique downloads | 40 |