A Spectrogram and Local Feature-Assisted Convolutional Neural Network for Amharic Speech Emotion Identification

Read the full article See related articles

Discuss this preprint

Start a discussion What are Sciety discussions?

Listed in

This article is not in any list yet, why not save it to one of your lists.
Log in to save this article

Abstract

Speech Emotion Recognition (SER) plays a significant role in improving human-computer interaction and human-human communication. Nevertheless, speech emotion recognition in low-resource languages like Amharic is still a difficult task because of the lack of datasets and language diversity. In this paper, a Convolutional Neural Network (CNN)-based approach, which combines spectrogram features and local acoustic features such as Mel-Frequency Cepstral Coefficients (MFCCs), chroma, zero-crossing rate (ZCR), energy, and pitch, is proposed for efficient Amharic speech emotion recognition. A dataset of 1,650 three-second Amharic speech samples was created, and the samples were labeled with five emotional classes: anger, fear, happy, neutral, and sad. Advanced preprocessing methods such as spectral subtraction and wavelet denoising were used to improve the quality of the signals and speed up the training process. The experimental results show that the proposed CNN-based approach has a classification accuracy of 90 percent, which is better than the recurrent neural network-based approaches: Long Short-Term Memory (LSTM) with 58.48 percent, Bidirectional Long Short-Term Memory (BiLSTM) with 63.33 percent, and Gated Recurrent Unit (GRU) with 40 percent, as well as the single-feature models: local acoustic features with 73 percent and spectrogram features with 79 percent. These results confirm that the integration of spectrogram and local acoustic features within a CNN architecture improves accuracy and efficiency in speech emotion recognition in low-resource languages, setting a standard for future Amharic SER research.

Article activity feed