Publication

Speech emotion recognition

Ahmed, Fatima
Citations
Altmetric:
Supervisor
Subject
Date
Research Projects
Organizational Units
Journal Issue
Abstract
In this study, a deep learning model was developed to recognize emotions in speech. The model used a combination of Mel-Frequency Cepstral Coefficients (MFCCs) for feature extraction and a Long Short-Term Memory (LSTM) layer to capture contextual information. The model was trained and tested on a speech database called the Toronto Emotional Speech Set. The results showed that the model was able to achieve high accuracy in emotion recognition, exceeding 95% for model accuracy and 97% for validation accuracy. The authors suggest that this type of model could be used to improve the ability of AI systems to understand and respond to human emotions, potentially enhancing the user experience in tasks such as voice commands, messaging, and recommendation systems.
Sponsor
Effat
Copyright
Book title
Journal title
DOI
Embedded videos