ISSN : 2663-2187

IMPLEMENTATION OF EMOTION PREDICTION THROUGH SPEECH USING NEURAL NETWORK

Main Article Content

S BABU, M.E., (PH.D)., DR. K. LOGESH, M.TECH., PH.D., V POORNIMA
ยป doi: 10.33472/AFJBS.6.10.2024.801-806

Abstract

The study explores the field of Speech Emotion Recognition (SER), which aims to identify human emotions from speech patterns. The abstract details the methodology and the technological framework utilized to achieve this goal. The focus is on the extraction of audio features from speech samples, utilizing a precise and clear database of actors' voices devoid of background noise. This is crucial for the accuracy of the SER system. An array of classifier algorithms is discussed, with a particular emphasis on the importance of selecting the most effective classifier to enhance the recognition process. Among various audio features, Mel-Frequency Cepstral Coefficients (MFCC), Mel Spectrograms, and chroma features are highlighted as critical in identifying emotions. These features capture essential elements of sound that are indicative of emotional states. The project employs a neural network approach to classify emotions based on the extracted audio signals. This system has been proposed for use in several practical applications, including call centers, educational settings, and support for physically disabled individuals, where understanding emotional cues is crucial for effective communication. The conclusion reaffirms the importance of a robust database, effective feature extraction, and a precise classification model in determining the success of SER systems. The proposed model uses advanced feature extraction techniques and neural network classifiers to ensure a high degree of accuracy in emotion recognition. This research not only contributes to the technical advancements in SER but also underscores the potential applications of this technology in enhancing interpersonal interactions and accessibility through emotional understanding

Article Details