Abstract
Emotion recognition from speech is an important aspect of human-computer interaction (HCI) systems,
allowing machines to better understand human emotions and respond accordingly. This paper explores the use of
machine learning techniques to recognize emotions in speech signals. We leverage the librosa library for feature
extraction from audio files and train multiple machine learning models, including Support Vector Machine (SVM),
Random Forest (RF), and k-Nearest Neighbors (k-NN), to classify speech emotions. The aim is to create an
automated system capable of identifying emotions like happy, sad, angry, neutral, and surprised from speech audio.