Music emotion classification based on vocal and instrumental sound features using artificial neural network / Nurlaila Rosli

Classifying emotion in a song remains as a challenge in various area of research. Most of existing work in music emotion classification (MEC) done by looking at features such as audio, lyrics, social tags or combination of two or more features as stated above. There were only few studies on MEC that...

Full description

Saved in:
Bibliographic Details
Main Author: Rosli, Nurlaila
Format: Thesis
Language:English
Published: 2013
Subjects:
Online Access:https://ir.uitm.edu.my/id/eprint/64250/1/64250.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Classifying emotion in a song remains as a challenge in various area of research. Most of existing work in music emotion classification (MEC) done by looking at features such as audio, lyrics, social tags or combination of two or more features as stated above. There were only few studies on MEC that exploit timbre features from vocal part of the song. Thus, this research present works on classifying emotion in music by extracting timbre features from both vocal and instrumental sound clips. Three timbre features, namely spectral centroid, spectral rolloff and zero-cross are extracted based on its attribute in distinguishing between sad audio features and happy audio features. The final system is able to use all of the musical timbre features extracted from vocal part and instrumental part of a song, as to classify the type of emotion in selected Malay popular music. For training and testing purposes, this system is using an Artificial Neural Network (ANN). The percentages of emotion classified in Malay popular songs are projected to be higher when both vocal and instrumental sound features are applied to the ANN classifier. The findings of this research will collectively improve MEC based on manipulation of vocal and instrumental sound timbre features, as well as contributing towards the literature of music information retrieval, affective computing and psychology. However, it is suggested that this research must be incorporated with others features, such as rhythm and spectrum along with timbre features. It is also suggested that other emotion such as anger, calmness, sorrow and etc must be considered for the improvement of this research in the future.