Multi-label annotation of music
| dc.contributor.author | Ahsan, H. | |
| dc.contributor.author | Kumar, V. | |
| dc.contributor.author | Jawahar, C.V. | |
| dc.date.accessioned | 2026-02-06T06:39:34Z | |
| dc.date.issued | 2015 | |
| dc.description.abstract | Automatic annotation of an audio or a music piece with multiple labels helps in understanding the composition of a music. Such meta-level information can be very useful in applications such as music transcription, retrieval, organization and personalization. In this work, we formulate the problem of annotation as multi-label classification which is considerably different from that of a popular single (binary or multi-class) label classification. We employ both the nearest neighbour and max-margin (SVM) formulations for the automatic annotation. We consider K-NN and SVM that are adapted for multi-label classification using one-vs-rest strategy and a direct multi-label classification formulation using ML-KNN and M3L. In the case of music, often the signatures of the labels (e.g. instruments and vocal signatures) are fused in the features. We therefore propose a simple feature augmentation technique based on non-negative matrix factorization (NMF) with an intuition to decompose a music piece into its constituent components. We conducted our experiments on two data sets - Indian classical instruments dataset and Emotions dataset [1], and validate the methods. © 2015 IEEE. | |
| dc.identifier.citation | ICAPR 2015 - 2015 8th International Conference on Advances in Pattern Recognition, 2015, Vol., , p. - | |
| dc.identifier.uri | https://doi.org/10.1109/ICAPR.2015.7050685 | |
| dc.identifier.uri | https://idr.nitk.ac.in/handle/123456789/32384 | |
| dc.publisher | Institute of Electrical and Electronics Engineers Inc. | |
| dc.subject | multi-label classification | |
| dc.subject | Music annotation | |
| dc.title | Multi-label annotation of music |
