Classification of vocal and non-vocal regions from audio songs using spectral features and pitch variations

dc.contributor.authorVishnu Srinivasa Murthy, Y.V.S.
dc.contributor.authorKoolagudi, S.G.
dc.date.accessioned2026-02-06T06:39:32Z
dc.date.issued2015
dc.description.abstractIn this work, an effort has been made to identify vocal and non-vocal regions from a given song using signal processing techniques and machine learning algorithm. Initially spectral features like mel-frequency cepstral coefficients (MFCCs) are used to develop the baseline system. Statistical values of pitch, jitter and shimmer are considered to improve performance of the system. Artificial neural networks (ANNs) are used to capture the characteristics of vocal and non-vocal segments of the songs. The experiment is conducted on 60 vocal and 60 non-vocal clips extracted from Telugu albums. 11-point moving window is used to ensure the continuity of vocal and non-vocal segments, thus improving the accuracy of system. With this approach system achieves 85.59% accuracy for vocal and 88.52% for non-vocal segment classification. © 2015 IEEE.
dc.identifier.citationCanadian Conference on Electrical and Computer Engineering, 2015, Vol.2015-June, June, p. 1271-1276
dc.identifier.issn8407789
dc.identifier.urihttps://doi.org/10.1109/CCECE.2015.7129461
dc.identifier.urihttps://idr.nitk.ac.in/handle/123456789/32359
dc.publisherInstitute of Electrical and Electronics Engineers Inc.
dc.subjectArtificial neural networks
dc.subjectJitter
dc.subjectMel-frequency cepstral coefficients
dc.subjectNon-vocal regions
dc.subjectPitch
dc.subjectShimmer
dc.subjectVocal regions
dc.titleClassification of vocal and non-vocal regions from audio songs using spectral features and pitch variations

Files