Conference Papers
Permanent URI for this collectionhttps://idr.nitk.ac.in/handle/123456789/28506
Browse
5 results
Search Results
Item Raga classification for Carnatic music(Springer Verlag service@springer.de, 2015) Suma, S.M.; Koolagudi, S.G.In this work, an effort has been made to identify raga of given piece of Carnatic music. In the proposed method, direct raga classification without the use of note sequence has been performed using pitch as the primary feature. The primitive features that are extracted from the probability density function (pdf) of the pitch contour are used for classification. A feature vector of 36 dimension is obtained by extracting some parameters from the pdf. Since non-sequential features are extracted from the signal, artificial neural network (ANN) is used as a classifier. The database used for validating the system consists of 162 songs from 12 ragas. The average classification accuracy is found to be 89.5%. © Springer India 2015.Item Identification of allied raagas in Carnatic music(Institute of Electrical and Electronics Engineers Inc., 2015) Upadhyaya, P.; Suma, S.M.; Koolagudi, S.G.In this work, an effort has been made to differentiate the allied raagas in Carnatic music. Allied raagas are the raagas that are composed using same set of notes. The features derived from the pitch sequence are used for differentiating these raagas. The coefficients of legendre polynomials, used to fit the pitch contours of the song clips are used for identifying raagas. Obtained features are validated using different classifiers such as Neural networks, Naive Bayes, Multi class classifier, Bagging and Random forest. The proposed system is tested on 4 sets of allied raagas. Naive Bayes classifier gives an average accuracy of 86.67% for allied set of Todi-Dhanyasi and Multi class classifier gives an average accuracy of 86.67% for allied set of Kharaharapriya-Anandabhairavi-Reethigoula. In general, Neural network classifier performance is found to be better than other classifiers. © 2015 IEEE.Item Identifying gamakas in Carnatic music(Institute of Electrical and Electronics Engineers Inc., 2015) Vyas, H.M.; Suma, S.M.; Koolagudi, S.G.; Guruprasad, K.R.In this work, an effort has been made to identify the gamakas present in a given piece of Carnatic music clip. Gamakas are the beautification elements used to improve the melody. The identification of gamaka is very important stage in note transcription. In the proposed method, features that correspond to melodic variations such as pitch and energy are used for characterizing the gamakas. The input pitch contour is modelled using Hidden Markov Model with 3 states, namely Attack, Sustain and Decay. These states correspond to ups and downs in the melody of the music. The system is validated using a comprehensive data set consisting 160 songs from 8 different ragas. The average accuracy of 75.86% is achieved using this method. © 2015 IEEE.Item Rhythm and timbre analysis for carnatic music processing(Springer Science and Business Media Deutschland GmbH info@springer-sbm.com, 2016) Heshi, R.; Suma, S.M.; Koolagudi, S.G.; Bhandari, S.; Sreenivasa Rao, K.S.In this work, an effort has been made to analyze rhythm and timbre related features to identify raga and tala from a piece of Carnatic music. Raga and Tala classification is performed using both rhythm and timbre features. Rhythm patterns and rhythm histogram are used as rhythm features. Zero crossing rate (ZCR), centroid, spectral roll-off, flux, entropy are used as timbre features. Music clips contain both instrumental and vocals. To find similarity between the feature vectors T-Test is used as a similarity measure. Further, classification is done using Gaussian Mixture Models (GMM). The results shows that the rhythm patterns are able to distinguish different ragas and talas with an average accuracy of 89.98 and 86.67 % respectively. © Springer India 2016.Item Note Transcription from Carnatic Music(Springer, 2020) Suma, S.M.; Koolagudi, S.G.; Ramteke, P.B.; Sreenivasa Rao, K.S.In this work, an effort has been made to identify note sequence of different ragas of Carnatic Music. The proposed heuristic method makes use of standard just-intonation frequency ratios between notes for basic transcription of music piece into written sequence of notes. The notes present in a given piece of music are obtained using pitch histograms. The normalized pitch contour of the music piece is segmented based on detection of the note boundaries. These segments are labeled using note information already available. Without prior knowledge of raga, 30 out of 64 sequences are identified accurately and additional 18 sequences are identified with one note error. With the prior raga knowledge 76.56% accuracy is observed in note sequence identification. © 2020, Springer Nature Singapore Pte Ltd.
