Conference Papers
Permanent URI for this collectionhttps://idr.nitk.ac.in/handle/123456789/28506
Browse
2 results
Search Results
Item SanS: Classification of Sanskrit Mantras using Speech Processing(Association for Computing Machinery, 2024) Keerthan Kumar, T.G.; Udaya, S.; Koolagudi, S.G.Chandas is a classification metric used in Sanskrit Mantras. They are essentially meant to maintain a form of rhythm for each and every mantra, which gives the hymns their distinctive chanting pattern. Sanskrit hymns (mantras) have many different classifications. Often, before chanting, they are invoked as such in the order of Rishi (author of the hymn), Chanda (rhythm), Devata (god to which invoked), and Viniyoga (use of such hymn). One such example is the Gayatri hymn, which is from Gayatri Chanda. Other examples are verses of the Bhagavad Gita, which are entirely in Anushtup Chanda. Chandas is an essential component of Sanskrit Mantras and forms an integral part of it. Knowledge of Chandas is essential for the study of Sanskrit Poetry, and without knowing what a Chanda is, one cannot analyze Sanskrit hymns. Essentially, Chandas are the meters used to keep track of the mantras. Chandas formulate the rhythm of the mantra and the way that it should be chanted. Based on the number of syllables, there are seven different Chandas, and mantras are usually classified into one of these seven. Identification of Chandas is usually specified in the beginning before chanting; however, in cases where the Chandas are not specified, one may have to do it manually, which may be cumbersome, especially for Chandas with many syllables. In this work, we propose a novel approach called Classification of Sanskrit Mantras using Speech Processing Technology (SanS) to determine the Chanda of a mantra, given the audio file of a Sanskrit mantra and using a Wavenet architecture. The proposed SanS predicts the type of Chanda by counting the number of syllables, giving 81.57% accuracy compared to other works. © 2024 Copyright held by the owner/author(s).Item Binarization in DeepFake Audio Detection: A Comparative Study and Performance Analysis(Institute of Electrical and Electronics Engineers Inc., 2025) Gowhar, S.; Pandey, A.; Rudra, B.DeepFake audio, generated through advanced AI techniques, poses significant risks such as fraud, misinformation, and identity theft. As the quality of synthetic audio improves, detecting such fakes has become increasingly challenging. Traditional detection methods struggle to keep pace as AI-generated voices replicate speech patterns, tone, and pitch convincingly. While computationally intensive large-scale models can help detect DeepFakes generated by AI, their resource requirements make them impractical for deployment on mobile devices as well as on resource-constrained devices. This paper proposes a lightweight yet effective approach using binarized neural networks (BNNs) and further enhancements using additional dense layers and stacked modeling to overcome these challenges. We conduct a comprehensive performance analysis of the network and compare it with various machine learning and neural network methods to evaluate the tradeoff between detection accuracy and computational efficiency as an effect of binarization and precision loss in feature embeddings. © 2025 IEEE.
