Conference Papers

Permanent URI for this collectionhttps://idr.nitk.ac.in/handle/123456789/28506

Browse

Search Results

Now showing 1 - 10 of 11
  • Item
    Dynamic 3D graph visualizations in Julia
    (The Society for Modeling and Simulation International www.scs.org, 2016) Anilkumar, A.; Mathew, K.T.; Mulimani, M.; Koolagudi, S.G.; Jamadagni, C.
    A major problem with graph visualization libraries and packages is the lack of interactivity and 3D visualization. This makes understanding and analyzing complex graphs and topologies difficult. Existing packages and tools which do provide similar functionality are difficult to use, install, integrate and have many dependencies. This paper discusses NetworkViz.jl, a Julia package which addresses the issues of existing graph visualization platforms while ensuring simplicity, efficiency, a diverse set of features and easy integration with other packages. This package supports two-And three- dimensional visualizations and uses a force-directed graph drawing approach to generate aesthetically pleasing and easy-To-use graphs. The library was built entirely in Julia due to its good documentation, large open source community and in order to fully utilize (he inherent advantages provided by the language. As graph visualizations are important for analyzing complex networks, testing routing algorithms, as teaching AIDS, etc., we believe that NetworkViz.jl will be of integral use in the fields of research and education. © 2016 Society for Modeling & Simulation International (SCS).
  • Item
    Image processing approach to diagnose eye diseases
    (Springer Verlag service@springer.de, 2017) Prashasthi, P.; Shravya, K.S.; Deepak, A.; Mulimani, M.; Shashidhar, K.G.
    Image processing and machine learning techniques are used for automatic detection of abnormalities in eye. The proposed methodology requires a clear photograph of eye (not necessarily a fundoscopic image) from which the chromatic and spatial property of the sclera and iris is extracted. These features are used in the diagnosis of various diseases considered. The changes in the colour of iris is a symptom for corneal infections and cataract, the spatial distribution of different colours distinguishes diseases like subconjunctival haemorrhage and conjunctivitis, and the spatial arrangement of iris and sclera is an indicator of palsy. We used various classifiers of which adaboost classifier which was found to give a substantially high accuracy i.e., about 95% accuracy when compared to others (k-NN and naive-Bayes). To enumerate the accuracy of the method proposed, we used 150 samples in which 23% were used for testing and 77% were used for training. © Springer International Publishing AG 2017.
  • Item
    Currency recognition system using image processing
    (Institute of Electrical and Electronics Engineers Inc., 2017) Abburu, V.; Gupta, S.; Rimitha, S.R.; Mulimani, M.; Koolagudi, S.G.
    In this paper, we propose a system for automated currency recognition using image processing techniques. The proposed method can be used for recognizing both the country or origin as well as the denomination or value of a given banknote. Only paper currencies have been considered. This method works by first identifying the country of origin using certain predefined areas of interest, and then extracting the denomination value using characteristics such as size, color, or text on the note, depending on how much the notes within the same country differ. We have considered 20 of the most traded currencies, as well as their denominations. Our system is able to accurately and quickly identify test notes. © 2017 IEEE.
  • Item
    Acoustic event classification using graph signals
    (Institute of Electrical and Electronics Engineers Inc., 2017) Mulimani, M.; Jahnavi, U.P.; Koolagudi, S.G.
    In this paper, a graph signal is generated from spectrogram and features are investigated from graph signal for Acoustic Event Classification (AEC). Different acoustic events are selected from Sound Scene Database of Real Word Computing Partnership (RWCP) group. Three different noises are selected from NOISEX'92 database and added to test samples at different noise conditions separately. The recognition performance of acoustic events using proposed features and Mel-frequency cepstral coefficients (MFCCs) with clean and noisy test samples are compared. The proposed features show significantly improved recognition accuracy over MFCCs in noisy conditions. © 2017 IEEE.
  • Item
    Gender Detection using Handwritten Signatures
    (Institute of Electrical and Electronics Engineers Inc., 2018) Mohit Reddy, J.; Guru Pradeep Reddy, T.; Mishra, S.; Mulimani, M.; Koolagudi, S.G.
    In this paper, a method is proposed which uses both Image Processing and Machine Learning techniques which detects the gender of a person using handwritten signature. A photograph of a handwritten signature is given as input to the model which then extracts different features like pen pressure, slant angle, count external and internal contours etc. The features extracted from multiple images in the dataset are used to train the model, which then predicts the output of a new input given to it. Our objective is to collect unbiased datasets from a set of people and feed those signatures to the model, carrying out the statistical analysis and calculating the accuracy of the algorithm after every signature classification. We have used Adaboost classifier which gave a cross-validation accuracy of 73.2% compared to other classifiers like Gradient Boosting Classifier, Random Forest Trees and Multi-Layer Perceptron which gave 73.2%, 63.2% and 59.6% accuracies respectively. Copy Right © INDIACom-2018.
  • Item
    Robust acoustic event classification using bag-of-visual-words
    (International Speech Communication Association publication@isca-speech.org 4 Rue des Fauvettes - Lous Tourils Baixas 66390, 2018) Mulimani, M.; Koolagudi, S.G.
    This paper presents a novel Bag-of-Visual-Words (BoVW) approach, to represent the grayscale spectrograms of acoustic events. Such, BoVW representations are referred as histograms of visual features, used for Acoustic Event Classification (AEC). Further, Chi-square distance between histograms of visual features evaluated, which generates kernel to Support Vector Machines (Chi-square SVM) classifier. Evaluation of the proposed histograms of visual features together with Chi-square SVM classifier is conducted on different categories of acoustic events from UPC-TALP corpora in clean and different noise conditions. Results show that proposed approach is more robust to noise and achieves improved recognition accuracy compared to other methods. © 2018 International Speech Communication Association. All rights reserved.
  • Item
    Acoustic Event Classification Using Spectrogram Features
    (Institute of Electrical and Electronics Engineers Inc., 2018) Mulimani, M.; Koolagudi, S.G.
    This paper investigates a new feature extraction method to extract different features from the spectrogram of an audio signal for Acoustic Event Classification (AEC). A new set of features is formulated and extracted from local spectrogram regions named blocks. The average recognition performance of proposed spectrogram based features and Mel-frequency cepstral coefficients (MFCCs) with their deltas and accelerations on Support Vector Machines (SVM) is compared. In this work, different categories of acoustic events are considered from the Freiburg-106 dataset. Proposed features show significantly improved performance over conventional Mel-frequency cepstral coefficients (MFCCs) for Acoustic Event Classification. © 2018 IEEE.
  • Item
    Sobriety Testing Based on Thermal Infrared Images Using Convolutional Neural Networks
    (Institute of Electrical and Electronics Engineers Inc., 2018) Kamath, A.K.; Karthik, A.T.; Monis, L.; Mulimani, M.; Koolagudi, S.G.
    This paper proposes a method to test the sobriety of an individual using infrared images of the persons eyes, face, hand, and facial profile. The database we used consisted of images of forty different individuals. The process is broken down into two main stages. In the first stage, the data set was divided according to body part and each one was run through its own Convolutional Neural Network (CNN). We then tested the resulting network against a validation data set. The results obtained gave us an indication of which body parts were better suited for identifying signs of drunken state and sobriety. In the second stage, we took the weights of CNN giving best validation accuracy from the first stage. We then grouped the body parts according to the person they belong to. The body parts were fed together into a CNN using the weights obtained in the first stage. The result for each body part was passed to a simple back-propagation neural network (BPNN) to get final results. We tried to identify the most optimal configuration of neural networks for each stage of the process. The results we obtained showed that facial profile images tend to give very good indications of sobriety. The results also showed that combining the results of multiple body parts using a simple BPNN gives a higher accuracy than that of individual ones. © 2018 IEEE.
  • Item
    Locality-constrained linear coding based fused visual features for robust acoustic event classification
    (International Speech Communication Association, 2019) Mulimani, M.; Koolagudi, G.K.
    In this paper, a novel Fused Visual Features (FVFs) are proposed for Acoustic Event Classification (AEC) in the meeting room and office environments. The codes of Visual Features (VFs) are evaluated from row vectors and Scale Invariant Feature Transform (SIFT) vectors of the grayscale Gammatonegram of an acoustic event separately using Locality-constrained Linear Coding (LLC). Further, VFs from row vectors and SIFT vectors of the grayscale Gammatonegram are fused to get FVFs. Performance of the proposed FVFs is evaluated on acoustic events of publicly available UPC-TALP and DCASE datasets in clean and noisy conditions. Results show that proposed FVFs are robust to noise and achieve overall recognition accuracy of 96.40% and 90.45% on UPC-TALP and DCASE datasets, respectively. © 2019 ISCA
  • Item
    Polyphonic sound event detection using transposed convolutional recurrent neural network
    (Institute of Electrical and Electronics Engineers Inc., 2020) Chatterjee, C.C.; Mulimani, M.; Koolagudi, S.G.
    In this paper we propose a Transposed Convolutional Recurrent Neural Network (TCRNN) architecture for polyphonic sound event recognition. Transposed convolution layer, which caries out a regular convolution operation but reverts the spatial transformation and it is combined with a bidirectional Recurrent Neural Network (RNN) to get TCRNN. Instead of the traditional mel spectrogram features, the proposed methodology incorporates mel-IFgram (Instantaneous Frequency spectrogram) features. The performance of the proposed approach is evaluated on sound events of publicly available TUT-SED 2016 and Joint sound scene and polyphonic sound event recognition datasets. Results show that the proposed approach outperforms state-of-the-art methods. © 2020 Institute of Electrical and Electronics Engineers Inc.. All rights reserved.