Faculty Publications

Permanent URI for this communityhttps://idr.nitk.ac.in/handle/123456789/18736

Publications by NITK Faculty

Browse

Search Results

Now showing 1 - 10 of 13
  • Item
    A Lightweight Convolutional Neural Network Model for Tuberculosis Bacilli Detection From Microscopic Sputum Smear Images
    (wiley, 2021) Panicker, R.O.; Pawan, S.J.; Rajan, J.; Sabu, M.K.
    This chapter describes a lightweight convolutional neural network model that automatically detects Tuberculosis (TB) bacilli from sputum smear microscopic images. According to WHO, about onefourth of the population in the universe is infected with TB, and every day five thousand people are killed due to TB disease. There are well-known recommended diagnostics are available for TB detection, among them sputum smear microscopic examination is a primary and most efficient recommended method for most of the developing and moderately developed countries. However, this manual detection method is highly error-prone and time-consuming. In this chapter, we proposed a lightweight CNN model for classifying Tuberculosis bacilli from non-bacilli objects. We adopted a Convolutional Neural Network (CNN) architecture with a skip connection of variable lengths that can identify TB bacilli from sputum smear microscopic images. The performance of the proposed model in terms of accuracy is close to the state-of-the-art. However, the number of parameters in the proposed model is significantly less than other recently proposed models. © 2021 Scrivener Publishing LLC.
  • Item
    Automatic detection and localization of Focal Cortical Dysplasia lesions in MRI using fully convolutional neural network
    (Elsevier Ltd, 2019) Bijay Dev, K.M.; Pawan, P.S.; Niyas, S.; Vinayagamani, S.; Kesavadas, C.; Rajan, J.
    Focal cortical dysplasia (FCD) is the leading cause of drug-resistant epilepsy in both children and adults. At present, the only therapeutic approach in patients with drug-resistant epilepsy is surgery. Hence, the quantification of FCD via non-invasive imaging techniques helps physicians to decide on surgical interventions. The properties like non-invasiveness and capability to produce high-resolution images makes magnetic resonance imaging an ideal tool for detecting the FCD to an extent. The FCD lesions vary in size, shape, and location for different patients and make the manual detection time consuming and sensitive to the experience of the observer. Automatic segmentation of FCD lesions is challenging due to the difference in signal strength in images acquired with different machines, noise, and other kinds of distortions such as motion artifacts. Most of the methods proposed in the literature use conventional machine learning and image processing techniques in which their accuracy relies on the trained features. Hence, feature extraction should be done more precisely which requires human expertise. The ability to learn the appropriate features/representations from the training data without any human interventions makes the convolutional neural network (CNN) the suitable method for addressing these drawbacks. As far as we are aware, this work is the first one to use a CNN based model to solve the aforementioned problem using only MRI FLAIR images. We customized the popular U-Net architecture and trained the proposed model from scratch (using MRI images acquired with 1.5T and 3T scanners). FCD detection rate (recall) of the proposed model is 82.5 (33/40 patients detected correctly). © 2019
  • Item
    Stack generalized deep ensemble learning for retinal layer segmentation in Optical Coherence Tomography images
    (Elsevier Sp. z o.o., 2020) Anoop, B.N.; Pavan, R.; Girish, G.N.; Kothari, A.R.; Rajan, J.
    Segmentation of retinal layers is a vital and important step in computerized processing and the study of retinal Optical Coherence Tomography (OCT) images. However, automatic segmentation of retinal layers is challenging due to the presence of noise, widely varying reflectivity of image components, variations in morphology and alignment of layers in the presence of retinal diseases. In this paper, we propose a Fully Convolutional Network (FCN) termed as DelNet based on a deep ensemble learning approach to selectively segment retinal layers from OCT scans. The proposed model is tested on a publicly available DUKE DME dataset. Comparative analysis with other state-of-the-art methods on a benchmark dataset shows that the performance of DelNet is superior to other methods. © 2020 Nalecz Institute of Biocybernetics and Biomedical Engineering of the Polish Academy of Sciences
  • Item
    A cascaded convolutional neural network architecture for despeckling OCT images
    (Elsevier Ltd, 2021) Anoop, B.N.; Kalmady, K.S.; Udathu, A.; Siddharth, V.; Girish, G.N.; Kothari, A.R.; Rajan, J.
    Optical Coherence Tomography (OCT) is an imaging technique widely used for medical imaging. Noise in an OCT image generally degrades its quality, thereby obscuring clinical features and making the automated segmentation task suboptimal. Obtaining higher quality images requires sophisticated equipment and technology, available only in selected research settings, and is expensive to acquire. Developing effective denoising methods to improve the quality of the images acquired on systems currently in use has potential for vastly improving image quality and automated quantitative analysis. Noise characteristics in images acquired from machines of different makes and models may vary. Our experiments show that any single state-of-the-art method for noise reduction fails to perform equally well on images from various sources. Therefore, detailed analysis is required to determine the exact noise type in images acquired using different OCT machines. In this work we studied noise characteristics in the publicly available DUKE and OPTIMA datasets to build a more efficient model for noise reduction. These datasets have OCT images acquired using machines of different manufacturers. We further propose a patch-wise training methodology to build a system to effectively denoise OCT images. We have performed an extensive range of experiments to show that the proposed method performs superior to other state-of-the-art-methods. © 2021 Elsevier Ltd
  • Item
    Multi-Res-Attention UNet: A CNN Model for the Segmentation of Focal Cortical Dysplasia Lesions from Magnetic Resonance Images
    (Institute of Electrical and Electronics Engineers Inc., 2021) Thomas, E.; Pawan, S.J.; Kumar, S.; Horo, A.; Niyas, S.; Vinayagamani, S.; Kesavadas, C.; Rajan, J.
    In this work, we have focused on the segmentation of Focal Cortical Dysplasia (FCD) regions from MRI images. FCD is a congenital malformation of brain development that is considered as the most common causative of intractable epilepsy in adults and children. To our knowledge, the latest work concerning the automatic segmentation of FCD was proposed using a fully convolutional neural network (FCN) model based on UNet. While there is no doubt that the model outperformed conventional image processing techniques by a considerable margin, it suffers from several pitfalls. First, it does not account for the large semantic gap of feature maps passed from the encoder to the decoder layer through the long skip connections. Second, it fails to leverage the salient features that represent complex FCD lesions and suppress most of the irrelevant features in the input sample. We propose Multi-Res-Attention UNet; a novel hybrid skip connection-based FCN architecture that addresses these drawbacks. Moreover, we have trained it from scratch for the detection of FCD from 3 T MRI 3D FLAIR images and conducted 5-fold cross-validation to evaluate the model. FCD detection rate (Recall) of 92% was achieved for patient wise analysis. © 2013 IEEE.
  • Item
    Capsule Network–based architectures for the segmentation of sub-retinal serous fluid in optical coherence tomography images of central serous chorioretinopathy
    (Springer Science and Business Media Deutschland GmbH, 2021) Pawan, S.J.; Sankar, R.; Jain, A.; Jain, M.; Darshan, D.V.; Anoop, B.N.; Kothari, A.R.; Venkatesan, M.; Rajan, J.
    Central serous chorioretinopathy (CSCR) is a chorioretinal disorder of the eye characterized by serous detachment of the neurosensory retina at the posterior pole of the eye. CSCR results from the accumulation of subretinal fluid (SRF) due to idiopathic defects at the level of the retinal pigment epithelial (RPE) that allows serous fluid from the choriocapillaris to diffuse into the subretinal space between RPE and neurosensory retinal layers. This condition is presently investigated by clinicians using invasive angiography or non-invasive optical coherence tomography (OCT) imaging. OCT images provide a representation of the fluid underlying the retina, and in the absence of automated segmentation tools, currently only a qualitative assessment of the same is used to follow the progression of the disease. Automated segmentation of the SRF can prove to be extremely useful for the assessment of progression and for the timely management of CSCR. In this paper, we adopt an existing architecture called SegCaps, which is based on the recently introduced Capsule Networks concept, for the segmentation of SRF from CSCR OCT images. Furthermore, we propose an enhancement to SegCaps, which we have termed as DRIP-Caps, that utilizes the concepts of Dilation, Residual Connections, Inception Blocks, and Capsule Pooling to address the defined problem. The proposed model outperforms the benchmark UNet architecture while reducing the number of trainable parameters by 54.21%. Moreover, it reduces the computation complexity of SegCaps by reducing the number of trainable parameters by 37.85%, with competitive performance. The experiments demonstrate the generalizability of the proposed model, as evidenced by its remarkable performance even with a limited number of training samples. [Figure not available: see fulltext.]. © 2021, International Federation for Medical and Biological Engineering.
  • Item
    Segmentation of focal cortical dysplasia lesions from magnetic resonance images using 3D convolutional neural networks
    (Elsevier Ltd, 2021) Niyas, S.; Chethana Vaisali, S.; Show, I.; Chandrika, T.G.; Vinayagamani, S.; Kesavadas, C.; Rajan, J.
    Computer-aided diagnosis using advanced Artific ial Intelligence (AI) techniques has become much popular over the last few years. This work automates the segmentation of Focal Cortical Dysplasia (FCD) lesions from three-dimensional (3D) Magnetic Resonance (MR) images. FCD is a type of neuronal malformation in the brain cortex and is the leading cause of intractable epilepsy, irrespective of gender or age differences. Since the neuron related abnormalities are usually resistant to drug therapy, surgical resection has been the main treatment approach for patients with intractable epilepsy. Automating the identification and segmentation of FCD is useful for neuroradiologists in pre-surgical evaluations. Convolutional Neural Networks (CNNs) have the ability to learn appropriate features from the training data without any human intervention. But, most of the state-of-the-art FCD segmentation approaches use two-dimensional (2D) CNN models despite the availability of 3D Magnetic resonance imaging (MRI) volumes, and hence fail to leverage the inter-slice information present in the MRI volumes. The major hurdles in considering a 3D CNN model are the need for a large 3D dataset, big memory, and high computation cost. A deep 3D CNN segmentation model, which can extract inter-slice information and overcomes the drawbacks of conventional 3D CNN methods to an extent, is proposed in this paper. The model uses a 3D version of U-Net with residual blocks that works on shallow depth 3D sub-volumes generated from MRI volumes. The proposed method shows superior performance over the state-of-the-art FCD segmentation methods in both qualitative and quantitative analysis. © 2021 Elsevier Ltd
  • Item
    Crossover based technique for data augmentation
    (Elsevier Ireland Ltd, 2022) Raj, R.; Mathew, J.; Kannath, S.K.; Rajan, J.
    Background and Objective: Medical image classification problems are frequently constrained by the availability of datasets. “Data augmentation” has come as a data enhancement and data enrichment solution to the challenge of limited data. Traditionally data augmentation techniques are based on linear and label preserving transformations; however, recent works have demonstrated that even non-linear, non-label preserving techniques can be unexpectedly effective. This paper proposes a non-linear data augmentation technique for the medical domain and explores its results. Methods: This paper introduces “Crossover technique”, a new data augmentation technique for Convolutional Neural Networks in Medical Image Classification problems. Our technique synthesizes a pair of samples by applying two-point crossover on the already available training dataset. By this technique, we create N new samples from N training samples. The proposed crossover based data augmentation technique, although non-label preserving, has performed significantly better in terms of increased accuracy and reduced loss for all the tested datasets over varied architectures. Results: The proposed method was tested on three publicly available medical datasets with various network architectures. For the mini-MIAS database of mammograms, our method improved the accuracy by 1.47%, achieving 80.15% using VGG-16 architecture. Our method works fine for both gray-scale as well as RGB images, as on the PH2 database for Skin Cancer, it improved the accuracy by 3.57%, achieving 85.71% using VGG-19 architecture. In addition, our technique improved accuracy on the brain tumor dataset by 0.40%, achieving 97.97% using VGG-16 architecture. Conclusion: The proposed novel crossover technique for training the Convolutional Neural Network (CNN) is painless to implement by applying two-point crossover on two images to form new images. The method would go a long way in tackling the challenges of limited datasets and problems of class imbalances in medical image analysis. Our code is available at https://github.com/rishiraj-cs/Crossover-augmentation © 2022
  • Item
    Stroke classification from computed tomography scans using 3D convolutional neural network
    (Elsevier Ltd, 2022) Neethi, A.S.; Niyas, S.; Kannath, S.K.; Mathew, J.; Anzar, A.M.; Rajan, J.
    Stroke is a cerebrovascular condition with a significant morbidity and mortality rate and causes physical disabilities for survivors. Once the symptoms are identified, it requires a time-critical diagnosis with the help of the most commonly available imaging techniques. Computed tomography (CT) scans are used worldwide for preliminary stroke diagnosis. It demands the expertise and experience of a radiologist to identify the stroke type, which is critical for initiating the treatment. This work attempts to gather those domain skills and build a model from CT scans to diagnose stroke. The non-contrast computed tomography (NCCT) scan of the brain comprises volumetric images or a 3D stack of image slices. So, a model that aims to solve the problem by targeting a 2D slice may fail to address the volumetric nature. We propose a 3D-based fully convolutional classification model to identify stroke cases from CT images that take into account the contextual longitudinal composition of volumetric data. We formulate a custom pre-processing module to enhance the scans and aid in improving the classification performance. Some of the significant challenges faced by 3D CNN are the less number of training samples, and the number of scans is mostly biased in favor of normal patients. In this work, the limitation of insufficient training volume and class imbalanced data have been rectified with the help of a strided slicing approach. A block-wise design was used to formulate the proposed network, with the initial part focusing on adjusting the dimensionality, at the same time retaining the features. Later on, the accumulated feature maps were effectively learned utilizing bundled convolutions and skip connections. The results of the proposed method were compared against 3D CNN stroke classification models on NCCT, various 3D CNN architectures on other brain imaging modalities, and 3D extensions of some of the classical CNN architectures. The proposed method achieved an improvement of 14.28% in the F1-score over the state-of-the-art 3D CNN stroke classification model. © 2022 Elsevier Ltd
  • Item
    A novel deep classifier framework for automated molecular subtyping of breast carcinoma using immunohistochemistry image analysis
    (Elsevier Ltd, 2022) Mathew, T.; Niyas, S.; Johnpaul, C.I.; Kini, J.; Rajan, J.
    Breast carcinoma has various subtypes based on the genetic factors involved in the pathogenesis of the malignancy. Identifying the exact subtype and providing targeted treatment to the patient can improve the survival chances. Molecular subtyping through immunohistochemistry analysis is a pathology procedure to determine the subtype of breast cancer. The existing manual procedure is tedious and involves assessing the status of the four vital molecular biomarkers present in the tumor tissues. In this paper, a deep learning-based framework for automated molecular subtyping of breast cancer is proposed. Digital slide images of the four biomarkers are separately processed by the proposed framework. In the preprocessing stage, the non-informative background regions from the images are separated. The patches extracted from the foreground regions are classified into target classes using convolutional neural network models trained for this purpose. Classification results are post-processed to predict the status of all the four biomarkers. The predictions for the individual biomarkers are finally consolidated as per clinical guidelines to determine the subtype of the cancer. The proposed system is evaluated for the performance of individual biomarker status prediction and patient-level subtype classification.For patient-level evaluation of biomarkers ER, PR, K67, and HER2, the proposed method gives F1 Scores 1.00, 1.00, 0.90, and 0.94 respectively, whereas for molecular subtyping an F1 score of 0.89 is obtained. In both these aspects, the proposed framework has given significant results that show the effectiveness of our approach. © 2022 Elsevier Ltd