Faculty Publications

Permanent URI for this communityhttps://idr.nitk.ac.in/handle/123456789/18736

Publications by NITK Faculty

Browse

Search Results

Now showing 1 - 5 of 5
  • Item
    COVID-19: Automatic detection from X-ray images by utilizing deep learning methods
    (Elsevier Ltd, 2021) Nigam, B.; Nigam, A.; Jain, R.; Dodia, S.; Arora, N.; Annappa, B.
    In recent months, a novel virus named Coronavirus has emerged to become a pandemic. The virus is spreading not only humans, but it is also affecting animals. First ever case of Coronavirus was registered in city of Wuhan, Hubei province of China on 31st of December in 2019. Coronavirus infected patients display very similar symptoms like pneumonia, and it attacks the respiratory organs of the body, causing difficulty in breathing. The disease is diagnosed using a Real-Time Reverse Transcriptase Polymerase Chain reaction (RT-PCR) kit and requires time in the laboratory to confirm the presence of the virus. Due to insufficient availability of the kits, the suspected patients cannot be treated in time, which in turn increases the chance of spreading the disease. To overcome this solution, radiologists observed the changes appearing in the radiological images such as X-ray and CT scans. Using deep learning algorithms, the suspected patients’ X-ray or Computed Tomography (CT) scan can differentiate between the healthy person and the patient affected by Coronavirus. In this paper, popular deep learning architectures are used to develop a Coronavirus diagnostic systems. The architectures used in this paper are VGG16, DenseNet121, Xception, NASNet, and EfficientNet. Multiclass classification is performed in this paper. The classes considered are COVID-19 positive patients, normal patients, and other class. In other class, chest X-ray images of pneumonia, influenza, and other illnesses related to the chest region are included. The accuracies obtained for VGG16, DenseNet121, Xception, NASNet, and EfficientNet are 79.01%, 89.96%, 88.03%, 85.03% and 93.48% respectively. The need for deep learning with radiologic images is necessary for this critical condition as this will provide a second opinion to the radiologists fast and accurately. These deep learning Coronavirus detection systems can also be useful in the regions where expert physicians and well-equipped clinics are not easily accessible. © 2021 Elsevier Ltd
  • Item
    Stroke classification from computed tomography scans using 3D convolutional neural network
    (Elsevier Ltd, 2022) Neethi, A.S.; Niyas, S.; Kannath, S.K.; Mathew, J.; Anzar, A.M.; Rajan, J.
    Stroke is a cerebrovascular condition with a significant morbidity and mortality rate and causes physical disabilities for survivors. Once the symptoms are identified, it requires a time-critical diagnosis with the help of the most commonly available imaging techniques. Computed tomography (CT) scans are used worldwide for preliminary stroke diagnosis. It demands the expertise and experience of a radiologist to identify the stroke type, which is critical for initiating the treatment. This work attempts to gather those domain skills and build a model from CT scans to diagnose stroke. The non-contrast computed tomography (NCCT) scan of the brain comprises volumetric images or a 3D stack of image slices. So, a model that aims to solve the problem by targeting a 2D slice may fail to address the volumetric nature. We propose a 3D-based fully convolutional classification model to identify stroke cases from CT images that take into account the contextual longitudinal composition of volumetric data. We formulate a custom pre-processing module to enhance the scans and aid in improving the classification performance. Some of the significant challenges faced by 3D CNN are the less number of training samples, and the number of scans is mostly biased in favor of normal patients. In this work, the limitation of insufficient training volume and class imbalanced data have been rectified with the help of a strided slicing approach. A block-wise design was used to formulate the proposed network, with the initial part focusing on adjusting the dimensionality, at the same time retaining the features. Later on, the accumulated feature maps were effectively learned utilizing bundled convolutions and skip connections. The results of the proposed method were compared against 3D CNN stroke classification models on NCCT, various 3D CNN architectures on other brain imaging modalities, and 3D extensions of some of the classical CNN architectures. The proposed method achieved an improvement of 14.28% in the F1-score over the state-of-the-art 3D CNN stroke classification model. © 2022 Elsevier Ltd
  • Item
    StrokeViT with AutoML for brain stroke classification
    (Elsevier Ltd, 2023) Raj, R.; Mathew, J.; Kannath, S.K.; Rajan, J.
    Stroke, categorized under cardiovascular and circulatory diseases, is considered the second foremost cause of death worldwide, causing approximately 11% of deaths annually. Stroke diagnosis using a Computed Tomography (CT) scan is considered ideal for identifying whether the stroke is hemorrhagic or ischemic. However, most methods for stroke classification are based on a single slice-level prediction mechanism, meaning that the most imperative CT slice has to be manually selected by the radiologist from the original CT volume. This paper proposes an integration of Convolutional Neural Network (CNN), Vision Transformers (ViT), and AutoML to obtain slice-level predictions as well as patient-wise prediction results. While the CNN with inductive bias captures local features, the transformer captures long-range dependencies between sequences. This collaborative local-global feature extractor improves upon the slice-wise predictions of the CT volume. We propose stroke-specific feature extraction from each slice-wise prediction to obtain the patient-wise prediction using AutoML. While the slice-wise predictions helps the radiologist to verify close and corner cases, the patient-wise predictions makes the outcome clinically relevant and closer to real-world scenario. The proposed architecture has achieved an accuracy of 87% for single slice-level prediction and an accuracy of 92% for patient-wise prediction. For comparative analysis of slice-level predictions, standalone architectures of VGG-16, VGG-19, ResNet50, and ViT were considered. The proposed architecture has outperformed the standalone architectures by 9% in terms of accuracy. For patient-wise predictions, AutoML considers 13 different ML algorithms, of which 3 achieve an accuracy of more than 90%. The proposed architecture helps in reducing the manual effort by the radiologist to manually select the most imperative CT from the original CT volume and shows improvement over other standalone architectures for classification tasks. The proposed architecture can be generalized for volumetric scans aiding in the patient diagnosis of head and neck, lungs, diseases of hepatobiliary tract, genitourinary diseases, women's imaging including breast cancer and various musculoskeletal diseases. Code for proposed stroke-specific feature extraction with the pre-trained weights of the trained model is available at: https://github.com/rishiraj-cs/StrokeViT_With_AutoML. © 2022 Elsevier Ltd
  • Item
    KAC SegNet: A Novel Kernel-Based Active Contour Method for Lung Nodule Segmentation and Classification Using Dense AlexNet Framework
    (World Scientific, 2024) Dodia, S.; Annappa, B.; Mahesh, P.A.
    Lung cancer is known to be one of the leading causes of death worldwide. There is a chance of increasing the survival rate of the patients if detected at an early stage. Computed Tomography (CT) scans are prominently used to detect and classify lung cancer nodules/tumors in the thoracic region. There is a need to develop an efficient and reliable computer-aided diagnosis model to detect lung cancer nodules accurately from CT scans. This work proposes a novel kernel-based active-contour (KAC) SegNet deep learning model to perform lung cancer nodule detection from CT scans. The active contour uses a snake method to detect internal and external boundaries of the curves, which is used to extract the Region Of Interest (ROI) from the CT scan. From the extracted ROI, the nodules are further classified into benign and malignant using a Dense AlexNet deep learning model. The key contributions of this work are the fusion of an edge detection method with a deep learning segmentation method which provides enhanced lung nodule segmentation performance, and an ensemble of state-of-the-art deep learning classifiers, which encashes the advantages of both DenseNet and AlexNet to learn better discriminative information from the detected lung nodules. The experimental outcome shows that the proposed segmentation approach achieves a Dice Score Coefficient of 97.8% and an Intersection-over-Union of 92.96%. The classification performance resulted in an accuracy of 95.65%, a False Positive Rate, and False Negative Rate values of 0.0572 and 0.0289. The proposed model is robust compared to the existing state-of-the-art methods. © 2024 World Scientific Publishing Company.
  • Item
    CAEB7-UNet: An Attention-Based Deep Learning Framework for Automated Segmentation of C-Spine Vertebrae in CT Images
    (Institute of Electrical and Electronics Engineers Inc., 2025) Pandey, A.K.; Senapati, K.; Pateel, G.P.
    Accurate segmentation of vertebrae in computed tomography (CT) images possess serious challenges due to the irregular vertebral boundaries, low contrast and brightness, and noise in CT scans. This study presents a novel channel attention-based EfficientNetB7-UNet (CAEB7-UNet) method to address this complex task effectively. The proposed model introduces an upgraded ReLU-based channel attention module (CAM) in the skip connection which restrains the nonessential attributes by suppressing them and accentuates the relevant features by emphasizing them to boost the overall segmentation performance. In this work, an improved EfficientNetB7 is employed as the encoder for feature extraction, the fusion of local and global features is enhanced through the upgraded CAM in skip connection, and the up-sampling is performed in the decoder. Further, the model is optimized by incorporating hyperparameter optimization, specifically, hybrid learning rate scheduler strategies, along with the AdamW optimizer and custom data augmentation. A total of 34,782 CT images obtained from the RSNA-2022 cervical spine fracture detection challenge is utilized in this study. The proposed model achieves outstanding performance, yielding a dice score index (DSI) of 96.14% and mean intersection over union (mIoU) of 91.46%. Moreover, a comparative performance analysis of CAEB7-UNet with two state-of-the-art models is carried out on the same dataset. Our approach outperforms both the models, with the best one by 8.1%, 6.73%, 12.7%, and 11.98% in terms of DSI, mIoU, precision, and F1-score respectively. Additionally, it requires merely 0.38 seconds to generate the segmentation mask of a single slice of a CT scan. © 2013 IEEE.