Faculty Publications

Permanent URI for this communityhttps://idr.nitk.ac.in/handle/123456789/18736

Publications by NITK Faculty

Browse

Search Results

Now showing 1 - 7 of 7
  • Item
    NucleiSegNet: Robust deep learning architecture for the nuclei segmentation of liver cancer histopathology images
    (Elsevier Ltd, 2021) Lal, S.; Das, D.; Alabhya, K.; Kanfade, A.; Kumar, A.; Kini, J.R.
    The nuclei segmentation of hematoxylin and eosin (H&E) stained histopathology images is an important prerequisite in designing a computer-aided diagnostics (CAD) system for cancer diagnosis and prognosis. Automated nuclei segmentation methods enable the qualitative and quantitative analysis of tens of thousands of nuclei within H&E stained histopathology images. However, a major challenge during nuclei segmentation is the segmentation of variable sized, touching nuclei. To address this challenge, we present NucleiSegNet - a robust deep learning network architecture for the nuclei segmentation of H&E stained liver cancer histopathology images. Our proposed architecture includes three blocks: a robust residual block, a bottleneck block, and an attention decoder block. The robust residual block is a newly proposed block for the efficient extraction of high-level semantic maps. The attention decoder block uses a new attention mechanism for efficient object localization, and it improves the proposed architecture's performance by reducing false positives. When applied to nuclei segmentation tasks, the proposed deep-learning architecture yielded superior results compared to state-of-the-art nuclei segmentation methods. We applied our proposed deep learning architecture for nuclei segmentation to a set of H&E stained histopathology images from two datasets, and our comprehensive results show that our proposed architecture outperforms state-of-the-art methods. As part of this work, we also introduced a new liver dataset (KMC liver dataset) of H&E stained liver cancer histopathology image tiles, containing 80 images with annotated nuclei procured from Kasturba Medical College (KMC), Mangalore, Manipal Academy of Higher Education (MAHE), Manipal, Karnataka, India. The proposed model's source code is available at https://github.com/shyamfec/NucleiSegNet. © 2020 Elsevier Ltd
  • Item
    Efficient deep learning architecture with dimension-wise pyramid pooling for nuclei segmentation of histopathology images
    (Elsevier Ltd, 2021) Aatresh, A.A.; Yatgiri, R.P.; Chanchal, A.K.; Kumar, A.; Ravi, A.; Das, D.; Raghavendra, B.S.; Lal, S.; Kini, J.
    Image segmentation remains to be one of the most vital tasks in the area of computer vision and more so in the case of medical image processing. Image segmentation quality is the main metric that is often considered with memory and computation efficiency overlooked, limiting the use of power hungry models for practical use. In this paper, we propose a novel framework (Kidney-SegNet) that combines the effectiveness of an attention based encoder-decoder architecture with atrous spatial pyramid pooling with highly efficient dimension-wise convolutions. The segmentation results of the proposed Kidney-SegNet architecture have been shown to outperform existing state-of-the-art deep learning methods by evaluating them on two publicly available kidney and TNBC breast H&E stained histopathology image datasets. Further, our simulation experiments also reveal that the computational complexity and memory requirement of our proposed architecture is very efficient compared to existing deep learning state-of-the-art methods for the task of nuclei segmentation of H&E stained histopathology images. The source code of our implementation will be available at https://github.com/Aaatresh/Kidney-SegNet. © 2021 Elsevier Ltd
  • Item
    DIResUNet: Architecture for multiclass semantic segmentation of high resolution remote sensing imagery data
    (Springer, 2022) Priyanka; Sravya, N.; Lal, S.; Nalini, J.; Chintala, C.S.; Dell’Acqua, F.
    Scene understanding is an important task in information extraction from high-resolution aerial images, an operation which is often involved in remote sensing applications. Recently, semantic segmentation using deep learning has become an important method to achieve state-of-the-art performance in pixel-level classification of objects. This latter is still a challenging task due to large pixel variance within classes possibly coupled with small pixel variance between classes. This paper proposes an artificial-intelligence (AI)-based approach to this problem, by designing the DIResUNet deep learning model. The model is built by integrating the inception module, a modified residual block, and a dense global spatial pyramid pooling (DGSPP) module, in combination with the well-known U-Net scheme. The modified residual blocks and the inception module extract multi-level features, whereas DGSPP extracts contextual intelligence. In this way, both local and global information about the scene are extracted in parallel using dedicated processing structures, resulting in a more effective overall approach. The performance of the proposed DIResUNet model is evaluated on the Landcover and WHDLD high resolution remote sensing (HRRS) datasets. We compared DIResUNet performance with recent benchmark models such as U-Net, UNet++, Attention UNet, FPN, UNet+SPP, and DGRNet to prove the effectiveness of our proposed model. Results show that the proposed DIResUNet model outperforms benchmark models on two HRRS datasets. © 2022, The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature.
  • Item
    BCDetNet: a deep learning architecture for building change detection from bi-temporal high resolution satellite images
    (Springer Science and Business Media Deutschland GmbH, 2023) Basavaraju, K.S.; Hiren, N.S.; Sravya, N.; Lal, S.; Nalini, J.; Chintala, C.S.
    Change detection is becoming more and more popular technology for the analysis of remote sensing data and is very important for an accurate understanding of changes that are happening in the Earth’s surface. Different Deep Learning methods proposed till now are mainly focused on simple networks which results in poor detection for small changed areas because they can not differentiate between the bi-temporal image’s characteristics. To solve this problem, this article proposes a novel Building Change Detection Network (BCDetNet) for building object change detection and its analysis from bi-temporal high resolution satellite image. The proposed BCDetNet model can detect small change areas with the help of multiple feature extraction block. The proposed BCDetNet model executes building change detection using bi-temporal high resolution satellite images. The proposed BCDetNet model is trained on two publicly available datasets namely LEVIR and WHU change detection(CD) datasets. These datasets contain RGB images with dimensions of (1024 × 1024) and (512 × 512), respectively. The BCDetNet model can learn from scratch during training and performs better than the benchmark change detection models with fewer trainable parameters. The BCDetNet model gives Recall—94.06%, Precision—93.00%, Jaccard score—88.40%, Accuracy—98.73%, F1 score—93.52% and Kappa coefficient—87.05% on LEVIR CD dataset and Recall—89.51%, Precision —92.78%, Jaccard score - 84.38%, Accuracy—96.78%, F1 score—91.06% and Kappa coefficient - 82.12% on WHU CD dataset. This work is a step in the direction of achieving best results in building change detection from high resolution satellite images. © 2023, The Author(s), under exclusive licence to Springer-Verlag GmbH Germany, part of Springer Nature.
  • Item
    FPGA implementation of deep learning architecture for kidney cancer detection from histopathological images
    (Springer, 2024) Lal, S.; Chanchal, A.K.; Kini, J.; Upadhyay, G.K.
    Kidney cancer is the most common type of cancer, and designing an automated system to accurately classify the cancer grade is of paramount importance for a better prognosis of the disease from histopathological kidney cancer images. Application of deep learning neural networks (DLNNs) for histopathological image classification is thriving and implementation of these networks on edge devices has been gaining the ground correspondingly due to high computational power and low latency requirements. This paper designs an automated system that classifies histopathological kidney cancer images. For experimentation, we have collected Kidney histopathological images of Non-cancerous, cancerous, and their respective grade of Renal Cell Carcinoma (RCC) from Kasturba Medical College (KMC), Mangalore, Karnataka, India. We have implemented and analyzed performances of deep learning architectures on a Field Programmable Gate Array (FPGA) board. Results yield that the Inception-V3 network provides better accuracy for kidney cancer detection as compared to other deep learning models on Kidney histopathological images. Further, the DenseNet-169 network provides better accuracy for kidney cancer grading as compared to other existing deep learning architecture on the FPGA board. © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2023.
  • Item
    TransSARNet: a deep learning framework for despeckling of SAR images
    (Institute of Physics, 2025) Kevala, V.D.; Sravya, N.; Lal, S.; Suresh, S.; Dell’Acqua, F.
    Synthetic Aperture Radar(SAR) images are extensively used for Earth observation because of their all-weather, day, and night imaging capabilities. However, speckle noise in SAR images significantly reduces their usability in a variety of applications. Deep learning models developed for SAR despeckling exhibit promising noise reduction capabilities. Bringing a balance between reducing graininess and preserving texture details is a challenging task. In addition, supervised training of a robust deep learning model requires noisy images that capture the SAR speckle dynamics and the corresponding speckle-free ground truth, which is generally not available. This study proposes the first hybrid CNN-Halo attention-based transformer model for SAR despeckling. CNN-based feature extraction modules provide multiscale and multidirectional and large-scale feature maps. A halo-attention transformer block is used in the skip connection. It aids in the better preservation of radiometric information in the despeckled SAR images. TransSARNet is trained in a supervised manner using a new synthetic SAR dataset, which is a combination of the Kylberg and UCMerced land-use datasets. This study also analyzed the effect of combining the Kylberg and UCMerced datasets on texture preservation in despeckled SAR images. The visual and qualitative metrics evaluated on Sentinel-1 Single Look Complex SAR data showed that the proposed TransSARNet approach outperformed the other models under consideration. TransSARNet achieves a harmonious balance between model complexity, despeckling ability, edge preservation, radiometric information preservation, and smoothing in homogeneous regions. © 2025 IOP Publishing Ltd. All rights, including for text and data mining, AI training, and similar technologies, are reserved.
  • Item
    Transformer assisted framework for automated multi-class abnormality classification for video capsule endoscopy
    (Institute of Physics, 2025) Prabhu, M.M.; Kaliki, V.S.; Lal, S.
    Video Capsule Endoscopy (VCE) is a minimally invasive imaging technique used for diagnosing gastrointestinal (GI) disorders, enabling detailed visualization of the digestive tract. This study introduces CASCRNet, a novel and parameter-efficient deep learning architecture designed to enhance interpretability and computational efficiency in multi-class abnormality classification for VCE. CASCRNet integrates focal loss, Atrous Spatial Pyramid Pooling, and Shared Channel Residual blocks to improve feature extraction and address class imbalance. In addition to CASCRNet, this study conducts a comprehensive evaluation of several deep learning models, including ResNet50, DenseNet121, RCCGNet, Hiera, and AIMv2. Among these, AIMv2, a fine-tuned transformer-based model, achieved the highest overall performance, serving as a new benchmark for accuracy. The proposed framework demonstrates robust results on the Capsule Vision 2024 dataset and highlights the potential of both lightweight and transformer-based solutions to improve diagnostic efficiency and clinical workflow in gastrointestinal imaging. © 2025 IOP Publishing Ltd. All rights, including for text and data mining, AI training, and similar technologies, are reserved.