Faculty Publications
Permanent URI for this communityhttps://idr.nitk.ac.in/handle/123456789/18736
Publications by NITK Faculty
Browse
2 results
Search Results
Item NucleiSegNet: Robust deep learning architecture for the nuclei segmentation of liver cancer histopathology images(Elsevier Ltd, 2021) Lal, S.; Das, D.; Alabhya, K.; Kanfade, A.; Kumar, A.; Kini, J.R.The nuclei segmentation of hematoxylin and eosin (H&E) stained histopathology images is an important prerequisite in designing a computer-aided diagnostics (CAD) system for cancer diagnosis and prognosis. Automated nuclei segmentation methods enable the qualitative and quantitative analysis of tens of thousands of nuclei within H&E stained histopathology images. However, a major challenge during nuclei segmentation is the segmentation of variable sized, touching nuclei. To address this challenge, we present NucleiSegNet - a robust deep learning network architecture for the nuclei segmentation of H&E stained liver cancer histopathology images. Our proposed architecture includes three blocks: a robust residual block, a bottleneck block, and an attention decoder block. The robust residual block is a newly proposed block for the efficient extraction of high-level semantic maps. The attention decoder block uses a new attention mechanism for efficient object localization, and it improves the proposed architecture's performance by reducing false positives. When applied to nuclei segmentation tasks, the proposed deep-learning architecture yielded superior results compared to state-of-the-art nuclei segmentation methods. We applied our proposed deep learning architecture for nuclei segmentation to a set of H&E stained histopathology images from two datasets, and our comprehensive results show that our proposed architecture outperforms state-of-the-art methods. As part of this work, we also introduced a new liver dataset (KMC liver dataset) of H&E stained liver cancer histopathology image tiles, containing 80 images with annotated nuclei procured from Kasturba Medical College (KMC), Mangalore, Manipal Academy of Higher Education (MAHE), Manipal, Karnataka, India. The proposed model's source code is available at https://github.com/shyamfec/NucleiSegNet. © 2020 Elsevier LtdItem ProsGradNet: An effective and structured CNN approach for prostate cancer grading from histopathology images(Elsevier Ltd, 2025) Prabhu, A.; Sravya, N.; Lal, S.; Kini, J.Prostate cancer (PCa) is one of the most prevalent and potentially fatal malignancies affecting men globally. The incidence of prostate cancer is expected to double by 2040, posing significant health challenges. This anticipated increase underscores the urgent need for early and precise diagnosis to facilitate effective treatment and management. Histopathological analysis using Gleason grading system plays a pivotal role in clinical decision making by classifying cancer subtypes based on their cellular characteristics. This paper proposes a novel deep CNN model named as Prostate Grading Network (ProsGradNet), for the automatic grading of PCa from histopathological images. Central to the approach is the novel Context Guided Shared Channel Residual (CGSCR) block, that introduces structured methods for channel splitting and clustering, by varying group sizes. By grouping channels into 2, 4, and 8, it prioritizes deeper layer features, enhancing local semantic content and abstract feature representation. This methodological advancement significantly boosts classification accuracy, achieving an impressive 92.88% on Prostate Gleason dataset, outperforming other CNN models. To demonstrate the generalizability of ProsGradNet over different datasets, experiments are performed on Kasturba Medical College (KMC) Kidney dataset as well. The results further confirm the superiority of the proposed ProsGradNet model, with a classification accuracy of 92.68% on the KMC Kidney dataset. This demonstrates the model's potential to be applied effectively across various histopathological datasets, making it a valuable tool to fight against cancer. © 2025 Elsevier Ltd
