Faculty Publications
Permanent URI for this communityhttps://idr.nitk.ac.in/handle/123456789/18736
Publications by NITK Faculty
Browse
7 results
Search Results
Item Dense refinement residual network for road extraction from aerial imagery data(Institute of Electrical and Electronics Engineers Inc., 2019) Eerapu, K.K.; Ashwath, B.; Lal, S.; Dell’Acqua, F.; Narasimha Dhan, A.V.Extraction of roads from high-resolution aerial images with a high degree of accuracy is a prerequisite in various applications. In aerial images, road pixels and background pixels are generally in the ratio of ones-to-tens, which implies a class imbalance problem. Existing semantic segmentation architectures generally do well in road-dominated cases but fail in background-dominated scenarios. This paper proposes a dense refinement residual network (DRR Net) for semantic segmentation of aerial imagery data. The proposed semantic segmentation architecture is composed of multiple DRR modules for the extraction of diversified roads alleviating the class imbalance problem. Each module of the proposed architecture utilizes dense convolutions at various scales only in the encoder for feature learning. Residual connections in each module of the proposed architecture provide the guided learning path by propagating the combined features to subsequent DRR modules. Segmentation maps undergo various levels of refinement based on the number of DRR modules utilized in the architecture. To emphasize more on small object instances, the proposed architecture has been trained with a composite loss function. The qualitative and quantitative results are reported by utilizing the Massachusetts roads dataset. The experimental results report that the proposed architecture provides better results as compared to other recent architectures. © 2019 Institute of Electrical and Electronics Engineers Inc.. All rights reserved.Item NucleiSegNet: Robust deep learning architecture for the nuclei segmentation of liver cancer histopathology images(Elsevier Ltd, 2021) Lal, S.; Das, D.; Alabhya, K.; Kanfade, A.; Kumar, A.; Kini, J.R.The nuclei segmentation of hematoxylin and eosin (H&E) stained histopathology images is an important prerequisite in designing a computer-aided diagnostics (CAD) system for cancer diagnosis and prognosis. Automated nuclei segmentation methods enable the qualitative and quantitative analysis of tens of thousands of nuclei within H&E stained histopathology images. However, a major challenge during nuclei segmentation is the segmentation of variable sized, touching nuclei. To address this challenge, we present NucleiSegNet - a robust deep learning network architecture for the nuclei segmentation of H&E stained liver cancer histopathology images. Our proposed architecture includes three blocks: a robust residual block, a bottleneck block, and an attention decoder block. The robust residual block is a newly proposed block for the efficient extraction of high-level semantic maps. The attention decoder block uses a new attention mechanism for efficient object localization, and it improves the proposed architecture's performance by reducing false positives. When applied to nuclei segmentation tasks, the proposed deep-learning architecture yielded superior results compared to state-of-the-art nuclei segmentation methods. We applied our proposed deep learning architecture for nuclei segmentation to a set of H&E stained histopathology images from two datasets, and our comprehensive results show that our proposed architecture outperforms state-of-the-art methods. As part of this work, we also introduced a new liver dataset (KMC liver dataset) of H&E stained liver cancer histopathology image tiles, containing 80 images with annotated nuclei procured from Kasturba Medical College (KMC), Mangalore, Manipal Academy of Higher Education (MAHE), Manipal, Karnataka, India. The proposed model's source code is available at https://github.com/shyamfec/NucleiSegNet. © 2020 Elsevier LtdItem O-SegNet: Robust Encoder and Decoder Architecture for Objects Segmentation from Aerial Imagery Data(Institute of Electrical and Electronics Engineers Inc., 2022) Eerapu, K.K.; Lal, S.; Narasimhadhan, A.V.The segmentation of diversified roads and buildings from high-resolution aerial images is essential for various applications, such as urban planning, disaster assessment, traffic congestion management, and up-to-date road maps. However, a major challenge during object segmentation is the segmentation of small-sized, diverse shaped roads, and buildings in dominant background scenarios. We introduce O-SegNet- the robust encoder and decoder architecture for objects segmentation from high-resolution aerial imagery data to address this challenge. The proposed O-SegNet architecture contains Guided-Attention (GA) blocks in the encoder and decoder to focus on salient features by representing the spatial dependencies between features of different scales. Further, GA blocks guide the successive stages of encoder and decoder by interrelating the pixels of the same class. To emphasize more on relevant context, the attention mechanism is provided between encoder and decoder after aggregating the global context via an 8 Level Pyramid Pooling Network (PPN). The qualitative and quantitative results of the proposed and existing semantic segmentation architectures are evaluated by utilizing the dataset provided by Kaiser et al. Further, we show that the proposed O-SegNet architecture outperforms state-of-the-art techniques by accurately preserving the road connectivity and structure of buildings. © 2017 IEEE.Item DIResUNet: Architecture for multiclass semantic segmentation of high resolution remote sensing imagery data(Springer, 2022) Priyanka; Sravya, N.; Lal, S.; Nalini, J.; Chintala, C.S.; Dell’Acqua, F.Scene understanding is an important task in information extraction from high-resolution aerial images, an operation which is often involved in remote sensing applications. Recently, semantic segmentation using deep learning has become an important method to achieve state-of-the-art performance in pixel-level classification of objects. This latter is still a challenging task due to large pixel variance within classes possibly coupled with small pixel variance between classes. This paper proposes an artificial-intelligence (AI)-based approach to this problem, by designing the DIResUNet deep learning model. The model is built by integrating the inception module, a modified residual block, and a dense global spatial pyramid pooling (DGSPP) module, in combination with the well-known U-Net scheme. The modified residual blocks and the inception module extract multi-level features, whereas DGSPP extracts contextual intelligence. In this way, both local and global information about the scene are extracted in parallel using dedicated processing structures, resulting in a more effective overall approach. The performance of the proposed DIResUNet model is evaluated on the Landcover and WHDLD high resolution remote sensing (HRRS) datasets. We compared DIResUNet performance with recent benchmark models such as U-Net, UNet++, Attention UNet, FPN, UNet+SPP, and DGRNet to prove the effectiveness of our proposed model. Results show that the proposed DIResUNet model outperforms benchmark models on two HRRS datasets. © 2022, The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature.Item DPPNet: An Efficient and Robust Deep Learning Network for Land Cover Segmentation From High-Resolution Satellite Images(Institute of Electrical and Electronics Engineers Inc., 2023) Sravya, N.; Priyanka; Lal, S.; Nalini, J.; Chintala, C.S.; Dell’Acqua, F.Visual understanding of land cover is an important task in information extraction from high-resolution satellite images, an operation which is often involved in remote sensing applications. Multi-class semantic segmentation of high-resolution satellite images turned out to be an important research topic because of its wide range of real-life applications. Although scientific literature reports several deep learning methods that can provide good results in segmenting remotely sensed images, these are generally computationally expensive. There still exists an open challenge towards developing a robust deep learning model capable of improving performances while requiring less computational complexity. In this article, we propose a new model termed DPPNet (Depth-wise Pyramid Pooling Network), which uses the newly designed Depth-wise Pyramid Pooling (DPP) block and a dense block with multi-dilated depth-wise residual connections. This proposed DPPNet model is evaluated and compared with the benchmark semantic segmentation models on the Land-cover and WHDLD high-resolution Space-borne Sensor (HRS) datasets. The proposed model provides DC, IoU, OA, Ka scores of (88.81%, 78.29%), (76.35%, 60.92%), (87.15%, 81.02%), (77.86%, 72.73%) on the Land-cover and WHDLD HRS datasets respectively. Results show that the proposed DPPNet model provides better performances, in both quantitative and qualitative terms, on these standard benchmark datasets than current state-of-art methods. © 2017 IEEE.Item RSCDNet: A Robust Deep Learning Architecture for Change Detection From Bi-Temporal High Resolution Remote Sensing Images(Institute of Electrical and Electronics Engineers Inc., 2023) Deepanshi; Barkur, R.; Suresh, D.; Lal, S.; Chintala, C.S.; Diwakar, P.G.Accurate change detection from high-resolution satellite and aerial images is of great significance in remote sensing for precise comprehension of Land cover (LC) variations. The current methods compromise with the spatial context; hence, they fail to detect and delineate small change areas and are unable to capture the difference between features of the bi-temporal images. This paper proposes Remote Sensing Change Detection Network (RSCDNet) - a robust end-to-end deep learning architecture for pixel-wise change detection from bi-temporal high-resolution remote-sensing (HRRS) images. The proposed RSCDNet model is based on an encoder-decoder framework integrated with the Modified Self-Attention (MSA) andthe Gated Linear Atrous Spatial Pyramid Pooling (GL-ASPP) blocks; both efficient mechanisms to regulate the field-of-view while finding the most suitable trade-off between accurate localization and context assimilation. The paper documents the design and development of the proposed RSCDNet model and compares its qualitative and quantitative results with state-of-the-art HRRS change detection architectures. The above mentioned novelties in the proposed architecture resulted in an F1-score of 98%, 98%, 88%, and 75% on the four publicly available HRRS datasets namely, Staza-Tisadob, Onera, CD-LEVIR, and WHU. In addition to the improvement in the performance metrics, the strategic connections in the proposed GL-ASPP and MSA units significantly reduce the prediction time per image (PTPI) and provide robustness against perturbations. Experimental results yield that the proposed RSCDNet model outperforms the most recent change detection benchmark models on all four HRRS datasets. © 2017 IEEE.Item SFSCDNet: A Deep Learning Model With Spatial Flow-Based Semantic Change Detection From Bi-Temporal Satellite Images(Institute of Electrical and Electronics Engineers Inc., 2024) Basavaraju, K.S.; Sravya, N.; Kevala, V.D.; Suresh, S.; Lal, S.Semantic change detection in remote sensing imagery plays a pivotal role in urban planning, environmental monitoring, and disaster assessment applications. Existing deep learning-based methods, particularly those relying on triple-branch architectures, often struggle to accurately localize and predict changes in complex spatial environments characterized by diverse land-cover types. To overcome these limitations, this paper proposes a novel network called the Spatial Flow-based Semantic Change Detection Network. This network processes bi-temporal satellite images using a dual-encoder, triple-decoder architecture that progressively refines spatial features at each network stage, improving semantic change detection results. The Attention-Based Siamese Encoder, Cascaded Convolutional Attention Fusion Block, Cascaded Convolutional Attention Refinement Block and Differentiable Binarization layer helps in improving semantic change detection performance. Experimental results of proposed network on the SECOND dataset demonstrate that the proposed model significantly improves the ability to localize critical changes and distinguish between change and no-change regions. The proposed network achieves an overall accuracy of 86.32%, a mean Intersection over Union of 70.33%, a Separated Kappa of 21.21%, and an F1-score for semantic change detection of 66.01%, with a score of 35.94%. These results represent substantial improvements over previous state-of-the-art models, including a 0.26% increase in overall accuracy, a 2.21% increase in mean Intersection over Union, a 2.62% enhancement in Separated Kappa, and a 3.6% improvement in F1-score for semantic change detection compared to the best-performing models. Notably, the proposed network achieves these results with only 14.56 million parameters, making it more effective and efficient than its competitors, which utilize over 22 million parameters. © 2013 IEEE.
