Faculty Publications
Permanent URI for this communityhttps://idr.nitk.ac.in/handle/123456789/18736
Publications by NITK Faculty
Browse
3 results
Search Results
Item DIResUNet: Architecture for multiclass semantic segmentation of high resolution remote sensing imagery data(Springer, 2022) Priyanka; Sravya, N.; Lal, S.; Nalini, J.; Chintala, C.S.; Dell’Acqua, F.Scene understanding is an important task in information extraction from high-resolution aerial images, an operation which is often involved in remote sensing applications. Recently, semantic segmentation using deep learning has become an important method to achieve state-of-the-art performance in pixel-level classification of objects. This latter is still a challenging task due to large pixel variance within classes possibly coupled with small pixel variance between classes. This paper proposes an artificial-intelligence (AI)-based approach to this problem, by designing the DIResUNet deep learning model. The model is built by integrating the inception module, a modified residual block, and a dense global spatial pyramid pooling (DGSPP) module, in combination with the well-known U-Net scheme. The modified residual blocks and the inception module extract multi-level features, whereas DGSPP extracts contextual intelligence. In this way, both local and global information about the scene are extracted in parallel using dedicated processing structures, resulting in a more effective overall approach. The performance of the proposed DIResUNet model is evaluated on the Landcover and WHDLD high resolution remote sensing (HRRS) datasets. We compared DIResUNet performance with recent benchmark models such as U-Net, UNet++, Attention UNet, FPN, UNet+SPP, and DGRNet to prove the effectiveness of our proposed model. Results show that the proposed DIResUNet model outperforms benchmark models on two HRRS datasets. © 2022, The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature.Item RSCDNet: A Robust Deep Learning Architecture for Change Detection From Bi-Temporal High Resolution Remote Sensing Images(Institute of Electrical and Electronics Engineers Inc., 2023) Deepanshi; Barkur, R.; Suresh, D.; Lal, S.; Chintala, C.S.; Diwakar, P.G.Accurate change detection from high-resolution satellite and aerial images is of great significance in remote sensing for precise comprehension of Land cover (LC) variations. The current methods compromise with the spatial context; hence, they fail to detect and delineate small change areas and are unable to capture the difference between features of the bi-temporal images. This paper proposes Remote Sensing Change Detection Network (RSCDNet) - a robust end-to-end deep learning architecture for pixel-wise change detection from bi-temporal high-resolution remote-sensing (HRRS) images. The proposed RSCDNet model is based on an encoder-decoder framework integrated with the Modified Self-Attention (MSA) andthe Gated Linear Atrous Spatial Pyramid Pooling (GL-ASPP) blocks; both efficient mechanisms to regulate the field-of-view while finding the most suitable trade-off between accurate localization and context assimilation. The paper documents the design and development of the proposed RSCDNet model and compares its qualitative and quantitative results with state-of-the-art HRRS change detection architectures. The above mentioned novelties in the proposed architecture resulted in an F1-score of 98%, 98%, 88%, and 75% on the four publicly available HRRS datasets namely, Staza-Tisadob, Onera, CD-LEVIR, and WHU. In addition to the improvement in the performance metrics, the strategic connections in the proposed GL-ASPP and MSA units significantly reduce the prediction time per image (PTPI) and provide robustness against perturbations. Experimental results yield that the proposed RSCDNet model outperforms the most recent change detection benchmark models on all four HRRS datasets. © 2017 IEEE.Item A Robust CNN Framework for Change Detection Analysis From Bitemporal Remote Sensing Images(Institute of Electrical and Electronics Engineers Inc., 2024) Sravya, N.; Bhaduka, K.; Lal, S.; Nalini, J.; Chintala, C.S.—Deep learning (DL) algorithms are currently the most effective methods for change detection (CD) from high-resolution multispectral (MS) remote-sensing (RS) images. Because a variety of satellites are able to provide a lot of data, it is now easy to find changes using efficient DL models. Current CD methods focus on simple structure and combining the features obtained by all the stages together rather than extracting multiscale features from a single stage since it may lead to information loss and an imbalance contribution of features at different stages. This in turn results in misclassification of small changed areas and poor edge and shape preservation of changed areas. This article introduces an enhanced RSCD network (ERSCDNet) for CD from bitemporal aerial and MS images. The proposed encoder–decoder-based ERSCDNet model uses an attention-based encoder and decoder block and a modified new spatial pyramid pooling block at each stage of the decoder part, which effectively utilize features at each encoder stages and prevent information loss. The learning, vision, and remote sensing CD (LEVIR-CD), Onera satellite change detection (OSCD), and Sun Yat-Sen University CD (SYSU-CD) datasets are used to evaluate the ERSCDNet model. The ERSCDNet gives better performance than all the models used in this article for comparison. It gives an F1 score, a Kappa coefficient, and a Jaccard index of (0.9306, 0.9282, 0.8703), (0.8945, 0.8887, 0.8091), and (0.7581, 0.6876, 0.6103) on OSCD, LEVIR-CD, and SYSU-CD datasets, respectively. © 2024 The Authors. This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License.
