Faculty Publications
Permanent URI for this communityhttps://idr.nitk.ac.in/handle/123456789/18736
Publications by NITK Faculty
Browse
8 results
Search Results
Item UCDNet: A Deep Learning Model for Urban Change Detection From Bi-Temporal Multispectral Sentinel-2 Satellite Images(Institute of Electrical and Electronics Engineers Inc., 2022) Basavaraju, K.S.; Sravya, N.; Lal, S.; Nalini, J.; Chintala, C.S.; Dell’Acqua, F.Change detection (CD) from satellite images has become an inevitable process in earth observation. Methods for detecting changes in multi-temporal satellite images are very useful tools when characterization and monitoring of urban growth patterns is concerned. Increasing worldwide availability of multispectral images with a high revisit frequency opened up more possibilities in the study of urban CD. Even though there exists several deep learning methods for CD, most of these available methods fail to predict the edges and preserve the shape of the changed area from multispectral images. This article introduces a deep learning model called urban CD network (UCDNet) for urban CD from bi-temporal multispectral Sentinel-2 satellite images. The model is based on an encoder-decoder architecture which uses modified residual connections and the new spatial pyramid pooling (NSPP) block, giving better predictions while preserving the shape of changed areas. The modified residual connections help locate the changes correctly, and the NSPP block can extract multiscale features and will give awareness about global context. UCDNet uses a proposed loss function which is a combination of weighted class categorical cross-entropy (WCCE) and modified Kappa loss. The Onera Satellite Change Detection (OSCD) dataset is used to train, evaluate, and compare the proposed model with the benchmark models. UCDNet gives better results from the reference models used here for comparison. It gives an accuracy of 99.3%, an $F1$ score ( $F1$ ) of 89.21%, a Kappa coefficient (Ka) of 88.85%, and a Jaccard index (JI) of 80.53% on the OSCD dataset. © 1980-2012 IEEE.Item DIResUNet: Architecture for multiclass semantic segmentation of high resolution remote sensing imagery data(Springer, 2022) Priyanka; Sravya, N.; Lal, S.; Nalini, J.; Chintala, C.S.; Dell’Acqua, F.Scene understanding is an important task in information extraction from high-resolution aerial images, an operation which is often involved in remote sensing applications. Recently, semantic segmentation using deep learning has become an important method to achieve state-of-the-art performance in pixel-level classification of objects. This latter is still a challenging task due to large pixel variance within classes possibly coupled with small pixel variance between classes. This paper proposes an artificial-intelligence (AI)-based approach to this problem, by designing the DIResUNet deep learning model. The model is built by integrating the inception module, a modified residual block, and a dense global spatial pyramid pooling (DGSPP) module, in combination with the well-known U-Net scheme. The modified residual blocks and the inception module extract multi-level features, whereas DGSPP extracts contextual intelligence. In this way, both local and global information about the scene are extracted in parallel using dedicated processing structures, resulting in a more effective overall approach. The performance of the proposed DIResUNet model is evaluated on the Landcover and WHDLD high resolution remote sensing (HRRS) datasets. We compared DIResUNet performance with recent benchmark models such as U-Net, UNet++, Attention UNet, FPN, UNet+SPP, and DGRNet to prove the effectiveness of our proposed model. Results show that the proposed DIResUNet model outperforms benchmark models on two HRRS datasets. © 2022, The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature.Item DPPNet: An Efficient and Robust Deep Learning Network for Land Cover Segmentation From High-Resolution Satellite Images(Institute of Electrical and Electronics Engineers Inc., 2023) Sravya, N.; Priyanka; Lal, S.; Nalini, J.; Chintala, C.S.; Dell’Acqua, F.Visual understanding of land cover is an important task in information extraction from high-resolution satellite images, an operation which is often involved in remote sensing applications. Multi-class semantic segmentation of high-resolution satellite images turned out to be an important research topic because of its wide range of real-life applications. Although scientific literature reports several deep learning methods that can provide good results in segmenting remotely sensed images, these are generally computationally expensive. There still exists an open challenge towards developing a robust deep learning model capable of improving performances while requiring less computational complexity. In this article, we propose a new model termed DPPNet (Depth-wise Pyramid Pooling Network), which uses the newly designed Depth-wise Pyramid Pooling (DPP) block and a dense block with multi-dilated depth-wise residual connections. This proposed DPPNet model is evaluated and compared with the benchmark semantic segmentation models on the Land-cover and WHDLD high-resolution Space-borne Sensor (HRS) datasets. The proposed model provides DC, IoU, OA, Ka scores of (88.81%, 78.29%), (76.35%, 60.92%), (87.15%, 81.02%), (77.86%, 72.73%) on the Land-cover and WHDLD HRS datasets respectively. Results show that the proposed DPPNet model provides better performances, in both quantitative and qualitative terms, on these standard benchmark datasets than current state-of-art methods. © 2017 IEEE.Item BCDetNet: a deep learning architecture for building change detection from bi-temporal high resolution satellite images(Springer Science and Business Media Deutschland GmbH, 2023) Basavaraju, K.S.; Hiren, N.S.; Sravya, N.; Lal, S.; Nalini, J.; Chintala, C.S.Change detection is becoming more and more popular technology for the analysis of remote sensing data and is very important for an accurate understanding of changes that are happening in the Earth’s surface. Different Deep Learning methods proposed till now are mainly focused on simple networks which results in poor detection for small changed areas because they can not differentiate between the bi-temporal image’s characteristics. To solve this problem, this article proposes a novel Building Change Detection Network (BCDetNet) for building object change detection and its analysis from bi-temporal high resolution satellite image. The proposed BCDetNet model can detect small change areas with the help of multiple feature extraction block. The proposed BCDetNet model executes building change detection using bi-temporal high resolution satellite images. The proposed BCDetNet model is trained on two publicly available datasets namely LEVIR and WHU change detection(CD) datasets. These datasets contain RGB images with dimensions of (1024 × 1024) and (512 × 512), respectively. The BCDetNet model can learn from scratch during training and performs better than the benchmark change detection models with fewer trainable parameters. The BCDetNet model gives Recall—94.06%, Precision—93.00%, Jaccard score—88.40%, Accuracy—98.73%, F1 score—93.52% and Kappa coefficient—87.05% on LEVIR CD dataset and Recall—89.51%, Precision —92.78%, Jaccard score - 84.38%, Accuracy—96.78%, F1 score—91.06% and Kappa coefficient - 82.12% on WHU CD dataset. This work is a step in the direction of achieving best results in building change detection from high resolution satellite images. © 2023, The Author(s), under exclusive licence to Springer-Verlag GmbH Germany, part of Springer Nature.Item A Robust CNN Framework for Change Detection Analysis From Bitemporal Remote Sensing Images(Institute of Electrical and Electronics Engineers Inc., 2024) Sravya, N.; Bhaduka, K.; Lal, S.; Nalini, J.; Chintala, C.S.—Deep learning (DL) algorithms are currently the most effective methods for change detection (CD) from high-resolution multispectral (MS) remote-sensing (RS) images. Because a variety of satellites are able to provide a lot of data, it is now easy to find changes using efficient DL models. Current CD methods focus on simple structure and combining the features obtained by all the stages together rather than extracting multiscale features from a single stage since it may lead to information loss and an imbalance contribution of features at different stages. This in turn results in misclassification of small changed areas and poor edge and shape preservation of changed areas. This article introduces an enhanced RSCD network (ERSCDNet) for CD from bitemporal aerial and MS images. The proposed encoder–decoder-based ERSCDNet model uses an attention-based encoder and decoder block and a modified new spatial pyramid pooling block at each stage of the decoder part, which effectively utilize features at each encoder stages and prevent information loss. The learning, vision, and remote sensing CD (LEVIR-CD), Onera satellite change detection (OSCD), and Sun Yat-Sen University CD (SYSU-CD) datasets are used to evaluate the ERSCDNet model. The ERSCDNet gives better performance than all the models used in this article for comparison. It gives an F1 score, a Kappa coefficient, and a Jaccard index of (0.9306, 0.9282, 0.8703), (0.8945, 0.8887, 0.8091), and (0.7581, 0.6876, 0.6103) on OSCD, LEVIR-CD, and SYSU-CD datasets, respectively. © 2024 The Authors. This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License.Item SFSCDNet: A Deep Learning Model With Spatial Flow-Based Semantic Change Detection From Bi-Temporal Satellite Images(Institute of Electrical and Electronics Engineers Inc., 2024) Basavaraju, K.S.; Sravya, N.; Kevala, V.D.; Suresh, S.; Lal, S.Semantic change detection in remote sensing imagery plays a pivotal role in urban planning, environmental monitoring, and disaster assessment applications. Existing deep learning-based methods, particularly those relying on triple-branch architectures, often struggle to accurately localize and predict changes in complex spatial environments characterized by diverse land-cover types. To overcome these limitations, this paper proposes a novel network called the Spatial Flow-based Semantic Change Detection Network. This network processes bi-temporal satellite images using a dual-encoder, triple-decoder architecture that progressively refines spatial features at each network stage, improving semantic change detection results. The Attention-Based Siamese Encoder, Cascaded Convolutional Attention Fusion Block, Cascaded Convolutional Attention Refinement Block and Differentiable Binarization layer helps in improving semantic change detection performance. Experimental results of proposed network on the SECOND dataset demonstrate that the proposed model significantly improves the ability to localize critical changes and distinguish between change and no-change regions. The proposed network achieves an overall accuracy of 86.32%, a mean Intersection over Union of 70.33%, a Separated Kappa of 21.21%, and an F1-score for semantic change detection of 66.01%, with a score of 35.94%. These results represent substantial improvements over previous state-of-the-art models, including a 0.26% increase in overall accuracy, a 2.21% increase in mean Intersection over Union, a 2.62% enhancement in Separated Kappa, and a 3.6% improvement in F1-score for semantic change detection compared to the best-performing models. Notably, the proposed network achieves these results with only 14.56 million parameters, making it more effective and efficient than its competitors, which utilize over 22 million parameters. © 2013 IEEE.Item ProsGradNet: An effective and structured CNN approach for prostate cancer grading from histopathology images(Elsevier Ltd, 2025) Prabhu, A.; Sravya, N.; Lal, S.; Kini, J.Prostate cancer (PCa) is one of the most prevalent and potentially fatal malignancies affecting men globally. The incidence of prostate cancer is expected to double by 2040, posing significant health challenges. This anticipated increase underscores the urgent need for early and precise diagnosis to facilitate effective treatment and management. Histopathological analysis using Gleason grading system plays a pivotal role in clinical decision making by classifying cancer subtypes based on their cellular characteristics. This paper proposes a novel deep CNN model named as Prostate Grading Network (ProsGradNet), for the automatic grading of PCa from histopathological images. Central to the approach is the novel Context Guided Shared Channel Residual (CGSCR) block, that introduces structured methods for channel splitting and clustering, by varying group sizes. By grouping channels into 2, 4, and 8, it prioritizes deeper layer features, enhancing local semantic content and abstract feature representation. This methodological advancement significantly boosts classification accuracy, achieving an impressive 92.88% on Prostate Gleason dataset, outperforming other CNN models. To demonstrate the generalizability of ProsGradNet over different datasets, experiments are performed on Kasturba Medical College (KMC) Kidney dataset as well. The results further confirm the superiority of the proposed ProsGradNet model, with a classification accuracy of 92.68% on the KMC Kidney dataset. This demonstrates the model's potential to be applied effectively across various histopathological datasets, making it a valuable tool to fight against cancer. © 2025 Elsevier LtdItem TransSARNet: a deep learning framework for despeckling of SAR images(Institute of Physics, 2025) Kevala, V.D.; Sravya, N.; Lal, S.; Suresh, S.; Dell’Acqua, F.Synthetic Aperture Radar(SAR) images are extensively used for Earth observation because of their all-weather, day, and night imaging capabilities. However, speckle noise in SAR images significantly reduces their usability in a variety of applications. Deep learning models developed for SAR despeckling exhibit promising noise reduction capabilities. Bringing a balance between reducing graininess and preserving texture details is a challenging task. In addition, supervised training of a robust deep learning model requires noisy images that capture the SAR speckle dynamics and the corresponding speckle-free ground truth, which is generally not available. This study proposes the first hybrid CNN-Halo attention-based transformer model for SAR despeckling. CNN-based feature extraction modules provide multiscale and multidirectional and large-scale feature maps. A halo-attention transformer block is used in the skip connection. It aids in the better preservation of radiometric information in the despeckled SAR images. TransSARNet is trained in a supervised manner using a new synthetic SAR dataset, which is a combination of the Kylberg and UCMerced land-use datasets. This study also analyzed the effect of combining the Kylberg and UCMerced datasets on texture preservation in despeckled SAR images. The visual and qualitative metrics evaluated on Sentinel-1 Single Look Complex SAR data showed that the proposed TransSARNet approach outperformed the other models under consideration. TransSARNet achieves a harmonious balance between model complexity, despeckling ability, edge preservation, radiometric information preservation, and smoothing in homogeneous regions. © 2025 IOP Publishing Ltd. All rights, including for text and data mining, AI training, and similar technologies, are reserved.
