Faculty Publications

Permanent URI for this communityhttps://idr.nitk.ac.in/handle/123456789/18736

Publications by NITK Faculty

Browse

Search Results

Now showing 1 - 8 of 8
  • Item
    A robust framework for quality enhancement of aerial remote sensing images
    (Elsevier B.V., 2018) Karuna Kumari, E.; Das, D.; Suresh, S.; Lal, S.; Narasimhadhan, A.V.
    This paper proposes a robust framework for quality restoration of remotely sensed aerial images. Proposed framework works in three steps: (1) Efficient color balancing and saturation adjustment, (2) Efficient color restoration, (3) Modified contrast enhancement using particle swarm optimization (PSO). In order to show the robustness, step-wise results of proposed framework is illustrated. Several aerial images from two publically available datasets are tested to support the robustness of the proposed framework over existing image quality restoration methods. The experimental results of proposed framework and other existing quality restoration methods are compared in terms of NIQMC, BIQME, MICHELSON, DE, EME and PIXDIST along with visual experimental results. Based on experimental results conducted on several aerial images suggest that the proposed framework is outperform over existing quality restoration methods. © 2018 Elsevier B.V.
  • Item
    Dense refinement residual network for road extraction from aerial imagery data
    (Institute of Electrical and Electronics Engineers Inc., 2019) Eerapu, K.K.; Ashwath, B.; Lal, S.; Dell’Acqua, F.; Narasimha Dhan, A.V.
    Extraction of roads from high-resolution aerial images with a high degree of accuracy is a prerequisite in various applications. In aerial images, road pixels and background pixels are generally in the ratio of ones-to-tens, which implies a class imbalance problem. Existing semantic segmentation architectures generally do well in road-dominated cases but fail in background-dominated scenarios. This paper proposes a dense refinement residual network (DRR Net) for semantic segmentation of aerial imagery data. The proposed semantic segmentation architecture is composed of multiple DRR modules for the extraction of diversified roads alleviating the class imbalance problem. Each module of the proposed architecture utilizes dense convolutions at various scales only in the encoder for feature learning. Residual connections in each module of the proposed architecture provide the guided learning path by propagating the combined features to subsequent DRR modules. Segmentation maps undergo various levels of refinement based on the number of DRR modules utilized in the architecture. To emphasize more on small object instances, the proposed architecture has been trained with a composite loss function. The qualitative and quantitative results are reported by utilizing the Massachusetts roads dataset. The experimental results report that the proposed architecture provides better results as compared to other recent architectures. © 2019 Institute of Electrical and Electronics Engineers Inc.. All rights reserved.
  • Item
    O-SegNet: Robust Encoder and Decoder Architecture for Objects Segmentation from Aerial Imagery Data
    (Institute of Electrical and Electronics Engineers Inc., 2022) Eerapu, K.K.; Lal, S.; Narasimhadhan, A.V.
    The segmentation of diversified roads and buildings from high-resolution aerial images is essential for various applications, such as urban planning, disaster assessment, traffic congestion management, and up-to-date road maps. However, a major challenge during object segmentation is the segmentation of small-sized, diverse shaped roads, and buildings in dominant background scenarios. We introduce O-SegNet- the robust encoder and decoder architecture for objects segmentation from high-resolution aerial imagery data to address this challenge. The proposed O-SegNet architecture contains Guided-Attention (GA) blocks in the encoder and decoder to focus on salient features by representing the spatial dependencies between features of different scales. Further, GA blocks guide the successive stages of encoder and decoder by interrelating the pixels of the same class. To emphasize more on relevant context, the attention mechanism is provided between encoder and decoder after aggregating the global context via an 8 Level Pyramid Pooling Network (PPN). The qualitative and quantitative results of the proposed and existing semantic segmentation architectures are evaluated by utilizing the dataset provided by Kaiser et al. Further, we show that the proposed O-SegNet architecture outperforms state-of-the-art techniques by accurately preserving the road connectivity and structure of buildings. © 2017 IEEE.
  • Item
    DIResUNet: Architecture for multiclass semantic segmentation of high resolution remote sensing imagery data
    (Springer, 2022) Priyanka; Sravya, N.; Lal, S.; Nalini, J.; Chintala, C.S.; Dell’Acqua, F.
    Scene understanding is an important task in information extraction from high-resolution aerial images, an operation which is often involved in remote sensing applications. Recently, semantic segmentation using deep learning has become an important method to achieve state-of-the-art performance in pixel-level classification of objects. This latter is still a challenging task due to large pixel variance within classes possibly coupled with small pixel variance between classes. This paper proposes an artificial-intelligence (AI)-based approach to this problem, by designing the DIResUNet deep learning model. The model is built by integrating the inception module, a modified residual block, and a dense global spatial pyramid pooling (DGSPP) module, in combination with the well-known U-Net scheme. The modified residual blocks and the inception module extract multi-level features, whereas DGSPP extracts contextual intelligence. In this way, both local and global information about the scene are extracted in parallel using dedicated processing structures, resulting in a more effective overall approach. The performance of the proposed DIResUNet model is evaluated on the Landcover and WHDLD high resolution remote sensing (HRRS) datasets. We compared DIResUNet performance with recent benchmark models such as U-Net, UNet++, Attention UNet, FPN, UNet+SPP, and DGRNet to prove the effectiveness of our proposed model. Results show that the proposed DIResUNet model outperforms benchmark models on two HRRS datasets. © 2022, The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature.
  • Item
    A new deep learning architecture for dehazing of aerial remote sensing images
    (Springer, 2022) Kalra, A.; Sequeira, A.; Manjunath, A.; Lal, S.; Raghavendra, R.
    A major problem in most aerial remote image processing applications is the presence of haze in images. It is a phenomenon by which particles in the atmosphere disperse light, thus altering the quality of the overall image. This can be detrimental to the performance of vision-based algorithms such as those concerned with object detection. There have been numerous attempts using traditional image processing techniques as well as using deep learning approaches to eliminate this haze. In most cases, models tend to make assumptions on the nature of haze that are rarely true in reality. In this paper, we propose an end-to-end deep learning architecture that can dehaze aerial remote sensing images efficiently with minimal deviation from the ground truth. Many of the assumptions made in other models are eliminated and the relationship between hazed and dehazed images is directly computed. The proposed model is based on the observation that identifying structural and statistical portions separately from an image and using those features to reconstruct the image can give a realistic dehazed image. It also makes use of information exposed by different color spaces to achieve this using lesser computation. The experimental quantitative and qualitative results of the proposed architecture are compared with recent benchmark dehaze models on NYU hazy dataset and real-world hazy images. Experimental results yield that the proposed architecture outperforms benchmark models on test aerial remote sensing images. © 2022, The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature.
  • Item
    RSCDNet: A Robust Deep Learning Architecture for Change Detection From Bi-Temporal High Resolution Remote Sensing Images
    (Institute of Electrical and Electronics Engineers Inc., 2023) Deepanshi; Barkur, R.; Suresh, D.; Lal, S.; Chintala, C.S.; Diwakar, P.G.
    Accurate change detection from high-resolution satellite and aerial images is of great significance in remote sensing for precise comprehension of Land cover (LC) variations. The current methods compromise with the spatial context; hence, they fail to detect and delineate small change areas and are unable to capture the difference between features of the bi-temporal images. This paper proposes Remote Sensing Change Detection Network (RSCDNet) - a robust end-to-end deep learning architecture for pixel-wise change detection from bi-temporal high-resolution remote-sensing (HRRS) images. The proposed RSCDNet model is based on an encoder-decoder framework integrated with the Modified Self-Attention (MSA) andthe Gated Linear Atrous Spatial Pyramid Pooling (GL-ASPP) blocks; both efficient mechanisms to regulate the field-of-view while finding the most suitable trade-off between accurate localization and context assimilation. The paper documents the design and development of the proposed RSCDNet model and compares its qualitative and quantitative results with state-of-the-art HRRS change detection architectures. The above mentioned novelties in the proposed architecture resulted in an F1-score of 98%, 98%, 88%, and 75% on the four publicly available HRRS datasets namely, Staza-Tisadob, Onera, CD-LEVIR, and WHU. In addition to the improvement in the performance metrics, the strategic connections in the proposed GL-ASPP and MSA units significantly reduce the prediction time per image (PTPI) and provide robustness against perturbations. Experimental results yield that the proposed RSCDNet model outperforms the most recent change detection benchmark models on all four HRRS datasets. © 2017 IEEE.
  • Item
    A Robust CNN Framework for Change Detection Analysis From Bitemporal Remote Sensing Images
    (Institute of Electrical and Electronics Engineers Inc., 2024) Sravya, N.; Bhaduka, K.; Lal, S.; Nalini, J.; Chintala, C.S.
    —Deep learning (DL) algorithms are currently the most effective methods for change detection (CD) from high-resolution multispectral (MS) remote-sensing (RS) images. Because a variety of satellites are able to provide a lot of data, it is now easy to find changes using efficient DL models. Current CD methods focus on simple structure and combining the features obtained by all the stages together rather than extracting multiscale features from a single stage since it may lead to information loss and an imbalance contribution of features at different stages. This in turn results in misclassification of small changed areas and poor edge and shape preservation of changed areas. This article introduces an enhanced RSCD network (ERSCDNet) for CD from bitemporal aerial and MS images. The proposed encoder–decoder-based ERSCDNet model uses an attention-based encoder and decoder block and a modified new spatial pyramid pooling block at each stage of the decoder part, which effectively utilize features at each encoder stages and prevent information loss. The learning, vision, and remote sensing CD (LEVIR-CD), Onera satellite change detection (OSCD), and Sun Yat-Sen University CD (SYSU-CD) datasets are used to evaluate the ERSCDNet model. The ERSCDNet gives better performance than all the models used in this article for comparison. It gives an F1 score, a Kappa coefficient, and a Jaccard index of (0.9306, 0.9282, 0.8703), (0.8945, 0.8887, 0.8091), and (0.7581, 0.6876, 0.6103) on OSCD, LEVIR-CD, and SYSU-CD datasets, respectively. © 2024 The Authors. This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License.
  • Item
    CAR-BRAINet: Sub-6 GHz aided spatial adaptive beam prediction with multi head attention for heterogeneous vehicular networks
    (Institute of Physics, 2025) Menon, A.G.; Krishnan, P.; Lal, S.
    Heterogeneous Vehicular Networks (HetVNets) play a crucial role by integrating different communication technologies, such as sub-6 GHz, mm-wave, and DSRC, to meet the diverse connectivity requirements of 5G/B5G vehicular networks. HetVNet helps address humongous user demands, but maintaining a steady connection in highly mobile, real-world conditions remains challenging. Though ample studies have been conducted on beam prediction models, a dedicated solution for HetVNets has been sparsely explored. Hence, developing a reliable beam prediction model, specifically for HetVNets, is necessary. This paper introduces a lightweight deep learning-based model termed ‘CAR-BRAINet’, which consists of convolutional neural networks with a powerful multi-head attention (MHA) mechanism. Existing literature on beam prediction is primarily studied under a limited, idealised vehicular scenario, often overlooking the real-time complexities and intricacies of vehicular networks. Therefore, this study aims to mimic the complexities of a real-time driving scenario by incorporating key factors, such as prominent MAC protocols (3GPP-C-V2X and IEEE 802.11BD), the effect of Doppler shifts under high velocity and varying distance, and SNR levels, into three high-quality dynamic data sets for urban, rural, and highway vehicular networks. CAR-BRAINet achieves a steady improvement of 11.6467% in spectral efficiency, with a 93.1638% lighter architecture compared to existing methods, resulting in a 94.7103% reduction in prediction time. Therefore, demonstrating a precise beam prediction across all vehicular scenarios, with minimal beam overhead. Thus, this study justifies the effectiveness of CAR-BRAINet in complex HetVNets, offering promising performance without relying on mobile users’ location, angle, and antenna dimensions, thereby reducing redundant sensor latency. © 2025 IOP Publishing Ltd. All rights, including for text and data mining, AI training, and similar technologies, are reserved.