TransSARNet: a deep learning framework for despeckling of SAR images
| dc.contributor.author | Kevala, V.D. | |
| dc.contributor.author | Sravya, N. | |
| dc.contributor.author | Lal, S. | |
| dc.contributor.author | Suresh, S. | |
| dc.contributor.author | Dell’Acqua, F. | |
| dc.date.accessioned | 2026-02-03T13:19:22Z | |
| dc.date.issued | 2025 | |
| dc.description.abstract | Synthetic Aperture Radar(SAR) images are extensively used for Earth observation because of their all-weather, day, and night imaging capabilities. However, speckle noise in SAR images significantly reduces their usability in a variety of applications. Deep learning models developed for SAR despeckling exhibit promising noise reduction capabilities. Bringing a balance between reducing graininess and preserving texture details is a challenging task. In addition, supervised training of a robust deep learning model requires noisy images that capture the SAR speckle dynamics and the corresponding speckle-free ground truth, which is generally not available. This study proposes the first hybrid CNN-Halo attention-based transformer model for SAR despeckling. CNN-based feature extraction modules provide multiscale and multidirectional and large-scale feature maps. A halo-attention transformer block is used in the skip connection. It aids in the better preservation of radiometric information in the despeckled SAR images. TransSARNet is trained in a supervised manner using a new synthetic SAR dataset, which is a combination of the Kylberg and UCMerced land-use datasets. This study also analyzed the effect of combining the Kylberg and UCMerced datasets on texture preservation in despeckled SAR images. The visual and qualitative metrics evaluated on Sentinel-1 Single Look Complex SAR data showed that the proposed TransSARNet approach outperformed the other models under consideration. TransSARNet achieves a harmonious balance between model complexity, despeckling ability, edge preservation, radiometric information preservation, and smoothing in homogeneous regions. © 2025 IOP Publishing Ltd. All rights, including for text and data mining, AI training, and similar technologies, are reserved. | |
| dc.identifier.citation | Engineering Research Express, 2025, 7, 3, pp. - | |
| dc.identifier.uri | https://doi.org/10.1088/2631-8695/ae037e | |
| dc.identifier.uri | https://idr.nitk.ac.in/handle/123456789/20068 | |
| dc.publisher | Institute of Physics | |
| dc.subject | Data mining | |
| dc.subject | Image denoising | |
| dc.subject | Learning systems | |
| dc.subject | Noise abatement | |
| dc.subject | Radar imaging | |
| dc.subject | Radiometry | |
| dc.subject | Speckle | |
| dc.subject | Supervised learning | |
| dc.subject | Synthetic aperture radar | |
| dc.subject | Textures | |
| dc.subject | De-speckling | |
| dc.subject | Deep learning | |
| dc.subject | Earth observations | |
| dc.subject | Imaging capabilities | |
| dc.subject | Learning frameworks | |
| dc.subject | Learning models | |
| dc.subject | Radiometrics | |
| dc.subject | Synthetic aperture radar despeckling | |
| dc.subject | Synthetic aperture radar images | |
| dc.subject | Transformer | |
| dc.title | TransSARNet: a deep learning framework for despeckling of SAR images |
