Enhanced Swin Transformer and Edge Spatial Attention for Remote Sensing Image Semantic Segmentation

Fuxiang Liu, Zhiqiang Hu, Lei Li*, Hanlu Li, Xinxin Liu

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Combining convolutional neural networks (CNNs) and transformers is a crucial direction in remote sensing image semantic segmentation. However, due to differences in the spatial information focus and feature extraction methods, existing feature transfer and fusion strategies do not effectively integrate the advantages of both approaches. To address these issues, we propose a CNN-transformer hybrid network for precise remote sensing image semantic segmentation. We propose a novel Swin Transformer block to optimize feature extraction and enable the model to handle remote sensing images of arbitrary sizes. Additionally, we design an Edge Spatial Attention module to focus attention on local edge structures, effectively integrating global features and local details. This facilitates efficient information flow between the Transformer encoder and CNN decoder. Finally, a multi-scale convolutional decoder is employed to fully leverage both global information from the Transformer and local features from the CNN, leading to accurate segmentation results. Our network achieved state-of-the-art performance on the Vaihingen and Potsdam datasets, reaching mIoU and F1 scores of 67.37% and 79.82%, as well as 72.39% and 83.68%, respectively.

Original languageEnglish
Pages (from-to)1296-1300
Number of pages5
JournalIEEE Signal Processing Letters
Volume32
DOIs
Publication statusPublished - 2025

Keywords

  • Edge detection
  • Swin transformer
  • remote sensing image
  • semantic segmentation

Fingerprint

Dive into the research topics of 'Enhanced Swin Transformer and Edge Spatial Attention for Remote Sensing Image Semantic Segmentation'. Together they form a unique fingerprint.

Cite this