Opto-Electronic Engineering, Volume. 51, Issue 12, 240237-1(2024)

Multi-scale feature enhanced Transformer network for efficient semantic segmentation

Yan Zhang, Chunming Ma, Shudong Liu, and Yemei Sun
Author Affiliations
  • College of Computer and Information Engineering, Tianjin Chengjian University, Tianjin 300380, China
  • show less

    To address the issues of insufficient utilization of multi-scale semantic information and high computational costs resulting from the generation of lengthy sequences in existing Transformer-based semantic segmentation networks, this paper proposes an efficient semantic segmentation backbone named MFE-Former, based on multi-scale feature enhancement. The network mainly includes the multi-scale pooling self-attention (MPSA) and the cross-spatial feed-forward network (CS-FFN). MPSA employs multi-scale pooling to downsample the feature map sequences, thereby reducing computational cost while efficiently extracting multi-scale contextual information, enhancing the Transformer’s capacity for multi-scale information modeling. CS-FFN replaces the traditional fully connected layers with simplified depth-wise convolution layers to reduce the parameters in the initial linear transformation of the feed-forward network and introduces a cross-spatial attention (CSA) to better capture different spaces interaction information, further enhancing the expressive power of the model. On the ADE20K, Cityscapes, and COCO-Stuff datasets, MFE-Former achieves mean intersection-over-union (mIoU) scores of 44.1%, 80.6%, and 38.0%, respectively. Compared to mainstream segmentation algorithms, MFE-Former demonstrates competitive segmentation accuracy at lower computational costs, effectively improving the utilization of multi-scale information and reducing computational burden.

    Keywords
    Tools

    Get Citation

    Copy Citation Text

    Yan Zhang, Chunming Ma, Shudong Liu, Yemei Sun. Multi-scale feature enhanced Transformer network for efficient semantic segmentation[J]. Opto-Electronic Engineering, 2024, 51(12): 240237-1

    Download Citation

    EndNote(RIS)BibTexPlain Text
    Save article for my favorites
    Paper Information

    Category: Article

    Received: Oct. 10, 2024

    Accepted: Nov. 19, 2024

    Published Online: Feb. 21, 2025

    The Author Email:

    DOI:10.12086/oee.2024.240237

    Topics