Laser & Optoelectronics Progress, Volume. 62, Issue 16, 1600002(2025)

Applications and Advancements of U-Net and Its Variants in Brain Tumor Image Segmentation

Nan Wang, Hua Wang, Dejian Wei, Liang Jiang, Peihong Han, and Hui Cao*
Author Affiliations
  • School of Medical Informational Engineering, Shandong University of Traditional Chinese Medicine, Jinan 250355, Shandong , China
  • show less
    Figures & Tables(14)
    Comparison of tumor images with and without noise. (a) Noise-free image; (b) noisy image
    Brain tumor images with different MRI modals. (a) T1 weighted modal; (b) T2 weighted modal; (c) T1C modal; (d) FLAIR modal; (e) expert segmentation result
    Standard network architecture and network architecture with residual connections[27]. (a) Standard network architecture; (b) network architecture with residual connections
    Comparison of receptive fields in CNN and Transformer models[47]
    DiffSwinTr structure diagram[49]
    MUNet model diagram[63]
    Simplified schematic diagram of adversarial U-Net structure[74]
    Performances comparison of different algorithms on BraTS 2019 validation set
    • Table 1. Commonly used brain tumor image datasets

      View table

      Table 1. Commonly used brain tumor image datasets

      DatasetMain applicationApplicable scenarioModalImage sizeSource
      BraTSBrain tumor segmentation, multimodal MRI dataTumor detection, segmentation model benchmarkingT1、T1-Gd、T2、FLAIR240 pixel ×240 pixel ×155 pixel (typical)https://www.med.upenn.edu/sbia/brats2020/
      TCIAGeneral medical imaging, cancer researchCancer diagnosis, image segmentation, model developmentMRI256 pixel ×256 pixel (variational)https://www.cancerimagingarchive.net/
      ISLESStroke lesion detection, MRI segmentationStroke detection, lesion localization, clinical studyMRI256 pixel ×256 pixel or 128 pixel ×128 pixelhttp://www.isles-challenge.org/
      RSNAClinical brain MRI data, disease diagnosisDisease diagnosis, brain structure analysisMRI512 pixel ×512 pixel (variational)https://www.rsna.org/en
      Lower-Grade GliomaLower-Grade Glioma segmentation, MRI dataLower-Grade Glioma research, tumor progression studyMRI240 pixel ×240 pixel ×155 pixel (typical)https://www.med.upenn.edu/sbia/lgg-challenge/
      CMB DatasetMicrobleed detection, MRI dataMicrobleed detection, small lesion analysisMRI256 pixel ×256 pixel (variational)https://pubmed.ncbi.nlm.nih.gov/
    • Table 2. Commonly used brain tumor image segmentation indices and their expressions

      View table

      Table 2. Commonly used brain tumor image segmentation indices and their expressions

      IndexFormula
      Dice coefficientEDice=2TTP2TTP+FFP+FFN
      SensitivityRSensitivity=TTPTTP+FFN
      SpecificityRSpecificity=TTNTTN+FFP
      PrecisionRPrecision=TTPTTP+FFP
    • Table 3. Comparison of the improved U-Net mechanisms

      View table

      Table 3. Comparison of the improved U-Net mechanisms

      MethodMechanismAdvantageLimitationAlgorithmFeature
      Network structure and feature optimizationMulti-scale feature extractionEffectively capture tumor features at different scales, improve segmentation accuracy for small tumors and boundary regionsHigh computational cost, complex model, which can burden training and inference processesRAPNet22Multi-scale, dilated convolutions, attention mechanism
      HA-RUnet23Attention mechanism, SE modules
      Ga-U-Net24Gabor convolutions, attention mechanisms
      U-shaped encoder-decoder model25Compact split attention, enhanced feature extraction
      Residual and skip connectionIncrease network depth and width while avoiding gradient vanishing problems, improve feature extraction and classification accuracyMay not fully solve gradient vanishing issues, limited ability to capture small tumor boundariesIRDNU-Net27Residual-Inception modules
      Residual learning U-Net29Residual learning, feature extraction
      dResU-Net30The training process is optimized and feature extraction is enhanced by jump connections between residual and convolutional blocks
      MMGAN31Residual learning, reduced parameters
      Lightweight designReduce model parameters, lower computational cost, and improve robustness and efficiencyMay lose detail in complex tumor areas, perform poorly on highly complex dataSEDNet36Hierarchical convolution, feature learning, robustness, efficiency, optimized architecture, segmentation, fewer parameters
      GA-UNet38Lightweight design, GhostV2 bottleneck, attention module
      Contextual information and attention mechanismContext information and attention mechanismImprove tumor localization and boundary recognition, reduce background interferenceLimited ability to capture complex boundaries and fine details, performance may be affected by background complexityMMS-Net39Triple attention modules, multi-modal MRI segmentation
      TDPC-Net403D attention, decoupled convolution units
      Dual attention U-Net43Dual attention mechanism, iterative feature aggregation
      3D U-Net with attention44Residual network, attention mechanisms, adaptive learning
      Transformer fusionStrong global context capture ability, improved segmentation accuracy, especially in multi-modal dataHigh computational and memory overhead, long training time, leading to lower processing efficiencyUNETR53Transformer encoder, global context modeling
      TransMVU55Transformer + U-Net, multi-view performance
      Swin-UNet58Swin-Transformer, global context modeling
      Spatial pyramid poolingExpand the receptive field and allow for multi-scale feature extraction, preserve fine detailsPerformance may degrade with very high-resolution images, less effective in very fine detailsAttention-UNet with ASPP61Attention mechanism, ASPP, multi-scale feature extraction, expanded receptive fields, preserved fine details, and improved segmentation accuracy
      Training strategy and performance improvementTraining strategy and performance improvementSignificantly enhance segmentation of small tumors and imbalanced data, improve boundary and overlap accuracyLess effective for large and complex tumors, may not handle large tumor structures wellWeighted loss + Dice loss62Generalized Dice loss, attention mechanism
      MUNet63mIoU loss, Dice loss, boundary loss, small tumor regions, overlap, similarity
      SBTC-Net64Transfer learning, segmentation and classification
    • Table 4. Comparison of pre-training methods for U-Net

      View table

      Table 4. Comparison of pre-training methods for U-Net

      MethodMechanismAdvantageLimitationAlgorithmFeature
      Multi-modal fusionUtilize complementary information from different modals (MRI, CT) for enhanced feature extractionImprove segmentation accuracy by integrating information from different modalsPotential risk of insufficient feature fusionACMINet68Cross-modal feature alignment, accurate tumor segmentation, volume brain tumor focus
      F2 Net69Effective feature fusion strategies that allow information from different modals to complement each other
      CKD-TransBTS70By integrating the radiologist's clinical experience with the model's structure and combining a two-branch hybrid encoder with the cross-attention mechanism
      SDPN71Lightweight dual-path network, multi-spectral attention, and tensor ring decomposition
      GAN and Image AugmentationCombine U-Net with GANs (cGAN) to generate realistic tumor images and improve segmentationEnhance segmentation precision and boundary details, improve model robustness to different tumor shapesStability issues with GAN parameters, high complexityAdversarial U-Net73Generative adversarial network integration, U-Net as generator, enhanced segmentation image quality
      Pix2Pix74Conditional GAN model, image-to-image translation, high-quality image generation, effective for paired datasets
      CycleGAN75Unpaired image-to-image translation, cycle-consistency loss, no need for paired datasets
      Self-supervised and semi-supervised learningUse self-supervised learning to pre-train models and semi-supervised learning with few labeled dataEnhance model performance with fewer labeled images, improve robustness and adaptabilitySelf-supervised learning still has limitations in complex segmentation tasksSSW-AN80Self-supervised learning, adjustable activation function, dynamic weight attention module, low/high-frequency channel separation
      Similarity-based algorithm81Semi-supervised learning, similarity constraints, effective with limited labeled data, improved generalization
      Real-time segmentation and clinical applicationCombine GAN, self-supervised learning, and efficient architectures to support fast inference and clinical applicationsIncrease segmentation speed while maintaining accuracy, suitable for real-time clinical useDependent on high-quality input data, limited application in low-quality data scenariosCascaded method82Two-stage approach, WT region detection, fine-grained sub-region segmentation
      TransU2-Net84Transformer module integration, lightweight design, efficient feature extraction
      EfficientNet-U-Net85EfficientNet integration, optimized computational efficiency, high segmentation accuracy
    • Table 5. Performance comparison of different types of U-Net in brain tumor image segmentation tasks

      View table

      Table 5. Performance comparison of different types of U-Net in brain tumor image segmentation tasks

      CategoryAdvantageLimitationMethodologyDatasetPrecision /%Others
      TCETWT

      U-Net+

      attention mechanism

      Dynamically highlight key features, improve local segmentation precision, excel in enhancing tumor regionsSensitive to parameter initialization, unstable training, increase model complexity and inference timeMMS-NetBraTS 201984.583.090.6
      TDPC-NetBraTS 201984.179.190.2
      DAMIA-U-NetBraTS 201971.571.380.5
      RF-ATT-U-NetBraTS 201886.079.086.0
      GCA-U-NetBraTS 202189.785.691.0Average Hausdorff distance: 9.12 mm
      MUNetBraTS 201881.581.590.1

      Hausdorff95: 6.152

      4.389 6.243

      U-Net+

      skip connection mechanism

      Preserve spatial information and improve gradient flow, enhance segmentation precision by combining low-level and high-level featuresIncrease computational complexity and may struggle with processing complex structuresmResU-NetBraTS 202192.989.792.8
      DSDU-NetPreserve low-level features, multi-scale patch extraction for tumor region separation
      ResNet with Attention Gates

      Combine ResNet with U-Net, attention gates improve

      tumor localization precision

      Maximized U-NetEnhance sensitivity to multi-scale information, improve feature expression ability

      U-Net+

      residual linkage mechanism

      Mitigate gradient vanishing, enhance deep feature representation, achieve more accurate segmentationHighly dependent on data generalization, increased model complexity and parametersdResU-Net

      Residual network as encoder to mitigate gradient vanishing;

      U-Net decoder improves segmentation performance

      RAPNetBraTS 201985.285.285.2

      U-Net+

      dense connection mechanism

      Enhance feature reuse and gradient flow, improve segmentation performance by allowing the network to efficiently learn from earlier layersIncrease model complexity and computational costImproved U-NetEnhance segmentation precision and robustness by adjusting network depth and channel count
      Data Augmentation Strategy with U-NetCombine effective data augmentation to handle data diversity in brain tumor segmentation

      U-Net+

      Transformer

      Capture long-range dependencies and global features, excel in complex tasks with superior performanceHigh computational complexity, require strong hardware support, unsuitable for real-time applicationsDiffSwinTr79.983.0785.38Average Hausdorff distance: 3.361 mm; 3.334 mm; 2.975 mm(95%)

      U-Net+

      GAN

      Generate high-quality synthetic data, alleviate data scarcity, excel in small-sample segmentation tasksUnstable training, heavily rely on parameter tuning, segmentation performance depend on generated data qualitySwin-UNet78.078.089.0
      Inception-UDet81.077.089.0
      ESGanBraTS 201382.076.089.0
      MMGanBraTS 201886.078.087.0

      U-Net+

      multimodal

      Combine multimodal information to enhance segmentation precisionRely on high-quality multimodal data, which can be difficult to acquire and alignF2 NetSignificant improvement in technology
      CKD-TransBTS90.288.593.3Sensitivity: 90.6%(TC) 90.0%(ET) 90.3%(WT)
      TransDoubleU-NetBraTS 2019/2020

      Promising results in

      MRI

      modal grading

      U-Net++

      EfficientNet

      LGGAverage Dice coefficient: 91.8%
      GA-UNetISIC-2018F1-score: 89.6%
      LGGF1-score: 89.6%
      U-Net+DSMBraTS 2019

      Sensitivity: 98.59%

      Specificity: 98.64%

      Precision: 98.64%

      Average Dice coefficient: 98.02%

      SAU-NetBraTS 2017

      Dice loss: 54.8%(TC)

      56.3%(ET)

      68.6%(WT)

      TransU2-NetBraTS 2021Average Dice coefficient: 88.17%
      HF-U-NetBraTS 2018

      Average Dice coefficient: 91.34%;

      Hausdorff distance: 3.74 mm

      U-Net-EfficientNetBraTS 201981.582.089.0
      ETumorNet-YOLOv8-U-NetTCIA

      Precision: 98.6%;

      Recall: 95.2%;

      F1-score: 96.3%;

      Specificity: 89.1%

    • Table 6. Performance comparison of U-Net with different training strategies on BraTS 2020

      View table

      Table 6. Performance comparison of U-Net with different training strategies on BraTS 2020

      CategoryAdvantageLimitationMethodologyPrecision /%Others
      TCETWT

      U-Net+

      attention mechanism

      Multi-scale feature fusion, enhanced boundary preservationHigh computational complexityHA-RUnet81.3078.7086.70

      Sensitivity: 83.0%(TC);

      88.0%(ET);

      93.0%(WT)

      Hierarchical feature extraction enhances tumor localizationRequire extensive hyperparameter tuningMUNet82.3083.5091.50

      Hausdorff95: 6.437

      2.421

      3.755

      U-Net+

      residual linkage mechanism

      Cross-resolution feature fusion enhances robustnessChallenging deployment on resource-limited systemsMCG-CR3D-U-Net89.9084.9091.70

      Hausdorff distance: 9.203 mm(TC)

      12.171 mm(ET)

      6.021 mm(WT)

      U-Net+

      GAN

      Generative augmentation mitigates data scarcityTraining instability due to GAN optimization challengesVit+GAN

      Precision: 97.65%;

      Sensitivity:97.70%

      Lightweight improvementDepthwise separable convolutions minimize parametersLower precision compared to heavy-weight modelsSEDNet90.2694.5193.08Hausdorff distance: NTC (Nearest Point Correspondence) is 0.704 mm, ED (Euclidean Distance) is 1.286 mm, ET(Error Threshold) is 0.776 mm

      U-Net+

      transfer learning

      Incorporate spatial and boundary attention for precise segmentationPotential overfitting on small datasetsSBTC-NetPrecision: 99.97%
    Tools

    Get Citation

    Copy Citation Text

    Nan Wang, Hua Wang, Dejian Wei, Liang Jiang, Peihong Han, Hui Cao. Applications and Advancements of U-Net and Its Variants in Brain Tumor Image Segmentation[J]. Laser & Optoelectronics Progress, 2025, 62(16): 1600002

    Download Citation

    EndNote(RIS)BibTexPlain Text
    Save article for my favorites
    Paper Information

    Category: Reviews

    Received: Dec. 6, 2024

    Accepted: Mar. 12, 2025

    Published Online: Aug. 8, 2025

    The Author Email: Hui Cao (caohui63@163.com)

    DOI:10.3788/LOP242385

    CSTR:32186.14.LOP242385

    Topics