Bias Corrected Twin Squeeze-and-Excitation Attention Enhanced UNet for brain tumor segmentation

https://doi.org/10.1016/j.bspc.2025.109390Get rights and content

Highlights

  • Twin SE module is embedded into UNet backbone for precise Brain Tumor segmentation.
  • Bias-Field correction is imbibed in preprocessing step to negate intensity variation.
  • Model is evaluated in two phases on MICCAI BraTS 2019, 2020 and 2023 datasets.
  • Comparable Dice and mean IoU scores achieved, attributed to twin SE and UNet fusion.

Abstract

Brain tumor segmentation is crucial in the context of deep learning-based medical image analysis, where accurate delineation aids in detecting abnormalities, treatment planning, and monitoring therapeutic outcomes for brain cancer. Despite the success of deep learning algorithms in medical image segmentation, challenges remain in capturing long-range dependencies, extracting relevant features, and addressing intensity variations across different imaging modalities. In this paper, we propose a novel deep learning architecture, Bias-Corrected Twin Squeeze-and-Excitation Attention Enhanced UNet (BC-TSEA-UNet), which integrates twin squeeze-and-excitation (SE) attention blocks into the UNet backbone to tackle these challenges. Unlike standard SE blocks, the twin SE configuration applies dual channel recalibration at multiple semantic levels, thereby capturing both shallow and deep contextual dependencies more effectively. The motivation behind incorporating SE blocks is to enhance feature recalibration, allowing the model to better focus on both local and global contextual information critical for accurate segmentation. Furthermore, a bias correction mechanism is employed during preprocessing to mitigate intensity non-uniformities in MRI scans, ensuring more consistent data representation across modalities. The proposed architecture is extensively evaluated on the BraTS 2019, BraTS 2020, and BraTS 2023 datasets, demonstrating average absolute improvements of 0.1112, 0.1339, and 0.1986 in Dice scores and 0.1601, 0.1396, and 0.1847 in mean intersection-over-union (IoU) on tumor subregions compared to the baseline UNet model, respectively. By emphasizing salient features and mitigating bias, BC-TSEA-UNet significantly improves feature representation, leading to more accurate and reliable tumor delineation across multiple datasets.

Introduction

Medical image segmentation plays a pivotal role in identifying anatomically significant regions within medical images, facilitating crucial clinical tasks such as disease diagnosis [1], disease progression monitoring [2], [3], and effective therapy planning [4]. Precise segmentation is particularly indispensable in detecting anomalies and tumors. Brain tumors, for instance, pose a significant health concern worldwide, accounting for 85% to 90% of all primary central nervous system (CNS) tumors [5]. They are also the leading cause of cancer deaths among children and adolescents younger than 20 years.
Brain tumors are classified as either primary or secondary. Primary tumors originate from brain cells, whereas secondary tumors metastasize from other parts of the body. Among primary tumors, gliomas are the most prevalent and originate from glial cells. They can be of high grade (HGG) or low grade (LGG), with high-grade gliomas being highly aggressive and requiring urgent medical attention. Delays in diagnosis and treatment of gliomas can lead to advanced-stage cancer and mortality. Accurate segmentation of them at an early stage is, therefore, required for effective diagnosis and treatment planning, given their aggressive nature and potential for metastasis. But, due to the heterogeneity in shapes and sizes of gliomas, identifying and delineating tumor boundaries is challenging. Gliomas exhibit notable heterogeneity not only in their morphological attributes, such as shape and size, but also in their histological composition and genetic profiles. The complex morphology of gliomas, ranging from compact masses to infiltrative patterns, further exacerbates the challenge of accurately capturing their extent within medical images. Further the anatomical variations also complicate glioma segmentation. Variations in brain anatomy, such as differences in shape, size, and spatial orientation, can obscure tumor boundaries and confound segmentation algorithms.
Traditionally, brain tumor segmentation has relied on manual annotation by medical experts, which is time-consuming and subjective process. Thresholding-based methods [6] which involve segmenting brain tumor regions based on intensity thresholds are relatively simple but struggle in the presence of noise and intensity variations. Region growing algorithms [7] that iteratively group pixels with similar properties are effective but sensitive to initialization and parameter selection. The widely used watershed transform algorithm [8] and active contour models [9] although accurate, suffers from over-segmentation. Atlas-based segmentation methods [10] register a pre-segmented atlas image to the target image and propagate the corresponding segmentation labels. They are sensitive to anatomical variability across different subjects. The traditional machine learning approaches including Support Vector Machines (SVMs) and Random Forests have revolutionized medical image segmentation [11], [12] but shows limitations in capturing complex spatial relationships particularly required in segmenting brain tumors across multi-modalities images and images with varying structures.
In the past decade, deep learning techniques have emerged as powerful tools for medical image analysis, showing great potential in automating the segmentation task. Particularly, Convolutional Neural Networks (CNNs) have become a potent tool for automated segmentation by leveraging their ability to learn discriminative features from large-scale medical imaging datasets [13], [14], [15]. However, despite these advancements, several key gaps remain in the current models. CNNs face significant challenges in capturing long-range dependencies and the spatial relationships between objects and their surroundings [16], which are crucial for accurately delineating tumor boundaries. This gap in the integration of contextual information arises from the reliance of traditional CNN-based architectures on local receptive fields. These models struggle to incorporate global context, making it difficult to accurately segment complex tumor subregions that often have irregular shapes and vary significantly in intensity, shape, and appearance across different patients. The UNet architecture, introduced by Ronneberger et al. [17], with its encoder–decoder structure and skip connections, has shown flexibility in handling these variations to some extent. Despite numerous adaptations aimed at enhancing its capabilities, UNet based architectures encounter difficulties in capturing intricate details and attending to the most relevant features, which is crucial for distinguishing subtle differences in tumor regions. Another significant gap lies in the insufficient handling of multimodal data, where current models fail to fully exploit the complementary information provided by different MRI modalities. The inability to effectively address these gaps often leads to errors, especially in ambiguous or noisy areas of the images, emphasizing the need for models that better integrate contextual information and leverage multimodal data for more robust and accurate brain tumor segmentation.
This motivates the need to integrate advanced attention mechanisms to boost the effectiveness of UNet-based segmentation models for accurate representation and discrimination of tumor regions while maintaining computational efficiency. In this article, a deep learning structure named Bias Corrected Twin Squeeze-and-Excitation Attention Enhanced UNet (BC-TSEA-UNet) is introduced. BC-TSEA-UNet merges the well-known UNet design with twin squeeze-and-excitation (SE) attention blocks, resulting in improved segmentation performance and heightened network discriminative power. SE Blocks selectively recalibrate feature maps by emphasizing informative features and suppressing less important ones, effectively capturing both local and global context. This approach enhances the model’s ability to integrate contextual information and improves segmentation performance. The primary contributions of the paper are:
  • A novel deep learning architecture, BC-TSEA-UNet, is proposed for brain tumor segmentation, that integrates twin SE blocks to enhance feature recalibration and improve the model’s ability to capture both local and global contextual information.
  • A comprehensive bias field correction mechanism is incorporated to mitigate intensity variations across brain images, ensuring consistent and reliable data for segmentation.
  • An extensive evaluation of the proposed BC-TSEA-UNet on the BraTS 2019, BraTS 2020, and BraTS 2023 datasets is performed that demonstrates its strong generalization capability across multiple brain tumor datasets.
The paper is organized as follows: Section 2 provides an overview of related work in medical image segmentation. Section 3 discusses the overall methodology of BC-TSEA-UNet. Section 4 outlines the experiments conducted, while Section 5 presents the result analysis performed on the BraTS datasets.

Access through your organization

Check access to the full text by signing in through your organization.

Access through your organization

Section snippets

Related work

Historically, brain tumor segmentation has primarily relied on traditional image processing techniques, such as thresholding, region-based approaches, edge detection techniques [18], the watershed segmentation [19], and active contour models [20]. These methods were instrumental in early tumor detection but exhibited limitations in terms of accuracy and precision. MRI data has been widely used in brain-related research. Robinson et al. [21] facilitated the use of fMRI data to identify brain

Methodology

The proposed architecture is the first extensive exploration of the integration of two Squeeze-and-Excitation attention (SE) blocks paired with UNet for medical image segmentation and also the first to adopt the idea of using twin equal-size SE modules in the UNet architecture to fully exploit the capability of the SE blocks. The integration of twin SE Blocks is strategically designed adaptation to enhance feature recalibration at both local and global levels. This novel configuration enables

Experiments

In this section, the experimental setup, dataset, evaluation metrics, and implementation details are discussed. The experiments are conducted in two phases on the MICCAI BraTS 2019, 2020, and 2023 datasets and are implemented using Python language and its libraries.

Results and discussion

This section presents the results of the extensive evaluation of BC-TSEA-UNet conducted on the BraTS dataset. Both the quantitative and qualitative results of the experiments are provided in the subsections below along with a detailed discussion of the outcomes.

Conclusion

In this work, the Bias Corrected Twin Squeeze-and-Excitation Attention Enhanced U-Net (BC-TSEA-UNet) was created for medical image segmentation, focusing on brain tumor segmentation using the BraTS dataset. The key innovation includes the integration of twin SE modules in the U-Net structure, enhancing feature adjustment and context modeling across multiple scales. This method extracts informative features while retaining essential spatial details for precise segmentation. Additionally, bias

CRediT authorship contribution statement

Abhishek Jadhav: Writing – review & editing, Writing – original draft, Methodology, Conceptualization. Akhtar Rasool: Validation, Supervision. Manasi Gyanchandani: Validation, Supervision.

Declaration of competing interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References (65)

  • Y. Liu et al.

    TransUNetplus: Redesigning the skip connection to enhance features in medical image segmentation

    Knowl.-Based Syst.

    (2022)
  • S.U.H. Dar et al.

    Parallel-stream fusion of scan-specific and scan-general priors for learning deep MRI reconstruction in low-data regimes

    Comput. Biol. Med.

    (2023)
  • A. Güngör et al.

    Adaptive diffusion priors for accelerated MRI reconstruction

    Med. Image Anal.

    (2023)
  • Y. Zhang et al.

    TAU: Transferable attention U-net for optic disc and cup segmentation

    Knowl.-Based Syst.

    (2021)
  • Q. Zhou et al.

    MOLS-net: Multi-organ and lesion segmentation network based on sequence feature pyramid and attention mechanism for aortic dissection diagnosis

    Knowl.-Based Syst.

    (2022)
  • L. Lu et al.

    Deformable attention-oriented feature pyramid network for semantic segmentation

    Knowl.-Based Syst.

    (2022)
  • C. Gao et al.

    Multiscale fused network with additive channel–spatial attention for image segmentation

    Knowl.-Based Syst.

    (2021)
  • D. Hao et al.

    Sequential vessel segmentation via deep channel attention network

    Neural Netw.

    (2020)
  • J. Ashburner et al.

    Unified segmentation

    Neuroimage

    (2005)
  • Y. Chang et al.

    Dpafnet: A residual dual-path attention-fusion convolutional neural network for multimodal brain tumor segmentation

    Biomed. Signal Process. Control.

    (2023)
  • D. Valenkova et al.

    A fuzzy rank-based ensemble of CNN models for MRI segmentation

    Biomed. Signal Process. Control.

    (2025)
  • I. Aboussaleh et al.

    3DUV-netr+: A 3D hybrid semantic architecture using transformers for brain tumor segmentation with MultiModal MR images

    Results Eng.

    (2024)
  • O. Bernard et al.

    Deep learning techniques for automatic MRI cardiac multi-structures segmentation and diagnosis: Is the problem solved?

    IEEE Trans. Med. Imaging

    (2018)
  • M. Zambrano-Vizuete et al.

    Segmentation of medical image using novel dilated ghost deep learning model

    Comput. Intell. Neurosci.

    (2022)
  • M.E. Tschuchnig et al.

    Anomaly detection in medical imaging - a mini review

  • R.L. Siegel et al.

    Cancer statistics, 2023

    CA: Cancer J. Clin.

    (2023)
  • Y.X. Dong

    Review of otsu segmentation algorithm

  • E.S. Biratu et al.

    Enhanced region growing for brain tumor MR image segmentation

    J. Imaging

    (2021)
  • M.A. Khan et al.

    Brain tumor detection and classification: A framework of marker-based watershed algorithm and multilevel priority features selection

    Microsc. Res. Tech.

    (2019)
  • S. Csaholczi et al.

    Automatic segmentation of brain tumor parts from MRI data using a random forest classifier

  • C. Zhou et al.

    One-pass multi-task convolutional neural networks for efficient brain tumor segmentation

  • O. Ronneberger et al.

    U-net: Convolutional networks for biomedical image segmentation

  • Cited by (0)

    View full text