Back to all papers

Multimodal brain tumor segmentation and classification based on optimized DeepLabV3 + and fused fire module with self-attention.

March 6, 2026pubmed logopapers

Authors

Ullah MS,Khan MA,Nam Y,Almujally NA,Alasiry A,Marzougui M,Gorriz JM,Hussain A

Affiliations (8)

  • Department of Computer Science, HITEC University, Taxila, Pakistan.
  • Center of AI, Prince Mohammad Bin Fahd University, Al-Khobar, Saudi Arabia. [email protected].
  • Department of Computer Science and Engineering, Soonchunhyang University, Asan, 31538, Republic of Korea. [email protected].
  • Department of Information Systems, College of Computer and Information Sciences, Princess Nourah Bint Abdulrahman University, P.O. Box 84428, Riyadh, 11671, Saudi Arabia.
  • College of Computer Science, King Khalid University, Abha, 61413, Saudi Arabia.
  • Dasci Institute, University of Granada, Granada, Spain.
  • School of Computing, Edinburgh Napier University, Edinburgh, UK.
  • Nuffield Department of Primary Care Health Sciences, University of Oxford, Oxford, UK.

Abstract

In this work, we propose a novel deep learning architecture for brain tumor segmentation and classification, SMDeepNet, which is based on an optimized DeepLabV3 +  + and a Fused Fire Module with Self-Attention. The segmentation framework comprises down- and up-sampling based on a DeepLabV3 + neural network and is optimized by dynamically initializing hyperparameters for training the backbone ResNet-50 architecture. During the down-sampling or encoder stage, the Atrous Spatial Pyramid Pooling (ASPP) module extracted features using convolutional layers with various filter sizes and dilations. These features are then passed to the up-sampling or decoder section for final segmentation. The classification framework comprises two sub-frameworks: a fire-residual bottleneck (Fire-RB) and a Hybrid Efficient Attention (Hybrid-EA). The Fire-RB framework comprises several parallel blocks: one side implements squeeze-and-expand behavior in the fire mechanism, and the other implements residual bottlenecks. The two parallel blocks are concatenated, and features are extracted from Fire-RB. The Hybrid-EA model is a custom variant of the pre-trained EfficientNetB0 model that incorporates a self-attention mechanism. The self-attention mechanism enhanced the functionality of the EfficientNetB0 model. Features from Fire-RB and Hybrid-EA are concatenated channel-wise, and the final modality classification is performed. The BraTS 2023 dataset is used in this work to evaluate the proposed methodologies. Segmentation results indicate that accuracy, Dice Score, and Intersection over Union (IOU) are 0.9871, 0.9420, and 0.8951, respectively. Modality classification results indicate an accuracy of 0.9920, which is improved over the recent state-of-the-art techniques.

Topics

Journal Article

Ready to Sharpen Your Edge?

Subscribe to join 11k+ peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.