SA-UMamba: Spatial attention convolutional neural networks for medical image segmentation.
Authors
Affiliations (4)
Affiliations (4)
- School of Computer Science and Technology, Huaibei Normal University, Huaibei, Anhui, China.
- Huaibei Key Laboratory of Digital Multimedia Intelligent Information Processing, Huaibei, Anhui, China.
- College of Electronic and Information Engineering, Hebei University, Baoding, Hebei, China.
- Huaibei People's Hospital, Huaibei, Anhui, China.
Abstract
Medical image segmentation plays an important role in medical diagnosis and treatment. Most recent medical image segmentation methods are based on a convolutional neural network (CNN) or Transformer model. However, CNN-based methods are limited by locality, whereas Transformer-based methods are constrained by the quadratic complexity of attention computations. Alternatively, the state-space model-based Mamba architecture has garnered widespread attention owing to its linear computational complexity for global modeling. However, Mamba and its variants are still limited in their ability to extract local receptive field features. To address this limitation, we propose a novel residual spatial state-space (RSSS) block that enhances spatial feature extraction by integrating global and local representations. The RSSS block combines the Mamba module for capturing global dependencies with a receptive field attention convolution (RFAC) module to extract location-sensitive local patterns. Furthermore, we introduce a residual adjust strategy to dynamically fuse global and local information, improving spatial expressiveness. Based on the RSSS block, we design a U-shaped SA-UMamba segmentation framework that effectively captures multi-scale spatial context across different stages. Experiments conducted on the Synapse, ISIC17, ISIC18 and CVC-ClinicDB datasets validate the segmentation performance of our proposed SA-UMamba framework.