MANGA-YOLO: A Mamba-inspired YOLO model with group attention for breast mass detection in mammograms.
Authors
Affiliations (3)
Affiliations (3)
- School of Information Technology, Monash University, Malaysia Campus, Subang Jaya, 47500, Malaysia.
- School of Information Technology, Monash University, Malaysia Campus, Subang Jaya, 47500, Malaysia. Electronic address: [email protected].
- School of Electrical Engineering, International University, Ho Chi Minh City, 700000, Viet Nam; Vietnam National University, Ho Chi Minh City, 700000, Viet Nam.
Abstract
Breast cancer continues to be the primary cause of mortality from cancer in women globally, highlighting the urgent necessity for early and accurate identification. Although mammogram is the gold standard for screening, several challenges still exist, such as low contrast, varying sizes of masses, and irregular mass shapes. To address these limitations, we propose MANGA-YOLO, a novel learning model for breast mass detection inspired by the Mamba and transformer architecture. Our model integrates Mamba-Inspired Attention with Contextual Awareness (MACA) to improve feature extraction by capturing long-range dependencies and enhancing contextual understanding. Additionally, we introduce a Spatial and Channel Group Attention (SCGA) module, which combines convolutional spatial refinement with group self-attention. This enriches feature representation by strengthening boundary sensitivity and emphasizing global information through a series of attention modules, thus improving the detection of masses with diverse sizes. These improvements enable more precise detection of small, low-contrast, and irregularly shaped masses. The proposed approach is evaluated on three mammography datasets - INbreast, VinDr-Mammo, and CBIS-DDSM - achieving 0.881, 0.69, and 0.664 for mAP50 metric, respectively. The results show the outperformance of our method over existing state-of-the-art models. It suggests that MANGA-YOLO offers a more robust and efficient approach to automated breast mass detection, potentially improving early diagnosis and clinical decision making. The code is available at https://github.com/Kien-Trang/MANGA-YOLO.