A novel approach for breast cancer detection using a Nesterov accelerated adam optimizer with an attention mechanism.
Authors
Affiliations (4)
Affiliations (4)
- Information Technology Department, Faculty of Computers and Artificial Intelligence, Damietta University, Damietta, 34517, Egypt. [email protected].
- Information Technology Department, Faculty of Computers and Artificial Intelligence, Damietta University, Damietta, 34517, Egypt.
- Department of Data Science, Faculty of Artifcial Intelligence, Kafrelsheikh University, Kafrelsheikh, 33511, Egypt.
- Faculty of Artificial Intelligence, Kafrelsheikh University, Kafrelsheikh, 33511, Egypt.
Abstract
Image-based automatic breast tumor detection has become a significant research focus, driven by recent advancements in machine learning (ML) algorithms. Traditional disease detection methods often involve manual feature extraction from images, a process requiring extensive expertise from specialists and pathologists. This labor-intensive approach is not only time-consuming but also impractical for widespread application. However, advancements in digital technologies and computer vision have enabled convolutional neural networks (CNNs) to learn features automatically, thereby overcoming these challenges. This paper presents a deep neural network model based on the MobileNet-V2 architecture, enhanced with a convolutional block attention mechanism for identifying tumor types in ultrasound images. The attention module improves the MobileNet-V2 model's performance by highlighting disease-affected areas within the images. The proposed model refines features extracted by MobileNet-V2 using the Nesterov-accelerated Adaptive Moment Estimation (Nadam) optimizer. This integration enhances convergence and stability, leading to improved classification accuracy. The proposed approach was evaluated on the BUSI ultrasound image dataset. Experimental results demonstrated strong performance, achieving an accuracy of 99.1%, sensitivity of 99.7%, specificity of 99.5%, precision of 97.7%, and an area under the curve (AUC) of 1.0 using an 80-20 data split. Additionally, under 10-fold cross-validation, the model achieved an accuracy of 98.7%, sensitivity of 99.1%, specificity of 98.3%, precision of 98.4%, F1-score of 98.04%, and an AUC of 0.99.