Back to all papers

AI-FLEET: Phase I-Multimodal Deep Learning Model for Phyllodes Tumor Classification.

February 6, 2026pubmed logopapers

Authors

Holt L,Chamberlain V,Shern T,Abir FF,Daly A,Anderman K,Gadd MA,Verdial FC,Kwait RM,Specht MC,Smith BL,Brattain L,Ozmen T

Affiliations (4)

  • Breast Section, Division of Gastrointestinal and Oncologic Surgery, MGH Center for Breast Cancer, Massachusetts General Hospital, Boston, MA, USA.
  • Department of Electrical and Computer Engineering, University of Central Florida, Orlando, FL, USA.
  • Department of Internal Medicine, University of Central Florida College of Medicine, Orlando, FL, USA.
  • Breast Section, Division of Gastrointestinal and Oncologic Surgery, MGH Center for Breast Cancer, Massachusetts General Hospital, Boston, MA, USA. [email protected].

Abstract

Fibroepithelial breast lesions, including fibroadenomas and phyllodes tumors (PTs), can be difficult to classify on needle biopsy. Misclassification may result in unnecessary excisions of benign fibroadenomas or delays and repeat operations for borderline/malignant PTs. Artificial Intelligence for Fibroepithelial Lesion Evaluation and Extrication Technology (AI-FLEET) is a multi-stage program designed to improve diagnostic accuracy and reduce inconclusive preoperative assessments by integrating radiologic, pathologic, and clinical data. In this first phase, we retrospectively analyzed patients with histologically confirmed PTs. Borderline and malignant PTs were grouped together owing to similarities in margin management and the limited number of cases. Models were trained to distinguish benign from borderline/malignant PTs using ultrasound images and clinical variables (age, body mass index (BMI), race/ethnicity, menopausal status, echogenicity, and tumor size). Multiple convolutional and attention-based encoders were evaluated using subject-stratified five-fold cross-validation. The cohort included 81 patients (65 benign, 16 borderline/malignant PTs) with 1638 ultrasound images. The multimodal ConvNeXt model achieved an accuracy of 0.91 (AUC 0.94), while the multimodal ResNet18 achieved an accuracy of 0.92 (AUC 0.94). Other multimodal architectures showed lower performance. Ultrasound-only and clinical-only models reached AUCs of 0.89 and 0.78, respectively. Saliency analyses identified intratumoral heterogeneity as an important predictive feature. Multimodal deep learning models combining ultrasound and clinical factors achieved high accuracy in differentiating benign from borderline/malignant PTs, demonstrating the feasibility of AI-assisted assessment of fibroepithelial lesions. Phase II will expand this work by incorporating histopathology and fibroadenoma cases to further enhance radiologic-pathologic integration.

Topics

Journal Article

Ready to Sharpen Your Edge?

Subscribe to join 11k+ peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.