Sort by:
Page 5 of 30293 results

Deep learning mammography-based breast cancer risk model, its serial change, and breast cancer mortality.

Shin S, Chang Y, Ryu S

pubmed logopapersSep 3 2025
Although numerous breast cancer risk prediction models have been developed to categorize individuals by risk, a substantial gap persists in evaluating how well these models predict actual mortality outcomes. This study aimed to investigate the association between Mirai, a deep learning model for risk prediction based on mammography, and breast cancer-specific mortality in a large cohort of Korean women. This retrospective cohort study examined 124,653 cancer-free women aged ≥ 34 years who underwent mammography screening between 2009-2020. Participants were stratified into tertiles by Mirai risk scores and categorized into four groups based on risk changes over time. Cox proportional hazards regression models were used to evaluate the associations of both baseline Mirai scores and temporal risk changes with breast cancer-specific mortality. Over 1,075,177 person-years of follow-up, 31 breast cancer-related deaths occurred. The highest Mirai risk tertile showed significantly higher breast cancer-specific mortality than the lowest tertile (hazard ratio [HR], 5.34; 95% confidence interval [CI] 1.17-24.39; p for trend = 0.020). Temporal Mirai score changes were associated with mortality risk: those remaining in the high-risk (HR, 5.92; 95% CI 1.43-24.49) or moving from low to high risk (HR, 5.57; 95% CI 1.31-23.63) had higher mortality rates than those staying in low-risk. The Mirai model, developed to predict breast cancer incidence, was significantly associated with breast cancer-specific mortality. Changes in Mirai risk scores over time were also linked to breast cancer-specific mortality, supporting AI-based risk models in guiding risk-stratified screening and prevention of breast cancer-related deaths.

SynBT: High-quality Tumor Synthesis for Breast Tumor Segmentation by 3D Diffusion Model

Hongxu Yang, Edina Timko, Levente Lippenszky, Vanda Czipczer, Lehel Ferenczi

arxiv logopreprintSep 3 2025
Synthetic tumors in medical images offer controllable characteristics that facilitate the training of machine learning models, leading to an improved segmentation performance. However, the existing methods of tumor synthesis yield suboptimal performances when tumor occupies a large spatial volume, such as breast tumor segmentation in MRI with a large field-of-view (FOV), while commonly used tumor generation methods are based on small patches. In this paper, we propose a 3D medical diffusion model, called SynBT, to generate high-quality breast tumor (BT) in contrast-enhanced MRI images. The proposed model consists of a patch-to-volume autoencoder, which is able to compress the high-resolution MRIs into compact latent space, while preserving the resolution of volumes with large FOV. Using the obtained latent space feature vector, a mask-conditioned diffusion model is used to synthesize breast tumors within selected regions of breast tissue, resulting in realistic tumor appearances. We evaluated the proposed method for a tumor segmentation task, which demonstrated the proposed high-quality tumor synthesis method can facilitate the common segmentation models with performance improvement of 2-3% Dice Score on a large public dataset, and therefore provides benefits for tumor segmentation in MRI images.

Using Explainable AI to Characterize Features in the Mirai Mammographic Breast Cancer Risk Prediction Model.

Wang YK, Klanecek Z, Wagner T, Cockmartin L, Marshall N, Studen A, Jeraj R, Bosmans H

pubmed logopapersSep 3 2025
<i>"Just Accepted" papers have undergone full peer review and have been accepted for publication in <i>Radiology: Artificial Intelligence</i>. This article will undergo copyediting, layout, and proof review before it is published in its final version. Please note that during production of the final copyedited article, errors may be discovered which could affect the content.</i> Purpose To evaluate whether features extracted by Mirai can be aligned with mammographic observations, and contribute meaningfully to the prediction. Materials and Methods This retrospective study examined the correlation of 512 Mirai features with mammographic observations in terms of receptive field and anatomic location. A total of 29,374 screening examinations with mammograms (10,415 women, mean age at examination 60 [SD: 11] years) from the EMBED Dataset (2013-2020) were used to evaluate feature importance using a feature-centric explainable AI pipeline. Risk prediction was evaluated using only calcification features (CalcMirai) or mass features (MassMirai) against Mirai. Performance was assessed in screening and screen-negative (time-to-cancer > 6 months) populations using the area under the receiver operating characteristic curve (AUC). Results Eighteen calcification features and 18 mass features were selected for CalcMirai and MassMirai, respectively. Both CalcMirai and MassMirai had lower performance than Mirai in lesion detection (screening population, 1-year AUC: Mirai, 0.81 [95% CI: 0.78, 0.84], CalcMirai, 0.76 [95% CI: 0.73, 0.80]; MassMirai, 0.74 [95% CI: 0.71, 0.78]; <i>P</i> values < 0.001). In risk prediction, there was no evidence of a difference in performance between CalcMirai and Mirai (screen-negative population, 5-year AUC: Mirai, 0.66 [95% CI: 0.63, 0.69], CalcMirai, 0.66 [95% CI: 0.64, 0.69]; <i>P</i> value: 0.71); however, MassMirai achieved lower performance than Mirai (AUC, 0.57 [95% CI: 0.54, 0.60]; <i>P</i> value < .001). Radiologist review of calcification features confirmed Mirai's use of benign calcification in risk prediction. Conclusion The explainable AI pipeline demonstrated that Mirai implicitly learned to identify mammographic lesion features, particularly calcifications, for lesion detection and risk prediction. ©RSNA, 2025.

Mammographic density assessed using deep learning in women at high risk of developing breast cancer: the effect of weight change on density.

Squires S, Harvie M, Howell A, Evans DG, Astley SM

pubmed logopapersSep 3 2025
High mammographic density (MD) and excess weight are both associated with increased risk of breast cancer. Classically defined percentage density measures tend to increase with reduced weight due to disproportionate loss of breast fat, however the effect of weight loss on artificial intelligence-based density scores is unknown. We investigated an artificial intelligence-based density method, reporting density changes in 46 women enrolled in a weight-loss study in a family history breast cancer clinic, using a volumetric density method as a comparison.&#xD;&#xD;Methods: We analysed data from women who had weight recorded and mammograms taken at the start and end of the 12-month weight intervention study. MD was assessed at both time points using a deep learning model trained on expert estimates of percent density called pVAS, and the volumetric density software VolparaTM.&#xD;&#xD;Results: Mean (standard deviation) weight of participants at the start and end of the study was 86.0 (12.2) and 82.5 (13.8) respectively; mean (standard deviation) pVAS scores were 35.8 (13.0) and 36.3 (12.4), and Volpara volumetric percent density scores were 7.05 (4.4) and 7.6 (4.4).The Spearman rank correlation between reduction in weight and change in density was 0.17 (-0.13 to 0.43, p=0.27) for pVAS and 0.59 (0.36 to 0.75, p<0.001) for Volpara volumetric percent density.&#xD;&#xD;Conclusion: pVAS percentage density measurements were not significantly affected by change in weight. Percent density measured with Volpara increased as weight decreased, driven by changes in fat volume.&#xD.

Commercial Artificial Intelligence Versus Radiologists: NPV and Recall Rate in Large Population-Based Digital Mammography and Tomosynthesis Screening Mammography Cohorts.

Chen IE, Joines M, Capiro N, Dawar R, Sears C, Sayre J, Chalfant J, Fischer C, Hoyt AC, Hsu W, Milch HS

pubmed logopapersSep 3 2025
<b>Background:</b> By reliably classifying screening mammograms as negative, artificial intelligence (AI) could minimize radiologists' time spent reviewing high volumes of normal examinations and help prioritize examinations with high likelihood of malignancy. <b>Objective:</b> To compare performance of AI, classified as positive at different thresholds, with that of radiologists, focusing on NPV and recall rates, in large population-based digital mammography (DM) and digital breast tomosynthesis (DBT) screening cohorts. <b>Methods:</b> This retrospective single-institution study included women enrolled in the observational population-based Athena Breast Health Network. Stratified random sampling was used to identify cohorts of DM and DBT screening examinations performed from January 2010 through December 2019. Radiologists' interpretations were extracted from clinical reports. A commercial AI system classified examinations as low, intermediate, or elevated risk. Breast cancer diagnoses within 1 year after screening examinations were identified from a state cancer registry. AI and radiologist performance were compared. <b>Results:</b> The DM cohort included 26,693 examinations in 20,409 women (mean age, 58.1 years). AI classified 58.2%, 27.7%, and 14.0% of examinations as low, intermediate, and elevated risk, respectively. Sensitivity, specificity, recall rate and NPV for radiologists were 88.6%, 93.3%, 7.2%, and 99.9%; for AI (defining positive as elevated risk), 74.4%, 86.3%, 14.0%, and 99.8%; and for AI (defining positive as intermediate/elevated risk), 94.0%, 58.6%, 41.8%, and 99.9%. The DBT cohort included 4824 examinations in 4379 women (mean age, 61.3 years). AI classified 68.1%, 19.8%, and 12.1% of examinations as low, intermediate, and elevated risk, respectively. Sensitivity, specificity, recall rate, and NPV for radiologists were 83.8%, 93.7%, 6.9%, and 99.9%; for AI (defining positive results as elevated risk), 78.4%, 88.4%, 12.1%, and 99.8%; and for AI (defining positive results as intermediate/elevated risk), 89.2%, 68.5%, 31.9%, and 99.8%. <b>Conclusion:</b> In large DM and DBT cohorts, AI at either diagnostic threshold achieved high NPV but had higher recall rates than radiologists. Defining positive AI results to include intermediate-risk examinations, versus only elevated-risk examinations, detected additional cancers but yielded markedly increased recall rates. <b>Clinical Impact:</b> The findings support AI's potential to aid radiologists' workflow efficiency. Yet, strategies are needed to address frequent false-positive results, particularly in the intermediate-risk category.

Toward a robust lesion detection model in breast DCE-MRI: adapting foundation models to high-risk women

Gabriel A. B. do Nascimento, Vincent Dong, Guilherme J. Cavalcante, Alex Nguyen, Thaís G. do Rêgo, Yuri Malheiros, Telmo M. Silva Filho, Carla R. Zeballos Torrez, James C. Gee, Anne Marie McCarthy, Andrew D. A. Maidment, Bruno Barufaldi

arxiv logopreprintSep 2 2025
Accurate breast MRI lesion detection is critical for early cancer diagnosis, especially in high-risk populations. We present a classification pipeline that adapts a pretrained foundation model, the Medical Slice Transformer (MST), for breast lesion classification using dynamic contrast-enhanced MRI (DCE-MRI). Leveraging DINOv2-based self-supervised pretraining, MST generates robust per-slice feature embeddings, which are then used to train a Kolmogorov--Arnold Network (KAN) classifier. The KAN provides a flexible and interpretable alternative to conventional convolutional networks by enabling localized nonlinear transformations via adaptive B-spline activations. This enhances the model's ability to differentiate benign from malignant lesions in imbalanced and heterogeneous clinical datasets. Experimental results demonstrate that the MST+KAN pipeline outperforms the baseline MST classifier, achieving AUC = 0.80 \pm 0.02 while preserving interpretability through attention-based heatmaps. Our findings highlight the effectiveness of combining foundation model embeddings with advanced classification strategies for building robust and generalizable breast MRI analysis tools.

Prediction of lymphovascular invasion in invasive breast cancer via intratumoral and peritumoral multiparametric magnetic resonance imaging machine learning-based radiomics with Shapley additive explanations interpretability analysis.

Chen S, Zhong Z, Chen Y, Tang W, Fan Y, Sui Y, Hu W, Pan L, Liu S, Kong Q, Guo Y, Liu W

pubmed logopapersSep 1 2025
The use of multiparametric magnetic resonance imaging (MRI) in predicting lymphovascular invasion (LVI) in breast cancer has been well-documented in the literature. However, the majority of the related studies have primarily focused on intratumoral characteristics, overlooking the potential contribution of peritumoral features. The aim of this study was to evaluate the effectiveness of multiparametric MRI in predicting LVI by analyzing both intratumoral and peritumoral radiomics features and to assess the added value of incorporating both regions in LVI prediction. A total of 366 patients underwent preoperative breast MRI from two centers and were divided into training (n=208), validation (n=70), and test (n=88) sets. Imaging features were extracted from intratumoral and peritumoral T2-weighted imaging, diffusion-weighted imaging, and dynamic contrast-enhanced MRI. Five models were developed for predicting LVI status based on logistic regression: the tumor area (TA) model, peritumoral area (PA) model, tumor-plus-peritumoral area (TPA) model, clinical model, and combined model. The combined model was created incorporating the highest radiomics score and clinical factors. Predictive efficacy was evaluated via the receiver operating characteristic (ROC) curve and area under the curve (AUC). The Shapley additive explanation (SHAP) method was used to rank the features and explain the final model. The performance of the TPA model was superior to that of the TA and PA models. A combined model was further developed via multivariable logistic regression, with the TPA radiomics score (radscore), MRI-assessed axillary lymph node (ALN) status, and peritumoral edema (PE) being incorporated. The combined model demonstrated good calibration and discrimination performance across the training, validation, and test datasets, with AUCs of 0.888 [95% confidence interval (CI): 0.841-0.934], 0.856 (95% CI: 0.769-0.943), and 0.853 (95% CI: 0.760-0.946), respectively. Furthermore, we conducted SHAP analysis to evaluate the contributions of TPA radscore, MRI-ALN status, and PE in LVI status prediction. The combined model, incorporating clinical factors and intratumoral and peritumoral radscore, effectively predicts LVI and may potentially aid in tailored treatment planning.

Ultrasound-based detection and malignancy prediction of breast lesions eligible for biopsy: A multi-center clinical-scenario study using nomograms, large language models, and radiologist evaluation

Ali Abbasian Ardakani, Afshin Mohammadi, Taha Yusuf Kuzan, Beyza Nur Kuzan, Hamid Khorshidi, Ashkan Ghorbani, Alisa Mohebbi, Fariborz Faeghi, Sepideh Hatamikia, U Rajendra Acharya

arxiv logopreprintAug 31 2025
To develop and externally validate integrated ultrasound nomograms combining BIRADS features and quantitative morphometric characteristics, and to compare their performance with expert radiologists and state of the art large language models in biopsy recommendation and malignancy prediction for breast lesions. In this retrospective multicenter, multinational study, 1747 women with pathologically confirmed breast lesions underwent ultrasound across three centers in Iran and Turkey. A total of 10 BIRADS and 26 morphological features were extracted from each lesion. A BIRADS, morphometric, and fused nomogram integrating both feature sets was constructed via logistic regression. Three radiologists (one senior, two general) and two ChatGPT variants independently interpreted deidentified breast lesion images. Diagnostic performance for biopsy recommendation (BIRADS 4,5) and malignancy prediction was assessed in internal and two external validation cohorts. In pooled analysis, the fused nomogram achieved the highest accuracy for biopsy recommendation (83.0%) and malignancy prediction (83.8%), outperforming the morphometric nomogram, three radiologists and both ChatGPT models. Its AUCs were 0.901 and 0.853 for the two tasks, respectively. In addition, the performance of the BIRADS nomogram was significantly higher than the morphometric nomogram, three radiologists and both ChatGPT models for biopsy recommendation and malignancy prediction. External validation confirmed the robust generalizability across different ultrasound platforms and populations. An integrated BIRADS morphometric nomogram consistently outperforms standalone models, LLMs, and radiologists in guiding biopsy decisions and predicting malignancy. These interpretable, externally validated tools have the potential to reduce unnecessary biopsies and enhance personalized decision making in breast imaging.

Fusion model integrating multi-sequence MRI radiomics and habitat imaging for predicting pathological complete response in breast cancer treated with neoadjuvant therapy.

Xu S, Ying Y, Hu Q, Li X, Li Y, Xiong H, Chen Y, Ye Q, Li X, Liu Y, Ai T, Du Y

pubmed logopapersAug 29 2025
This study aimed to develop a predictive model integrating multi-sequence MRI radiomics, deep learning features, and habitat imaging to forecast pathological complete response (pCR) in breast cancer patients undergoing neoadjuvant therapy (NAT). A retrospective analysis included 203 breast cancer patients treated with NAT from May 2018 to January 2023. Patients were divided into training (n = 162) and test (n = 41) sets. Radiomics features were extracted from intratumoral and peritumoral regions in multi-sequence MRI (T2WI, DWI, and DCE-MRI) datasets. Habitat imaging was employed to analyze tumor subregions, characterizing heterogeneity within the tumor. We constructed and validated machine learning models, including a fusion model integrating all features, using Receiver Operating Characteristic (ROC) and Precision-Recall (PR) curves, decision curve analysis (DCA), and confusion matrices. Shapley Additive Explanations (SHAP) and Local Interpretable Model-agnostic Explanations (LIME) analyses were performed for model interpretability. The fusion model achieved superior predictive performance compared to single-region models, with AUCs of 0.913 (95% CI: 0.770-1.000) in the test set. PR curve analysis showed improved precision-recall balance, while DCA indicated higher clinical benefit. Confusion matrix analysis confirmed the model's classification accuracy. SHAP revealed DCE_LLL_DependenceUniformity as the most critical feature for predicting pCR and PC72 for non-pCR. LIME provided patient-specific insights into feature contributions. Integrating multi-dimensional MRI features with habitat imaging enhances pCR prediction in breast cancer. The fusion model offers a robust, non-invasive tool for guiding individualized treatment strategies while providing transparent interpretability through SHAP and LIME analyses.

Multimodal Deep Learning for Phyllodes Tumor Classification from Ultrasound and Clinical Data

Farhan Fuad Abir, Abigail Elliott Daly, Kyle Anderman, Tolga Ozmen, Laura J. Brattain

arxiv logopreprintAug 29 2025
Phyllodes tumors (PTs) are rare fibroepithelial breast lesions that are difficult to classify preoperatively due to their radiological similarity to benign fibroadenomas. This often leads to unnecessary surgical excisions. To address this, we propose a multimodal deep learning framework that integrates breast ultrasound (BUS) images with structured clinical data to improve diagnostic accuracy. We developed a dual-branch neural network that extracts and fuses features from ultrasound images and patient metadata from 81 subjects with confirmed PTs. Class-aware sampling and subject-stratified 5-fold cross-validation were applied to prevent class imbalance and data leakage. The results show that our proposed multimodal method outperforms unimodal baselines in classifying benign versus borderline/malignant PTs. Among six image encoders, ConvNeXt and ResNet18 achieved the best performance in the multimodal setting, with AUC-ROC scores of 0.9427 and 0.9349, and F1-scores of 0.6720 and 0.7294, respectively. This study demonstrates the potential of multimodal AI to serve as a non-invasive diagnostic tool, reducing unnecessary biopsies and improving clinical decision-making in breast tumor management.
Page 5 of 30293 results
Show
per page

Ready to Sharpen Your Edge?

Join hundreds of your peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.