Sort by:
Page 14 of 2332330 results

The identification and severity staging of chronic obstructive pulmonary disease using quantitative CT parameters, radiomics features, and deep learning features.

Feng S, Zhang W, Zhang R, Yang Y, Wang F, Miao C, Chen Z, Yang K, Yao Q, Liang Q, Zhao H, Chen Y, Liang C, Liang X, Chen R, Liang Z

pubmed logopapersSep 25 2025
To evaluate the value of quantitative CT (QCT) parameters, radiomics features, and deep learning (DL) features based on inspiratory and expiratory CT for the identification and severity staging of chronic obstructive pulmonary disease (COPD). This retrospective analysis included 223 COPD patients and 59 healthy controls from the Guangzhou cohort. We stratified the participants into a training cohort and a testing cohort (7:3) and extracted DL features based on VGG-16 method, radiomics features based on pyradiomics package, and QCT parameters based on NeuLungCARE software. The Logistic regression method was employed to construct models for the identification and severity staging of COPD. The Shenzhen cohort was used as the external validation cohort to assess the generalizability of the models. In the COPD identification models, Model 5-B1 (the QCT combined with DL model in biphasic CT) showed the best predictive performance with AUC of 0.920, and 0.897 in testing cohort and external validation cohort, respectively. In the COPD severity staging models, the predictive performance of Model 4-B2 (the model combining QCT with radiomics features in biphasic CT) and Model 5-B2 (the model combining QCT with DL features in biphasic CT was superior to that of the other models. This biphasic CT-based multi-modal approach integrating QCT, radiomics, or DL features offers a clinically valuable tool for COPD identification and severity staging.

Proof-of-concept comparison of an artificial intelligence-based bone age assessment tool with Greulich-Pyle and Tanner-Whitehouse version 2 methods in a pediatric cohort.

Marinelli L, Lo Mastro A, Grassi F, Berritto D, Russo A, Patanè V, Festa A, Grassi E, Grandone A, Nasto LA, Pola E, Reginelli A

pubmed logopapersSep 25 2025
Bone age assessment is essential in evaluating pediatric growth disorders. Artificial intelligence (AI) systems offer potential improvements in accuracy and reproducibility compared to traditional methods. To compare the performance of a commercially available artificial intelligence-based software (BoneView BoneAge, Gleamer, Paris, France) against two human-assessed methods-the Greulich-Pyle (GP) atlas and Tanner-Whitehouse version 2 (TW2)-in a pediatric population. This proof-of-concept study included 203 pediatric patients (mean age, 9.0 years; range, 2.0-17.0 years) who underwent hand and wrist radiographs for suspected endocrine or growth-related conditions. After excluding technically inadequate images, 157 cases were analyzed using AI and GP-assessed methods. A subset of 35 patients was also evaluated using the TW2 method by a pediatric endocrinologist. Performance was measured using mean absolute error (MAE), root mean square error (RMSE), bias, and Pearson's correlation coefficient, using chronological age as reference. The AI model achieved a MAE of 1.38 years, comparable to the radiologist's GP-based estimate (MAE, 1.30 years), and superior to TW2 (MAE, 2.86 years). RMSE values were 1.75 years, 1.80 years, and 3.88 years, respectively. AI showed minimal bias (-0.05 years), while TW2-based assessments systematically underestimated bone age (bias, -2.63 years). Strong correlations with chronological age were observed for AI (r=0.857) and GP (r=0.894), but not for TW2 (r=0.490). BoneView demonstrated comparable accuracy to radiologist-assessed GP method and outperformed TW2 assessments in this cohort. AI-based systems may enhance consistency in pediatric bone age estimation but require careful validation, especially in ethnically diverse populations.

Interpreting Convolutional Neural Network Activation Maps with Hand-crafted Radiomics Features on Progression of Pediatric Craniopharyngioma after Irradiation Therapy

Wenjun Yang, Chuang Wang, Tina Davis, Jinsoo Uh, Chia-Ho Hua, Thomas E. Merchant

arxiv logopreprintSep 25 2025
Purpose: Convolutional neural networks (CNNs) are promising in predicting treatment outcome for pediatric craniopharyngioma while the decision mechanisms are difficult to interpret. We compared the activation maps of CNN with hand crafted radiomics features of a densely connected artificial neural network (ANN) to correlate with clinical decisions. Methods: A cohort of 100 pediatric craniopharyngioma patients were included. Binary tumor progression was classified by an ANN and CNN with input of T1w, T2w, and FLAIR MRI. Hand-crafted radiomic features were calculated from the MRI using the LifeX software and key features were selected by Group lasso regularization, comparing to the activation maps of CNN. We evaluated the radiomics models by accuracy, area under receiver operational curve (AUC), and confusion matrices. Results: The average accuracy of T1w, T2w, and FLAIR MRI was 0.85, 0.92, and 0.86 (ANOVA, F = 1.96, P = 0.18) with ANN; 0.83, 0.81, and 0.70 (ANOVA, F = 10.11, P = 0.003) with CNN. The average AUC of ANN was 0.91, 0.97, and 0.90; 0.86, 0.88, and 0.75 of CNN for the 3 MRI, respectively. The activation maps were correlated with tumor shape, min and max intensity, and texture features. Conclusions: The tumor progression for pediatric patients with craniopharyngioma achieved promising accuracy with ANN and CNN model. The activation maps extracted from different levels were interpreted with hand-crafted key features of ANN.

Automated and Interpretable Survival Analysis from Multimodal Data

Mafalda Malafaia, Peter A. N. Bosman, Coen Rasch, Tanja Alderliesten

arxiv logopreprintSep 25 2025
Accurate and interpretable survival analysis remains a core challenge in oncology. With growing multimodal data and the clinical need for transparent models to support validation and trust, this challenge increases in complexity. We propose an interpretable multimodal AI framework to automate survival analysis by integrating clinical variables and computed tomography imaging. Our MultiFIX-based framework uses deep learning to infer survival-relevant features that are further explained: imaging features are interpreted via Grad-CAM, while clinical variables are modeled as symbolic expressions through genetic programming. Risk estimation employs a transparent Cox regression, enabling stratification into groups with distinct survival outcomes. Using the open-source RADCURE dataset for head and neck cancer, MultiFIX achieves a C-index of 0.838 (prediction) and 0.826 (stratification), outperforming the clinical and academic baseline approaches and aligning with known prognostic markers. These results highlight the promise of interpretable multimodal AI for precision oncology with MultiFIX.

Robust Disease Prognosis via Diagnostic Knowledge Preservation: A Sequential Learning Approach

Rajamohan, H. R., Xu, Y., Zhu, W., Kijowski, R., Cho, K., Geras, K., Razavian, N., Deniz, C. M.

medrxiv logopreprintSep 25 2025
Accurate disease prognosis is essential for patient care but is often hindered by the lack of long-term data. This study explores deep learning training strategies that utilize large, accessible diagnostic datasets to pretrain models aimed at predicting future disease progression in knee osteoarthritis (OA), Alzheimers disease (AD), and breast cancer (BC). While diagnostic pretraining improves prognostic task performance, naive fine-tuning for prognosis can cause catastrophic forgetting, where the models original diagnostic accuracy degrades, a significant patient safety concern in real-world settings. To address this, we propose a sequential learning strategy with experience replay. We used cohorts with knee radiographs, brain MRIs, and digital mammograms to predict 4-year structural worsening in OA, 2-year cognitive decline in AD, and 5-year cancer diagnosis in BC. Our results showed that diagnostic pretraining on larger datasets improved prognosis model performance compared to standard baselines, boosting both the Area Under the Receiver Operating Characteristic curve (AUROC) (e.g., Knee OA external: 0.77 vs 0.747; Breast Cancer: 0.874 vs 0.848) and the Area Under the Precision-Recall Curve (AUPRC) (e.g., Alzheimers Disease: 0.752 vs 0.683). Additionally, a sequential learning approach with experience replay achieved prognostic performance comparable to dedicated single-task models (e.g., Breast Cancer AUROC 0.876 vs 0.874) while also preserving diagnostic ability. This method maintained high diagnostic accuracy (e.g., Breast Cancer Balanced Accuracy 50.4% vs 50.9% for a dedicated diagnostic model), unlike simpler multitask methods prone to catastrophic forgetting (e.g., 37.7%). Our findings show that leveraging large diagnostic datasets is a reliable and data-efficient way to enhance prognostic models while maintaining essential diagnostic skills. Author SummaryIn our research, we addressed a common problem in medical AI: how to accurately predict the future course of a disease when long-term patient data is rare. We focused on knee osteoarthritis, Alzheimers disease, and breast cancer. We found that we could significantly improve a models ability to predict disease progression by first training it on a much larger, more common type of data - diagnostic images used to assess a patients current disease state. We then developed a specialized training method that allows a single AI model to perform both diagnosis and prognosis tasks effectively. A key challenge is that models often "forget" their original diagnostic skills when they learn a new prognostic task. In a clinical setting, this poses a safety risk, as it could lead to missed diagnoses. We utilize experience replay to overcome this by continually refreshing the models diagnostic knowledge. This creates a more robust and efficient model that mirrors a clinicians workflow, offering the potential to improve patient care with limited amount of hard-to-get longitudinal data.

Deep-learning-based Radiomics on Mitigating Post-treatment Obesity for Pediatric Craniopharyngioma Patients after Surgery and Proton Therapy

Wenjun Yang, Chia-Ho Hua, Tina Davis, Jinsoo Uh, Thomas E. Merchant

arxiv logopreprintSep 25 2025
Purpose: We developed an artificial neural network (ANN) combining radiomics with clinical and dosimetric features to predict the extent of body mass index (BMI) increase after surgery and proton therapy, with advantage of improved accuracy and integrated key feature selection. Methods and Materials: Uniform treatment protocol composing of limited surgery and proton radiotherapy was given to 84 pediatric craniopharyngioma patients (aged 1-20 years). Post-treatment obesity was classified into 3 groups (<10%, 10-20%, and >20%) based on the normalized BMI increase during a 5-year follow-up. We developed a densely connected 4-layer ANN with radiomics calculated from pre-surgery MRI (T1w, T2w, and FLAIR), combining clinical and dosimetric features as input. Accuracy, area under operative curve (AUC), and confusion matrices were compared with random forest (RF) models in a 5-fold cross-validation. The Group lasso regularization optimized a sparse connection to input neurons to identify key features from high-dimensional input. Results: Classification accuracy of the ANN reached above 0.9 for T1w, T2w, and FLAIR MRI. Confusion matrices showed high true positive rates of above 0.9 while the false positive rates were below 0.2. Approximately 10 key features selected for T1w, T2w, and FLAIR MRI, respectively. The ANN improved classification accuracy by 10% or 5% when compared to RF models without or with radiomic features. Conclusion: The ANN model improved classification accuracy on post-treatment obesity compared to conventional statistics models. The clinical features selected by Group lasso regularization confirmed our practical observation, while the additional radiomic and dosimetric features could serve as imaging markers and mitigation methods on post-treatment obesity for pediatric craniopharyngioma patients.

Machine Learning-Based Classification of White Matter Functional Changes in Stroke Patients Using Resting-State fMRI.

Liu LH, Wang CX, Huang X, Chen RB

pubmed logopapersSep 25 2025
Neuroimaging studies of brain function are important research methods widely applied to stroke patients. Currently, a large number of studies have focused on functional imaging of the gray matter cortex. Relevant research indicates that certain areas of the gray matter cortex in stroke patients exhibit abnormal brain activity during resting state. However, studies on brain function based on white matter remain insufficient. The changes in functional connectivity caused by stroke in white matter, as well as the repair or compensation mechanisms of white matter function after stroke, are still unclear. The aim of this study is to investigate and demonstrate the changes in brain functional connectivity activity in the white matter region of stroke patients. Revealing the recombination characteristics of white matter functional networks after stroke, providing potential biomarkers for rehabilitation therapy Provide new clinical insights for the rehabilitation and treatment of stroke patients. We recruited 36 stroke patients and 36 healthy controls for resting-state functional magnetic resonance imaging (rs-fMRI). Regional Homogeneity (ReHo) and Degree Centrality (DC), which are sensitive to white matter functional abnormalities, were selected as feature vectors. ReHo reflects local neuronal synchrony, while DC quantifies global network hub properties. The combination of both effectively characterizes functional changes in white matter. ReHo evaluates the functional consistency of different white matter regions by calculating the activity similarity between adjacent brain regions. Additionally, DC analysis of white matter was used to investigate the connectivity patterns and organizational principles of functional networks between white matter regions. This was achieved by calculating the number of connections in each brain region to identify changes in neural activation of white matter regions that significantly impact the brain network. Furthermore, ReHo and DC metrics were used as feature vectors for classification using machine learning algorithms. The results indicated significant differences in white matter DC and ReHo values between stroke patients and healthy controls. In the two-sample t-test analysis of white matter DC, stroke patients showed a significant reduction in DC values in the corpus callosum genu (GCC), corpus callosum body (BCC), and left anterior corona radiata (ACRL) regions (GCC: 0.143 vs. 1.024; BCC: 0.238 vs. 1.143; ACRL: 0.143 vs. 0.821, p < 0.001). However, an increase in DC values was observed in the left superior longitudinal fasciculus (SLF_L) region (1.190 vs. 0.190, p < 0.001). In the two-sample t-test analysis of white matter ReHo, stroke patients exhibited a decrease in ReHo values in the GCC and BCC regions (GCC: 0.859 vs. 1.375; BCC: 1.156 vs. 1.687, p < 0.001), indicating values lower than those in the healthy control group. Using leave-one-out cross-validation (LOOCV) to evaluate the white matter DC and ReHo feature values through SVM classification models for stroke patients and healthy controls, the DC classification AUC was 0.89, and the ReHo classification AUC reached 0.98. These results suggest that the features possess validity and discriminative power. These findings suggest alterations in functional connectivity in specific white matter regions following stroke. Specifically, we observed a weakening of functional connectivity in the genu of the corpus callosum (GCC), the body of the corpus callosum (BCC), and the left anterior corona radiata (ACR_L) regions, while compensatory functional connectivity was enhanced in the left superior longitudinal fasciculus (SLF_L) region. These findings reveal the reorganization characteristics of white matter functional networks after stroke, which may provide potential biomarkers for the rehabilitation treatment of stroke patients and offer new clinical insights for their rehabilitation and treatment.

Revolutionizing Precise Low Back Pain Diagnosis via Contrastive Learning

Thanh Binh Le, Hoang Nhat Khang Vo, Tan-Ha Mai, Trong Nhan Phan

arxiv logopreprintSep 25 2025
Low back pain affects millions worldwide, driving the need for robust diagnostic models that can jointly analyze complex medical images and accompanying text reports. We present LumbarCLIP, a novel multimodal framework that leverages contrastive language-image pretraining to align lumbar spine MRI scans with corresponding radiological descriptions. Built upon a curated dataset containing axial MRI views paired with expert-written reports, LumbarCLIP integrates vision encoders (ResNet-50, Vision Transformer, Swin Transformer) with a BERT-based text encoder to extract dense representations. These are projected into a shared embedding space via learnable projection heads, configurable as linear or non-linear, and normalized to facilitate stable contrastive training using a soft CLIP loss. Our model achieves state-of-the-art performance on downstream classification, reaching up to 95.00% accuracy and 94.75% F1-score on the test set, despite inherent class imbalance. Extensive ablation studies demonstrate that linear projection heads yield more effective cross-modal alignment than non-linear variants. LumbarCLIP offers a promising foundation for automated musculoskeletal diagnosis and clinical decision support.

Artificial intelligence for diagnosis in interstitial lung disease and digital ontology for unclassified interstitial lung disease.

Baba T, Goto T, Kitamura Y, Iwasawa T, Okudela K, Takemura T, Osawa A, Ogura T

pubmed logopapersSep 24 2025
Multidisciplinary discussion (MDD) is the gold standard for diagnosis in interstitial lung disease (ILD). However, its inter-rater agreement is not satisfactory, and access to the MDD is limited due to a shortage of ILD experts. Therefore, artificial intelligence would be helpful for diagnosing ILD. We retrospectively analyzed data from 630 patients with ILD, including clinical information, CT images, and pathological results. The ILD classification into four clinicopathologic entities (i.e., idiopathic pulmonary fibrosis, non-specific interstitial pneumonia, hypersensitivity pneumonitis, connective tissue disease) consists of two stages: first, pneumonia pattern classification of CT images using a convolutional neural network (CNN) model; second, multimodal (clinical, radiological, and pathological) classification using a support vector machine (SVM). The performance of the classification algorithm was evaluated using 5-fold cross-validation. The mean accuracies of the CNN model and SVM were 62.4 % and 85.4 %, respectively. For multimodal classification using SVM, the overall accuracy was very high, especially with sensitivities for idiopathic pulmonary fibrosis and hypersensitivity pneumonitis exceeding 90 %. When pneumonia patterns from CT images, pathological results, or clinical information were not used, the SVM accuracy was 84.3 %, 70.3 % and 79.8 %, respectively, suggesting that pathological results contributed most to MDD diagnosis. When an unclassifiable interstitial pneumonia was input, the SVM output tended to align with the most likely diagnosis by the expert MDD team. The algorithm based on multimodal information can assist in diagnosing interstitial lung disease and is suitable for ontology diagnosis. (242 words).

Deep learning and radiomics integration of photoacoustic/ultrasound imaging for non-invasive prediction of luminal and non-luminal breast cancer subtypes.

Wang M, Mo S, Li G, Zheng J, Wu H, Tian H, Chen J, Tang S, Chen Z, Xu J, Huang Z, Dong F

pubmed logopapersSep 24 2025
This study aimed to develop a Deep Learning Radiomics integrated model (DLRN), which combines photoacoustic/ultrasound(PA/US)imaging with clinical and radiomics features to distinguish between luminal and non-luminal BC in a preoperative setting. A total of 388 BC patients were included, with 271 in the training group and 117 in the testing group. Radiomics and deep learning features were extracted from PA/US images using Pyradiomics and ResNet50, respectively. Feature selection was performed using independent sample t-tests, Pearson correlation analysis, and LASSO regression to build a Deep Learning Radiomics (DLR) model. Based on the results of univariate and multivariate logistic regression analyses, the DLR model was combined with valuable clinical features to construct the DLRN model. Model efficacy was assessed using AUC, accuracy, sensitivity, specificity, and NPV. The DLR model comprised 3 radiomic features and 6 deep learning features, which, when combined with significant clinical predictors, formed the DLRN model. In the testing set, the AUC of the DLRN model (0.924 [0.877-0.972]) was significantly higher than that of the DLR (AUC 0.847 [0.758-0.936], p = 0.026), DL (AUC 0.822 [0.725-0.919], p = 0.06), Rad (AUC 0.717 [0.597-0.838], p < 0.001), and clinical (AUC 0.820 [0.745-0.895], p = 0.002) models. These findings indicate that the DLRN model (integrated model) exhibited the most favorable predictive performance among all models evaluated. The DLRN model effectively integrates PA/US imaging with clinical data, showing potential for preoperative molecular subtype prediction and guiding personalized treatment strategies for BC patients.
Page 14 of 2332330 results
Show
per page

Ready to Sharpen Your Edge?

Join hundreds of your peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.