Sort by:
Page 3 of 1601593 results

MMRQA: Signal-Enhanced Multimodal Large Language Models for MRI Quality Assessment

Fankai Jia, Daisong Gan, Zhe Zhang, Zhaochi Wen, Chenchen Dan, Dong Liang, Haifeng Wang

arxiv logopreprintSep 29 2025
Magnetic resonance imaging (MRI) quality assessment is crucial for clinical decision-making, yet remains challenging due to data scarcity and protocol variability. Traditional approaches face fundamental trade-offs: signal-based methods like MRIQC provide quantitative metrics but lack semantic understanding, while deep learning approaches achieve high accuracy but sacrifice interpretability. To address these limitations, we introduce the Multimodal MRI Quality Assessment (MMRQA) framework, pioneering the integration of multimodal large language models (MLLMs) with acquisition-aware signal processing. MMRQA combines three key innovations: robust metric extraction via MRQy augmented with simulated artifacts, structured transformation of metrics into question-answer pairs using Qwen, and parameter-efficient fusion through Low-Rank Adaptation (LoRA) of LLaVA-OneVision. Evaluated on MR-ART, FastMRI, and MyConnectome benchmarks, MMRQA achieves state-of-the-art performance with strong zero-shot generalization, as validated by comprehensive ablation studies. By bridging quantitative analysis with semantic reasoning, our framework generates clinically interpretable outputs that enhance quality control in dynamic medical settings.

Cross-regional radiomics: a novel framework for relationship-based feature extraction with validation in Parkinson's disease motor subtyping.

Hosseini MS, Aghamiri SMR, Panahi M

pubmed logopapersSep 29 2025
Traditional radiomics approaches focus on single-region feature extraction, limiting their ability to capture complex inter-regional relationships crucial for understanding pathophysiological mechanisms in complex diseases. This study introduces a novel cross-regional radiomics framework that systematically extracts relationship-based features between anatomically and functionally connected brain regions. We analyzed T1-weighted magnetic resonance imaging (MRI) data from 140 early-stage Parkinson's disease patients (70 tremor-dominant, 70 postural instability gait difficulty) from the Parkinson's Progression Markers Initiative (PPMI) database across multiple imaging centers. Eight bilateral motor circuit regions (putamen, caudate nucleus, globus pallidus, substantia nigra) were segmented using standardized atlases. Two feature sets were developed: 48 traditional single-region of interest (ROI) features and 60 novel motor-circuit features capturing cross-regional ratios, asymmetry indices, volumetric relationships, and shape distributions. Six feature engineering scenarios were evaluated using center-based 5-fold cross-validation with six machine learning classifiers to ensure robust generalization across different imaging centers. Motor-circuit features demonstrated superior performance compared to single-ROI features across enhanced preprocessing scenarios. Peak performance was achieved with area under the curve (AUC) of 0.821 ± 0.117 versus 0.650 ± 0.220 for single-ROI features (p = 0.0012, Cohen's d = 0.665). Cross-regional ratios, particularly putamen-substantia nigra relationships, dominated the most discriminative features. Motor-circuit features showed superior generalization across multi-center data and better clinical utility through decision curve analysis and calibration curves. The proposed cross-regional radiomics framework significantly outperforms traditional single-region approaches for Parkinson's disease motor subtype classification. This methodology provides a foundation for advancing radiomics applications in complex diseases where inter-regional connectivity patterns are fundamental to pathophysiology.

Novel multi-task learning for Alzheimer's stage classification using hippocampal MRI segmentation, feature fusion, and nomogram modeling.

Hu W, Du Q, Wei L, Wang D, Zhang G

pubmed logopapersSep 29 2025
To develop and validate a comprehensive and interpretable framework for multi-class classification of Alzheimer's disease (AD) progression stages based on hippocampal MRI, integrating radiomic, deep, and clinical features. This retrospective multi-center study included 2956 patients across four AD stages (Non-Demented, Very Mild Demented, Mild Demented, Moderate Demented). T1-weighted MRI scans were processed through a standardized pipeline involving hippocampal segmentation using four models (U-Net, nnU-Net, Swin-UNet, MedT). Radiomic features (n = 215) were extracted using the SERA platform, and deep features (n = 256) were learned using an LSTM network with attention applied to hippocampal slices. Fused features were harmonized with ComBat and filtered by ICC (≥ 0.75), followed by LASSO-based feature selection. Classification was performed using five machine learning models, including Logistic Regression (LR), Support Vector Machine (SVM), Random Forest (RF), Multilayer Perceptron (MLP), and eXtreme Gradient Boosting (XGBoost). Model interpretability was addressed using SHAP, and a nomogram and decision curve analysis (DCA) were developed. Additionally, an end-to-end 3D CNN-LSTM model and two transformer-based benchmarks (Vision Transformer, Swin Transformer) were trained for comparative evaluation. MedT achieved the best hippocampal segmentation (Dice = 92.03% external). Fused features yielded the highest classification performance with XGBoost (external accuracy = 92.8%, AUC = 94.2%). SHAP identified MMSE, hippocampal volume, and APOE ε4 as top contributors. The nomogram accurately predicted early-stage AD with clinical utility confirmed by DCA. The end-to-end model performed acceptably (AUC = 84.0%) but lagged behind the fused pipeline. Statistical tests confirmed significant performance advantages for feature fusion and MedT-based segmentation. This study demonstrates that integrating radiomics, deep learning, and clinical data from hippocampal MRI enables accurate and interpretable classification of AD stages. The proposed framework is robust, generalizable, and clinically actionable, representing a scalable solution for AD diagnostics.

Convolutional neural network models of structural MRI for discriminating categories of cognitive impairment: a systematic review and meta-analysis.

Dong X, Li Y, Hao J, Zhou P, Yang C, Ai Y, He M, Zhang W, Hu H

pubmed logopapersSep 29 2025
Alzheimer's disease (AD) and mild cognitive impairment (MCI) pose significant challenges to public health and underscore the need for accurate and early diagnostic tools. Structural magnetic resonance imaging (sMRI) combined with advanced analytical techniques like convolutional neural networks (CNNs) seemed to offer a promising avenue for the diagnosis of these conditions. This systematic review and meta-analysis aimed to evaluate the diagnostic performance of CNN algorithms applied to sMRI data in differentiating between AD, MCI, and normal cognition (NC). Following the PRISMA-DTA guidelines, a comprehensive literature search was carried out in PubMed and Web of Science databases for studies published between 2018 and 2024. Studies were included if they employed CNNs for the diagnostic classification of sMRI data from participants with AD, MCI, or NC. The methodological quality of the included studies was assessed using the QUADAS-2 and METRICS tools. Data extraction and statistical analysis were performed to calculate pooled diagnostic accuracy metrics. A total of 21 studies were included in the study, comprising 16,139 participants in the analysis. The pooled sensitivity and specificity of CNN algorithms for differentiating AD from NC were 0.92 and 0.91, respectively. For distinguishing MCI from NC, the pooled sensitivity and specificity were 0.74 and 0.79, respectively. The algorithms also showed a moderate ability to differentiate AD from MCI, with a pooled sensitivity and specificity of 0.73 and 0.79, respectively. In the pMCI versus sMCI classification, a pooled sensitivity was 0.69 and a specificity was 0.81. Heterogeneity across studies was significant, as indicated by meta-regression results. CNN algorithms demonstrated promising diagnostic performance in differentiating AD, MCI, and NC using sMRI data. The highest accuracy was observed in distinguishing AD from NC and the lowest accuracy observed in distinguishing pMCI from sMCI. These findings suggest that CNN-based radiomics has the potential to serve as a valuable tool in the diagnostic armamentarium for neurodegenerative diseases. However, the heterogeneity among studies indicates a need for further methodological refinement and validation. This systematic review was registered in PROSPERO (Registration ID: CRD42022295408).

Evaluation of Context-Aware Prompting Techniques for Classification of Tumor Response Categories in Radiology Reports Using Large Language Model.

Park J, Sim WS, Yu JY, Park YR, Lee YH

pubmed logopapersSep 29 2025
Radiology reports are essential for medical decision-making, providing crucial data for diagnosing diseases, devising treatment plans, and monitoring disease progression. While large language models (LLMs) have shown promise in processing free-text reports, research on effective prompting techniques for radiologic applications remains limited. To evaluate the effectiveness of LLM-driven classification based on radiology reports in terms of tumor response category (TRC), and to optimize the model through a comparison of four different prompt engineering techniques for effectively performing this classification task in clinical applications, we included 3062 whole-spine contrast-enhanced magnetic resonance imaging (MRI) radiology reports for prompt engineering and validation. TRCs were labeled by two radiologists based on criteria modified from the Response Evaluation Criteria in Solid Tumors (RECIST) guidelines. The Llama3 instruct model was used to classify TRCs in this study through four different prompts: General, In-Context Learning (ICL), Chain-of-Thought (CoT), and ICL with CoT. AUROC, accuracy, precision, recall, and F1-score were calculated against each prompt and model (8B, 70B) with the test report dataset. The average AUROC for ICL (0.96 internally, 0.93 externally) and ICL with CoT prompts (0.97 internally, 0.94 externally) outperformed other prompts. Error increased with prompt complexity, including 0.8% incomplete sentence errors and 11.3% probability-classification inconsistencies. This study demonstrates that context-aware LLM prompts substantially improved the efficiency and effectiveness of classifying TRCs from radiology reports, despite potential intrinsic hallucinations. While further improvements are required for real-world application, our findings suggest that context-aware prompts have significant potential for segmenting complex radiology reports and enhancing oncology clinical workflows.

Optimized T<sub>1</sub>-weighted MP-RAGE MRI of the brain at 0.55 T using variable flip angle coherent gradient echo imaging and deep learning reconstruction.

Bieri O, Nickel MD, Weidensteiner C, Madörin P, Bauman G

pubmed logopapersSep 29 2025
To propose and evaluate an optimized MP-RAGE protocol for rapid T<sub>1</sub>-weighted imaging of the brain at 0.55 T. Incoherent and coherent steady state free precession (SSFP) RAGE kernels with constant and variable excitation angles were investigated in terms of the white matter SNR and the white matter-gray matter signal difference. Potential edge smearing from the transient signal readout was assessed based on a differential point spread function analysis. Finally, the prospects of a deep-learning reconstruction (DLR) method for accelerated MP-RAGE MRI of undersampled data were evaluated for the best performing variant. MP-RAGE imaging with a variable flip angle (vFA) SSFP-FID kernel outperformed all other investigated variants. As compared to the standard MPRAGE sequence using a spoiled gradient echo kernel with constant flip angle, vFA SSFP-FID offered an average gain in the white matter SNR of 21% ± 2% and an average improvement for the white matter-gray matter signal difference for cortical gray matter of 47% ± 7%. The differential point spread function was narrowest for the spoiled gradient echo but slightly increased by 8% for vFA SSFP-FID. For vFA SSFP-FID, DLR offered a considerable decrease in the overall scan time from 5:17 min down to 2:46 min without noticeable image artifacts and degradations. At 0.55 T, a vFA MP-RAGE variant using an SSFP-FID kernel combined with a DLR method offers excellent prospects for rapid T<sub>1</sub>-weighted whole brain imaging in less than 3 min with nearly 1 mm (1.12 × 1.17 × 1.25 mm<sup>3</sup>) isotropic resolution.

Enhancing Spinal Cord and Canal Segmentation in Degenerative Cervical Myelopathy : The Role of Interactive Learning Models with manual Click.

Han S, Oh JK, Cho W, Kim TJ, Hong N, Park SB

pubmed logopapersSep 29 2025
We aim to develop an interactive segmentation model that can offer accuracy and reliability for the segmentation of the irregularly shaped spinal cord and canal in degenerative cervical myelopathy (DCM) through manual click and model refinement. A dataset of 1444 frames from 294 magnetic resonance imaging records of DCM patients was used and we developed two different segmentation models for comparison : auto-segmentation and interactive segmentation. The former was based on U-Net and utilized a pretrained ConvNeXT-tiny as its encoder. For the latter, we employed an interactive segmentation model structured by SimpleClick, a large model that utilizes a vision transformer as its backbone, together with simple fine-tuning. The segmentation performance of the two models were compared in terms of their Dice scores, mean intersection over union (mIoU), Average Precision and Hausdorff distance. The efficiency of the interactive segmentation model was evaluated by the number of clicks required to achieve a target mIoU. Our model achieved better scores across all four-evaluation metrics for segmentation accuracy, showing improvements of +6.4%, +1.8%, +3.7%, and -53.0% for canal segmentation, and +11.7%, +6.0%, +18.2%, and -70.9% for cord segmentation with 15 clicks, respectively. The required clicks for the interactive segmentation model to achieve a 90% mIoU for spinal canal with cord cases and 80% mIoU for spinal cord cases were 11.71 and 11.99, respectively. We found that the interactive segmentation model significantly outperformed the auto-segmentation model. By incorporating simple manual inputs, the interactive model effectively identified regions of interest, particularly in the complex and irregular shapes of the spinal cord, demonstrating both enhanced accuracy and adaptability.

Classification of anterior cruciate ligament tears in knee magnetic resonance images using pre-trained model and custom model.

Thangaperumal S, Murugan PR, Hossen J, Wong WK, Ng PK

pubmed logopapersSep 29 2025
An anterior cruciate ligament (ACL) tear is a prevalent knee injury among athletes, and aged people with osteoporosis are at increased risk for it. For early detection and treatment, precise and rapid identification of ACL tears is significant. A fully automated system that can identify ACL tear is necessary to aid healthcare providers in determining the nature of injuries detected on Magnetic Resonance Imaging (MRI) scans. Two Convolutional Neural Networks (CNN), the pretrained model and the CustomNet model are trained and tested using 581 MRI scans of the knee. Feature extraction is done with the pre-trained ResNet-18 model, and the ISOMAP algorithm is used in the CustomNet model. Linear and nonlinear dimensionality reduction techniques are employed to extract the needed features from the image. For the ResNet-18 model, the accuracy rate ranges between 86% and 92% for various data partitions. After performing PCA, the improved classification rate ranges between 92% and 96.2%. The CustomNet model's accuracy rate ranges from 40 to 70%, 70-90%, 60-70%, and 50-70% for different hyperparameter ensembles. Five-fold cross validation is implemented in CustomNet and it achieved an overall accuracy of 85.6%. These two models demonstrate superior efficiency and accuracy in classifying normal and ACL torn Knee MR images.

Clinical and MRI markers for acute vs chronic temporomandibular disorders using a machine learning and deep neural networks.

Lee YH, Jeon S, Kim DH, Auh QS, Lee JH, Noh YK

pubmed logopapersSep 29 2025
Exploring the transition from acute to chronic temporomandibular disorders (TMD) remains challenging due to the multifactorial nature of the disease. This study aims to identify clinical, behavioral, and imaging-based predictors that contribute to symptom chronicity in patients with TMD. We enrolled 239 patients with TMD (161 women, 78 men; mean age 35.60 ± 17.93 years), classified as acute ( < 6 months) or chronic ( ≥ 6 months) based on symptom duration. TMD was diagnosed according to the Diagnostic Criteria for TMD (DC/TMD Axis I). Clinical data, sleep-related variables, and temporomandibular joint magnetic resonance imaging (MRI) were collected. MRI assessments included anterior disc displacement (ADD), joint space narrowing, osteoarthritis, and effusion using 3 T T2-weighted and proton density scans. Predictors were evaluated using logistic regression and deep neural networks (DNN), and performance was compared. Chronic TMD is observed in 51.05% of patients. Compared to acute cases, chronic TMD is more frequently associated with TMJ noise (70.5%), bruxism (31.1%), and higher pain intensity (VAS: 4.82 ± 2.47). They also have shorter sleep and higher STOP-Bang scores, indicating greater risk of obstructive sleep apnea. MRI findings reveal increased prevalence of ADD (86.9%), TMJ-OA (82.0%), and joint space narrowing (88.5%) in chronic TMD. Logistic regression achieves an AUROC of 0.7550 (95% CI: 0.6550-0.8550), identifying TMJ noise, bruxism, VAS, sleep disturbance, STOP-Bang≥5, ADD, and joint space narrowing as significant predictors. The DNN model improves accuracy to 79.49% compared to 75.50%, though the difference is not statistically significant (p = 0.3067). Behavioral and TMJ-related structural factors are key predictors of chronic TMD and may aid early identification. Timely recognition may support personalized strategies and improve outcomes.

Cycle Diffusion Model for Counterfactual Image Generation

Fangrui Huang, Alan Wang, Binxu Li, Bailey Trang, Ridvan Yesiloglu, Tianyu Hua, Wei Peng, Ehsan Adeli

arxiv logopreprintSep 29 2025
Deep generative models have demonstrated remarkable success in medical image synthesis. However, ensuring conditioning faithfulness and high-quality synthetic images for direct or counterfactual generation remains a challenge. In this work, we introduce a cycle training framework to fine-tune diffusion models for improved conditioning adherence and enhanced synthetic image realism. Our approach, Cycle Diffusion Model (CDM), enforces consistency between generated and original images by incorporating cycle constraints, enabling more reliable direct and counterfactual generation. Experiments on a combined 3D brain MRI dataset (from ABCD, HCP aging & young adults, ADNI, and PPMI) show that our method improves conditioning accuracy and enhances image quality as measured by FID and SSIM. The results suggest that the cycle strategy used in CDM can be an effective method for refining diffusion-based medical image generation, with applications in data augmentation, counterfactual, and disease progression modeling.
Page 3 of 1601593 results
Show
per page

Ready to Sharpen Your Edge?

Join hundreds of your peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.