Sort by:
Page 4 of 3293286 results

Performance of GPT-5 in Brain Tumor MRI Reasoning

Mojtaba Safari, Shansong Wang, Mingzhe Hu, Zach Eidex, Qiang Li, Xiaofeng Yang

arxiv logopreprintAug 14 2025
Accurate differentiation of brain tumor types on magnetic resonance imaging (MRI) is critical for guiding treatment planning in neuro-oncology. Recent advances in large language models (LLMs) have enabled visual question answering (VQA) approaches that integrate image interpretation with natural language reasoning. In this study, we evaluated GPT-4o, GPT-5-nano, GPT-5-mini, and GPT-5 on a curated brain tumor VQA benchmark derived from 3 Brain Tumor Segmentation (BraTS) datasets - glioblastoma (GLI), meningioma (MEN), and brain metastases (MET). Each case included multi-sequence MRI triplanar mosaics and structured clinical features transformed into standardized VQA items. Models were assessed in a zero-shot chain-of-thought setting for accuracy on both visual and reasoning tasks. Results showed that GPT-5-mini achieved the highest macro-average accuracy (44.19%), followed by GPT-5 (43.71%), GPT-4o (41.49%), and GPT-5-nano (35.85%). Performance varied by tumor subtype, with no single model dominating across all cohorts. These findings suggest that GPT-5 family models can achieve moderate accuracy in structured neuro-oncological VQA tasks, but not at a level acceptable for clinical use.

Artificial Intelligence based fractional flow reserve.

Bednarek A, Gąsior P, Jaguszewski M, Buszman PP, Milewski K, Hawranek M, Gil R, Wojakowski W, Kochman J, Tomaniak M

pubmed logopapersAug 14 2025
Fractional flow reserve (FFR) - a physiological indicator of coronary stenosis significance - has now become a widely used parameter also in the guidance of percutaneous coronary intervention (PCI). Several studies have shown the superiority of FFR compared to visual assessment, contributing to the reduction in clinical endpoints. However, the current approach to FFR assessment requires coronary instrumentation with a dedicated pressure wire and thus increasing invasiveness, cost, and duration of the procedure. Alternative, noninvasive methods of FFR assessment based on computational fluid dynamics are being widely tested; these approaches are generally not fully automated and may sometimes require substantial computational power. Nowadays, one of the most rapidly expanding fields in medicine is the use of artificial intelligence (AI) in therapy optimization, diagnosis, treatment, and risk stratification. AI usage contributes to the development of more sophisticated methods of imaging analysis and allows for the derivation of clinically important parameters in a faster and more accurate way. Over the recent years, AI utility in deriving FFR in a noninvasive manner has been increasingly reported. In this review, we critically summarize current knowledge in the field of AI-derived FFR based on data from computed tomography angiography, invasive angiography, optical coherence tomography, and intravascular ultrasound. Available solutions, possible future directions in optimizing cathlab performance, including the use of mixed reality, as well as current limitations standing behind the wide adoption of these techniques, are overviewed.

Enhancing cardiac MRI reliability at 3 T using motion-adaptive B<sub>0</sub> shimming.

Huang Y, Malagi AV, Li X, Guan X, Yang CC, Huang LT, Long Z, Zepeda J, Zhang X, Yoosefian G, Bi X, Gao C, Shang Y, Binesh N, Lee HL, Li D, Dharmakumar R, Han H, Yang HR

pubmed logopapersAug 14 2025
Magnetic susceptibility differences at the heart-lung interface introduce B<sub>0</sub>-field inhomogeneities that challenge cardiac MRI at high field strengths (≥ 3 T). Although hardware-based shimming has advanced, conventional approaches often neglect dynamic variations in thoracic anatomy caused by cardiac and respiratory motion, leading to residual off-resonance artifacts. This study aims to characterize motion-induced B<sub>0</sub>-field fluctuations in the heart and evaluate a deep learning-enabled motion-adaptive B<sub>0</sub> shimming pipeline to mitigate them. A motion-resolved B<sub>0</sub> mapping sequence was implemented at 3 T to quantify cardiac and respiratory-induced B<sub>0</sub> variations. A motion-adaptive shimming framework was then developed and validated through numerical simulations and human imaging studies. B<sub>0</sub>-field homogeneity and T<sub>2</sub>* mapping accuracy were assessed in multiple breath-hold positions using standard and motion-adaptive shimming. Respiratory motion significantly altered myocardial B<sub>0</sub> fields (p < 0.01), whereas cardiac motion had minimal impact (p = 0.49). Compared with conventional scanner shimming, motion-adaptive B<sub>0</sub> shimming yielded significantly improved field uniformity across both inspiratory (post-shim SD<sub>ratio</sub>: 0.68 ± 0.10 vs. 0.89 ± 0.11; p < 0.05) and expiratory (0.65 ± 0.16 vs. 0.84 ± 0.20; p < 0.05) breath-hold states. Corresponding improvements in myocardial T<sub>2</sub>* map homogeneity were observed, with reduced coefficient of variation (0.44 ± 0.19 vs. 0.39 ± 0.22; 0.59 ± 0.30 vs. 0.46 ± 0.21; both p < 0.01). The proposed motion-adaptive B<sub>0</sub> shimming approach effectively compensates for respiration-induced B<sub>0</sub> fluctuations, enhancing field homogeneity and reducing off-resonance artifacts. This strategy improves the robustness and reproducibility of T<sub>2</sub>* mapping, enabling more reliable high-field cardiac MRI.

A novel hybrid convolutional and recurrent neural network model for automatic pituitary adenoma classification using dynamic contrast-enhanced MRI.

Motamed M, Bastam M, Tabatabaie SM, Elhaie M, Shahbazi-Gahrouei D

pubmed logopapersAug 14 2025
Pituitary adenomas, ranging from subtle microadenomas to mass-effect macroadenomas, pose diagnostic challenges for radiologists due to increasing scan volumes and the complexity of dynamic contrast-enhanced MRI interpretation. A hybrid CNN-LSTM model was trained and validated on a multi-center dataset of 2,163 samples from Tehran and Babolsar, Iran. Transfer learning and preprocessing techniques (e.g., Wiener filters) were utilized to improve classification performance for microadenomas (< 10 mm) and macroadenomas (> 10 mm). The model achieved 90.5% accuracy, an area under the receiver operating characteristic curve (AUROC) of 0.92, and 89.6% sensitivity (93.5% for microadenomas, 88.3% for macroadenomas), outperforming standard CNNs by 5-18% across metrics. With a processing time of 0.17 s per scan, the model demonstrated robustness to variations in imaging conditions, including scanner differences and contrast variations, excelling in real-time detection and differentiation of adenoma subtypes. This dual-path approach, the first to synergize spatial and temporal MRI features for pituitary diagnostics, offers high precision and efficiency. Supported by comparisons with existing models, it provides a scalable, reproducible tool to improve patient outcomes, with potential adaptability to broader neuroimaging challenges.

Instantaneous T<sub>2</sub> Mapping via Reduced Field of View Multiple Overlapping-Echo Detachment Imaging: Application in Free-Breathing Abdominal and Myocardial Imaging.

Dai C, Cai C, Wu J, Zhu L, Qu X, Yang Q, Zhou J, Cai S

pubmed logopapersAug 14 2025
Quantitative magnetic resonance imaging (qMRI) has attracted more and more attention in clinical diagnosis and medical sciences due to its capability to non-invasively characterize tissue properties. Nevertheless, most qMRI methods are time-consuming and sensitive to motion, making them inadequate for quantifying organs with physiological movement. In this context, single-shot multiple overlapping-echo detachment (MOLED) imaging technique has been presented, but its acquisition efficiency and image quality are limited when the field of view (FOV) is smaller than the object, especially for abdominal organs and myocardium. A novel single-shot reduced FOV qMRI method was developed based on MOLED (termed rFOV-MOLED). This method combines zonal oblique multislice (ZOOM) and outer volume suppression (OVS) techniques to reduce the FOV and suppress signals outside the FOV. A deep neural network was trained using synthetic data generated from Bloch simulations to achieve high-quality T<sub>2</sub> map reconstruction from rFOV-MOLED iamges. Numerical simulation, water phantom and in vivo abdominal and myocardial imaging experiments were performed to evaluate the method. The coefficient of variation and repeatability index were used to evaluate the reproducibility. Multiple statistical analyses were utilized to evaluate the accuracy and significance of the method, including linear regression, Bland-Altman analysis, Wilcoxon signed-rank test, and Mann-Whitney U test, with the p-value significance level of 0.05. Experimental results show that rFOV-MOLED achieved excellent performance in reducing aliasing signals due to FOV reduction. It provided T<sub>2</sub> maps closely resembling the reference maps. Moreover, it gave finer tissue details than MOLED and was quite repeatable. rFOV-MOLED can ultrafast and stably provide accurate T2 maps for myocardium and specific abdominal organs with improved acquisition efficiency and image quality.

Deep learning-based non-invasive prediction of PD-L1 status and immunotherapy survival stratification in esophageal cancer using [<sup>18</sup>F]FDG PET/CT.

Xie F, Zhang M, Zheng C, Zhao Z, Wang J, Li Y, Wang K, Wang W, Lin J, Wu T, Wang Y, Chen X, Li Y, Zhu Z, Wu H, Li Y, Liu Q

pubmed logopapersAug 14 2025
This study aimed to develop and validate deep learning models using [<sup>18</sup>F]FDG PET/CT to predict PD-L1 status in esophageal cancer (EC) patients. Additionally, we assessed the potential of derived deep learning model scores (DLS) for survival stratification in immunotherapy. In this retrospective study, we included 331 EC patients from two centers, dividing them into training, internal validation, and external validation cohorts. Fifty patients who received immunotherapy were followed up. We developed four 3D ResNet10-based models-PET + CT + clinical factors (CPC), PET + CT (PC), PET (P), and CT (C)-using pre-treatment [<sup>18</sup>F]FDG PET/CT scans. For comparison, we also constructed a logistic model incorporating clinical factors (clinical model). The DLS were evaluated as radiological markers for survival stratification, and nomograms for predicting survival were constructed. The models demonstrated accurate prediction of PD-L1 status. The areas under the curve (AUCs) for predicting PD-L1 status were as follows: CPC (0.927), PC (0.904), P (0.886), C (0.934), and the clinical model (0.603) in the training cohort; CPC (0.882), PC (0.848), P (0.770), C (0.745), and the clinical model (0.524) in the internal validation cohort; and CPC (0.843), PC (0.806), P (0.759), C (0.667), and the clinical model (0.671) in the external validation cohort. The CPC and PC models exhibited superior predictive performance. Survival analysis revealed that the DLS from most models effectively stratified overall survival and progression-free survival at appropriate cut-off points (P < 0.05), outperforming stratification based on PD-L1 status (combined positive score ≥ 10). Furthermore, incorporating model scores with clinical factors in nomograms enhanced the predictive probability of survival after immunotherapy. Deep learning models based on [<sup>18</sup>F]FDG PET/CT can accurately predict PD-L1 status in esophageal cancer patients. The derived DLS can effectively stratify survival outcomes following immunotherapy, particularly when combined with clinical factors.

Deep Learning-Based Instance-Level Segmentation of Kidney and Liver Cysts in CT Images of Patients Affected by Polycystic Kidney Disease.

Gregory AV, Khalifa M, Im J, Ramanathan S, Elbarougy DE, Cruz C, Yang H, Denic A, Rule AD, Chebib FT, Dahl NK, Hogan MC, Harris PC, Torres VE, Erickson BJ, Potretzke TA, Kline TL

pubmed logopapersAug 14 2025
Total kidney and liver volumes are key image-based biomarkers to predict the severity of kidney and liver phenotype in autosomal dominant polycystic kidney disease (ADPKD). However, MRI-based advanced biomarkers like total cyst number (TCN) and cyst parenchyma surface area (CPSA) have been shown to more accurately assess cyst burden and improve the prediction of disease progression. The main aim of this study is to extend the calculation of advanced biomarkers to other imaging modalities; thus, we propose a fully automated model to segment kidney and liver cysts in CT images. Abdominal CTs of ADPKD patients were gathered retrospectively between 2001-2018. A 3D deep-learning method using the nnU-Net architecture was trained to learn cyst edges-cores and the non-cystic kidney/liver parenchyma. Separate segmentation models were trained for kidney cysts in contrast-enhanced CTs and liver cysts in non-contrast CTs using an active learning approach. Two experienced research fellows manually generated the reference standard segmentation, which were reviewed by an expert radiologist for accuracy. Two-hundred CT scans from 148 patients (mean age, 51.2 ± 14.1 years; 48% male) were utilized for model training (80%) and testing (20%). In the test set, both models showed good agreement with the reference standard segmentations, similar to the agreement between two independent human readers (model vs reader: TCNkidney/liver r=0.96/0.97 and CPSAkidney r=0.98), inter-reader: TCNkidney/liver r=0.96/0.98 and CPSAkidney r=0.99). Our study demonstrates that automated models can segment kidney and liver cysts accurately in CT scans of patients with ADPKD.

Development and validation of deep learning model for detection of obstructive coronary artery disease in patients with acute chest pain: a multi-center study.

Kim JY, Park J, Lee KH, Lee JW, Park J, Kim PK, Han K, Baek SE, Im DJ, Choi BW, Hur J

pubmed logopapersAug 14 2025
This study aimed to develop and validate a deep learning (DL) model to detect obstructive coronary artery disease (CAD, ≥ 50% stenosis) in coronary CT angiography (CCTA) among patients presenting to the emergency department (ED) with acute chest pain. The training dataset included 378 patients with acute chest pain who underwent CCTA (10,060 curved multiplanar reconstruction [MPR] images) from a single-center ED between January 2015 and December 2022. The external validation dataset included 298 patients from 3 ED centers between January 2021 and December 2022. A DL model based on You Only Look Once v4, requires manual preprocessing for curved MPR extraction and was developed using 15 manually preprocessed MPR images per major coronary artery. Model performance was evaluated per artery and per patient. The training dataset included 378 patients (mean age 61.3 ± 12.2 years, 58.2% men); the external dataset included 298 patients (mean age 58.3 ± 13.8 years, 54.6% men). Obstructive CAD prevalence in the external dataset was 27.5% (82/298). The DL model achieved per-artery sensitivity, specificity, positive predictive value, negative predictive value (NPV), and area under the curve (AUC) of 92.7%, 89.9%, 62.6%, 98.5%, and 0.919, respectively; and per-patient values of 93.3%, 80.7%, 67.7%, 96.6%, and 0.871, respectively. The DL model demonstrated high sensitivity and NPV for identifying obstructive CAD in patients with acute chest pain undergoing CCTA, indicating its potential utility in aiding ED physicians in CAD detection.

Severity Classification of Pediatric Spinal Cord Injuries Using Structural MRI Measures and Deep Learning: A Comprehensive Analysis across All Vertebral Levels.

Sadeghi-Adl Z, Naghizadehkashani S, Middleton D, Krisa L, Alizadeh M, Flanders AE, Faro SH, Wang Z, Mohamed FB

pubmed logopapersAug 14 2025
Spinal cord injury (SCI) in the pediatric population presents a unique challenge in diagnosis and prognosis due to the complexity of performing clinical assessments on children. Accurate evaluation of structural changes in the spinal cord is essential for effective treatment planning. This study aims to evaluate structural characteristics in pediatric patients with SCI by comparing cross-sectional area (CSA), anterior-posterior (AP) width, and right-left (RL) width across all vertebral levels of the spinal cord between typically developing (TD) and participants with SCI. We employed deep learning techniques to utilize these measures for detecting SCI cases and determining their injury severity. Sixty-one pediatric participants (ages 6-18), including 20 with chronic SCI and 41 TD, were enrolled and scanned by using a 3T MRI scanner. All SCI participants underwent the International Standards for Neurological Classification of Spinal Cord Injury (ISNCSCI) test to assess their neurologic function and determine their American Spinal Injury Association (ASIA) Impairment Scale (AIS) category. T2-weighted MRI scans were utilized to measure CSA, AP width, and RL widths along the entire cervical and thoracic cord. These measures were automatically extracted at every vertebral level of the spinal cord by using the spinal cord toolbox. Deep convolutional neural networks (CNNs) were utilized to classify participants into SCI or TD groups and determine their AIS classification based on structural parameters and demographic factors such as age and height. Significant differences (<i>P</i> < .05) were found in CSA, AP width, and RL width between SCI and TD participants, indicating notable structural alterations due to SCI. The CNN-based models demonstrated high performance, achieving 96.59% accuracy in distinguishing SCI from TD participants. Furthermore, the models determined AIS category classification with 94.92% accuracy. The study demonstrates the effectiveness of integrating cross-sectional structural imaging measures with deep learning methods for classification and severity assessment of pediatric SCI. The deep learning approach outperforms traditional machine learning models in diagnostic accuracy, offering potential improvements in patient care in pediatric SCI management.

GNN-based Unified Deep Learning

Furkan Pala, Islem Rekik

arxiv logopreprintAug 14 2025
Deep learning models often struggle to maintain generalizability in medical imaging, particularly under domain-fracture scenarios where distribution shifts arise from varying imaging techniques, acquisition protocols, patient populations, demographics, and equipment. In practice, each hospital may need to train distinct models - differing in learning task, width, and depth - to match local data. For example, one hospital may use Euclidean architectures such as MLPs and CNNs for tabular or grid-like image data, while another may require non-Euclidean architectures such as graph neural networks (GNNs) for irregular data like brain connectomes. How to train such heterogeneous models coherently across datasets, while enhancing each model's generalizability, remains an open problem. We propose unified learning, a new paradigm that encodes each model into a graph representation, enabling unification in a shared graph learning space. A GNN then guides optimization of these unified models. By decoupling parameters of individual models and controlling them through a unified GNN (uGNN), our method supports parameter sharing and knowledge transfer across varying architectures (MLPs, CNNs, GNNs) and distributions, improving generalizability. Evaluations on MorphoMNIST and two MedMNIST benchmarks - PneumoniaMNIST and BreastMNIST - show that unified learning boosts performance when models are trained on unique distributions and tested on mixed ones, demonstrating strong robustness to unseen data with large distribution shifts. Code and benchmarks: https://github.com/basiralab/uGNN
Page 4 of 3293286 results
Show
per page

Ready to Sharpen Your Edge?

Join hundreds of your peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.