Sort by:
Page 17 of 1061052 results

Machine Learning-Based Reconstruction of 2D MRI for Quantitative Morphometry in Epilepsy

Ratcliffe, C., Taylor, P. N., de Bezenac, C., Das, K., Biswas, S., Marson, A., Keller, S. S.

medrxiv logopreprintAug 6 2025
IntroductionStructural neuroimaging analyses require research quality images, acquired with costly MRI acquisitions. Isotropic (3D-T1) images are desirable for quantitative analyses, however a routine compromise in the clinical setting is to acquire anisotropic (2D-T1) analogues for qualitative visual inspection. ML (Machine learning-based) software have shown promise in addressing some of the limitations of 2D-T1 scans in research applications, yet their efficacy in quantitative research is generally poorly understood. Pathology-related abnormalities of the subcortical structures have previously been identified in idiopathic generalised epilepsy (IGE), which have been overlooked based on visual inspection, through the use of quantitative morphometric analyses. As such, IGE biomarkers present a suitable model in which to evaluate the applicability of image preprocessing methods. This study therefore explores subcortical structural biomarkers of IGE, first in our silver standard 3D-T1 scans, then in 2D-T1 scans that were either untransformed, resampled using a classical interpolation approach, or synthesised with a resolution and contrast agnostic ML model (the latter of which is compared to a separate model). Methods2D-T1 and 3D-T1 MRI scans were acquired during the same scanning session for 33 individuals with drug-responsive IGE (age mean 32.16 {+/-} SD = 14.20, male n = 14) and 42 individuals with drug-resistant IGE (31.76 {+/-} 11.12, 17), all diagnosed at the Walton Centre NHS Foundation Trust Liverpool, alongside 39 age- and sex-matched healthy controls (32.32 {+/-} 8.65, 16). The untransformed 2D-T1 scans were resampled into isotropic images using NiBabel (res-T1), and preprocessed into synthetic isotropic images using SynthSR (syn-T1). For the 3D-T1, 2D-T1, res-T1, and syn-T1 images, the recon-all command from FreeSurfer 8.0.0 was used to create parcellations of 174 anatomical regions (equivalent to the 174 regional parcellations provided as part of the DL+DiReCT pipeline), defined by the aseg and Destrieux atlases, and FSL run_first_all was used to segment subcortical surface shapes. The new ML FreeSurfer pipeline, recon-all-clinical, was also tested in the 2D-T1, 3D-T1, and res-T1 images. As a model comparison for SynthSR, the DL+DiReCT pipeline was used to provide segmentations of the 2D-T1 and res-T1 images, including estimates of regional volume and thickness. Spatial overlap and intraclass correlations between the morphometrics of the eight resulting parcellations were first determined, then subcortical surface shape abnormalities associated with IGE were identified by comparing the FSL run_first_all outputs of patients with controls. ResultsWhen standardised to the metrics derived from the 3D-T1 scans, cortical volume and thickness estimates trended lower for the 2D-T1, res-T1, syn-T1, and DL+DiReCT outputs, whereas subcortical volume estimates were more coherent. Dice coefficients revealed an acceptable spatial similarity between the cortices of the 3D-T1 scans and the other images overall, and was higher in the subcortical structures. Intraclass correlation coefficients were consistently lowest when metrics were computed for model-derived inputs, and estimates of thickness were less similar to the ground truth than those of volume. For the people with epilepsy, the 3D-T1 scans showed significant surface deflations across various subcortical structures when compared to healthy controls. Analysis of the 2D-T1 scans enabled the reliable detection of a subset of subcortical abnormalities, whereas analyses of the res-T1 and syn-T1 images were more prone to false-positive results. ConclusionsResampling and ML image synthesis methods do not currently attenuate partial volume effects resulting from low through plane resolution in anisotropic MRI scans, instead quantitative analyses using 2D-T1 scans should be interpreted with caution, and researchers should consider the potential implications of preprocessing. The recon-all-clinical pipeline is promising, but requires further evaluation, especially when considered as an alternative to the classical pipeline. Key PointsO_LISurface deviations indicative of regional atrophy and hypertrophy were identified in people with idiopathic generalised epilepsy. C_LIO_LIPartial volume effects are likely to attenuate subtle morphometric abnormalities, increasing the likelihood of erroneous inference. C_LIO_LIPriors in synthetic image creation models may render them insensitive to subtle biomarkers. C_LIO_LIResampling and machine-learning based image synthesis are not currently replacements for research quality acquisitions in quantitative MRI research. C_LIO_LIThe results of studies using synthetic images should be interpreted in a separate context to those using untransformed data. C_LI

Equivariant Spatiotemporal Transformers with MDL-Guided Feature Selection for Malignancy Detection in Dynamic PET

Dadashkarimi, M.

medrxiv logopreprintAug 6 2025
Dynamic Positron Emission Tomography (PET) scans offer rich spatiotemporal data for detecting malignancies, but their high-dimensionality and noise pose significant challenges. We introduce a novel framework, the Equivariant Spatiotemporal Transformer with MDL-Guided Feature Selection (EST-MDL), which integrates group-theoretic symmetries, Kolmogorov complexity, and Minimum Description Length (MDL) principles. By enforcing spatial and temporal symmetries (e.g., translations and rotations) and leveraging MDL for robust feature selection, our model achieves improved generalization and interpretability. Evaluated on three realworld PET datasets--LUNG-PET, BRAIN-PET, and BREAST-PET--our approach achieves AUCs of 0.94, 0.92, and 0.95, respectively, outperforming CNNs, Vision Transformers (ViTs), and Graph Neural Networks (GNNs) in AUC, sensitivity, specificity, and computational efficiency. This framework offers a robust, interpretable solution for malignancy detection in clinical settings.

The Effectiveness of Large Language Models in Providing Automated Feedback in Medical Imaging Education: A Protocol for a Systematic Review

Al-Mashhadani, M., Ajaz, F., Guraya, S. S., Ennab, F.

medrxiv logopreprintAug 6 2025
BackgroundLarge Language Models (LLMs) represent an ever-emerging and rapidly evolving generative artificial intelligence (AI) modality with promising developments in the field of medical education. LLMs can provide automated feedback services to medical trainees (i.e. medical students, residents, fellows, etc.) and possibly serve a role in medical imaging education. AimThis systematic review aims to comprehensively explore the current applications and educational outcomes of LLMs in providing automated feedback on medical imaging reports. MethodsThis study employs a comprehensive systematic review strategy, involving an extensive search of the literature (Pubmed, Scopus, Embase, and Cochrane), data extraction, and synthesis of the data. ConclusionThis systematic review will highlight the best practices of LLM use in automated feedback of medical imaging reports and guide further development of these models.

ERDES: A Benchmark Video Dataset for Retinal Detachment and Macular Status Classification in Ocular Ultrasound

Pouyan Navard, Yasemin Ozkut, Srikar Adhikari, Elaine Situ-LaCasse, Josie Acuña, Adrienne Yarnish, Alper Yilmaz

arxiv logopreprintAug 5 2025
Retinal detachment (RD) is a vision-threatening condition that requires timely intervention to preserve vision. Macular involvement -- whether the macula is still intact (macula-intact) or detached (macula-detached) -- is the key determinant of visual outcomes and treatment urgency. Point-of-care ultrasound (POCUS) offers a fast, non-invasive, cost-effective, and accessible imaging modality widely used in diverse clinical settings to detect RD. However, ultrasound image interpretation is limited by a lack of expertise among healthcare providers, especially in resource-limited settings. Deep learning offers the potential to automate ultrasound-based assessment of RD. However, there are no ML ultrasound algorithms currently available for clinical use to detect RD and no prior research has been done on assessing macular status using ultrasound in RD cases -- an essential distinction for surgical prioritization. Moreover, no public dataset currently supports macular-based RD classification using ultrasound video clips. We introduce Eye Retinal DEtachment ultraSound, ERDES, the first open-access dataset of ocular ultrasound clips labeled for (i) presence of retinal detachment and (ii) macula-intact versus macula-detached status. The dataset is intended to facilitate the development and evaluation of machine learning models for detecting retinal detachment. We also provide baseline benchmarks using multiple spatiotemporal convolutional neural network (CNN) architectures. All clips, labels, and training code are publicly available at https://osupcvlab.github.io/ERDES/.

CAPoxy: a feasibility study to investigate multispectral imaging in nailfold capillaroscopy

Taylor-Williams, M., Khalil, I., Manning, J., Dinsdale, G., Berks, M., Porcu, L., Wilkinson, S., Bohndiek, S., Murray, A.

medrxiv logopreprintAug 5 2025
BackgroundNailfold capillaroscopy enables visualisation of structural abnormalities in the microvasculature of patients with systemic sclerosis (SSc). The objective of this feasibility study was to determine whether multispectral imaging could provide functional assessment (differences in haemoglobin concentration or oxygenation) of capillaries to aid discrimination between healthy controls and patients with SSc. MSI of nailfold capillaries visualizes the smallest blood vessels and the impact of SSc on angiogenesis and their deformation, making it suitable for evaluating oxygenation-sensitive imaging techniques. Imaging of the nailfold capillaries offers tissue-specific oxygenation information, unlike pulse oximetry, which measures arterial blood oxygenation as a single-point measurement. MethodsThe CAPoxy study was a single-centre, cross-sectional, feasibility study of nailfold capillary multispectral imaging, comparing a cohort of patients with SSc to controls. A nine-band multispectral camera was used to image 22 individuals (10 patients with SSc and 12 controls). Linear mixed-effects models and summary statistics were used to compare the different regions of the nailfold (capillaries, surrounding edges, and outside area) between SSc and controls. A machine learning model was used to compare the two groups. ResultsPatients with SSc exhibited higher indicators of haemoglobin concentration in the capillary and adjacent regions compared to controls, which were significant in the regions surrounding the capillaries (p<0.001). There were also spectral differences between the SSc and controls groups that could indicate differences in oxygenation of the capillaries and surrounding tissue. Additionally, a machine learning model distinguished SSc patients from healthy controls with an accuracy of 84%, suggesting potential for multispectral imaging to classify SSc based on structural and functional microvascular changes. ConclusionsData indicates that multispectral imaging differentiates between patients with SSc from controls based on differences in vascular function. Further work to develop a targeted spectral camera would further improve the contrast between patients with SSc and controls, enabling better imaging. Key messagesMultispectral imaging holds promise for providing functional oxygenation measurement in nailfold capillaroscopy. Significant oxygenation differences between individuals with systemic sclerosis and healthy controls can be detected with multispectral imaging in the tissue surrounding capillaries.

BrainSignsNET: A Deep Learning Model for 3D Anatomical Landmark Detection in the Human Brain Imaging

shirzadeh barough, s., Ventura, C., Bilgel, M., Albert, M., Miller, M. I., Moghekar, A.

medrxiv logopreprintAug 5 2025
Accurate detection of anatomical landmarks in brain Magnetic Resonance Imaging (MRI) scans is essential for reliable spatial normalization, image alignment, and quantitative neuroimaging analyses. In this study, we introduce BrainSignsNET, a deep learning framework designed for robust three-dimensional (3D) landmark detection. Our approach leverages a multi-task 3D convolutional neural network that integrates an attention decoder branch with a multi-class decoder branch to generate precise 3D heatmaps, from which landmark coordinates are extracted. The model was trained and internally validated on T1-weighted Magnetization-Prepared Rapid Gradient-Echo (MPRAGE) scans from the Alzheimers Disease Neuroimaging Initiative (ADNI), the Baltimore Longitudinal Study of Aging (BLSA), and the Biomarkers of Cognitive Decline in Adults at Risk for AD (BIOCARD) datasets and externally validated on a clinical dataset from the Johns Hopkins Hydrocephalus Clinic. The study encompassed 14,472 scans from 6,299 participants, representing a diverse demographic profile with a significant proportion of older adult participants, particularly those over 70 years of age. Extensive preprocessing and data augmentation strategies, including traditional MRI corrections and tailored 3D transformations, ensured data consistency and improved model generalizability. Performance metrics demonstrated that on internal validation BrainSignsNET achieved an overall mean Euclidean distance of 2.32 {+/-} 0.41 mm and 94.8% of landmarks localized within their anatomically defined 3D volumes in the external validation dataset. This improvement in accurate anatomical landmark detection on brain MRI scans should benefit many imaging tasks, including registration, alignment, and quantitative analyses.

MAUP: Training-free Multi-center Adaptive Uncertainty-aware Prompting for Cross-domain Few-shot Medical Image Segmentation

Yazhou Zhu, Haofeng Zhang

arxiv logopreprintAug 5 2025
Cross-domain Few-shot Medical Image Segmentation (CD-FSMIS) is a potential solution for segmenting medical images with limited annotation using knowledge from other domains. The significant performance of current CD-FSMIS models relies on the heavily training procedure over other source medical domains, which degrades the universality and ease of model deployment. With the development of large visual models of natural images, we propose a training-free CD-FSMIS model that introduces the Multi-center Adaptive Uncertainty-aware Prompting (MAUP) strategy for adapting the foundation model Segment Anything Model (SAM), which is trained with natural images, into the CD-FSMIS task. To be specific, MAUP consists of three key innovations: (1) K-means clustering based multi-center prompts generation for comprehensive spatial coverage, (2) uncertainty-aware prompts selection that focuses on the challenging regions, and (3) adaptive prompt optimization that can dynamically adjust according to the target region complexity. With the pre-trained DINOv2 feature encoder, MAUP achieves precise segmentation results across three medical datasets without any additional training compared with several conventional CD-FSMIS models and training-free FSMIS model. The source code is available at: https://github.com/YazhouZhu19/MAUP.

R2GenKG: Hierarchical Multi-modal Knowledge Graph for LLM-based Radiology Report Generation

Futian Wang, Yuhan Qiao, Xiao Wang, Fuling Wang, Yuxiang Zhang, Dengdi Sun

arxiv logopreprintAug 5 2025
X-ray medical report generation is one of the important applications of artificial intelligence in healthcare. With the support of large foundation models, the quality of medical report generation has significantly improved. However, challenges such as hallucination and weak disease diagnostic capability still persist. In this paper, we first construct a large-scale multi-modal medical knowledge graph (termed M3KG) based on the ground truth medical report using the GPT-4o. It contains 2477 entities, 3 kinds of relations, 37424 triples, and 6943 disease-aware vision tokens for the CheXpert Plus dataset. Then, we sample it to obtain multi-granularity semantic graphs and use an R-GCN encoder for feature extraction. For the input X-ray image, we adopt the Swin-Transformer to extract the vision features and interact with the knowledge using cross-attention. The vision tokens are fed into a Q-former and retrieved the disease-aware vision tokens using another cross-attention. Finally, we adopt the large language model to map the semantic knowledge graph, input X-ray image, and disease-aware vision tokens into language descriptions. Extensive experiments on multiple datasets fully validated the effectiveness of our proposed knowledge graph and X-ray report generation framework. The source code of this paper will be released on https://github.com/Event-AHU/Medical_Image_Analysis.

Evaluating the Predictive Value of Preoperative MRI for Erectile Dysfunction Following Radical Prostatectomy

Gideon N. L. Rouwendaal, Daniël Boeke, Inge L. Cox, Henk G. van der Poel, Margriet C. van Dijk-de Haan, Regina G. H. Beets-Tan, Thierry N. Boellaard, Wilson Silva

arxiv logopreprintAug 5 2025
Accurate preoperative prediction of erectile dysfunction (ED) is important for counseling patients undergoing radical prostatectomy. While clinical features are established predictors, the added value of preoperative MRI remains underexplored. We investigate whether MRI provides additional predictive value for ED at 12 months post-surgery, evaluating four modeling strategies: (1) a clinical-only baseline, representing current state-of-the-art; (2) classical models using handcrafted anatomical features derived from MRI; (3) deep learning models trained directly on MRI slices; and (4) multimodal fusion of imaging and clinical inputs. Imaging-based models (maximum AUC 0.569) slightly outperformed handcrafted anatomical approaches (AUC 0.554) but fell short of the clinical baseline (AUC 0.663). Fusion models offered marginal gains (AUC 0.586) but did not exceed clinical-only performance. SHAP analysis confirmed that clinical features contributed most to predictive performance. Saliency maps from the best-performing imaging model suggested a predominant focus on anatomically plausible regions, such as the prostate and neurovascular bundles. While MRI-based models did not improve predictive performance over clinical features, our findings suggest that they try to capture patterns in relevant anatomical structures and may complement clinical predictors in future multimodal approaches.

MedCAL-Bench: A Comprehensive Benchmark on Cold-Start Active Learning with Foundation Models for Medical Image Analysis

Ning Zhu, Xiaochuan Ma, Shaoting Zhang, Guotai Wang

arxiv logopreprintAug 5 2025
Cold-Start Active Learning (CSAL) aims to select informative samples for annotation without prior knowledge, which is important for improving annotation efficiency and model performance under a limited annotation budget in medical image analysis. Most existing CSAL methods rely on Self-Supervised Learning (SSL) on the target dataset for feature extraction, which is inefficient and limited by insufficient feature representation. Recently, pre-trained Foundation Models (FMs) have shown powerful feature extraction ability with a potential for better CSAL. However, this paradigm has been rarely investigated, with a lack of benchmarks for comparison of FMs in CSAL tasks. To this end, we propose MedCAL-Bench, the first systematic FM-based CSAL benchmark for medical image analysis. We evaluate 14 FMs and 7 CSAL strategies across 7 datasets under different annotation budgets, covering classification and segmentation tasks from diverse medical modalities. It is also the first CSAL benchmark that evaluates both the feature extraction and sample selection stages. Our experimental results reveal that: 1) Most FMs are effective feature extractors for CSAL, with DINO family performing the best in segmentation; 2) The performance differences of these FMs are large in segmentation tasks, while small for classification; 3) Different sample selection strategies should be considered in CSAL on different datasets, with Active Learning by Processing Surprisal (ALPS) performing the best in segmentation while RepDiv leading for classification. The code is available at https://github.com/HiLab-git/MedCAL-Bench.
Page 17 of 1061052 results
Show
per page

Ready to Sharpen Your Edge?

Join hundreds of your peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.