Sort by:
Page 79 of 3493486 results

Effect of Data Augmentation on Conformal Prediction for Diabetic Retinopathy

Rizwan Ahamed, Annahita Amireskandari, Joel Palko, Carol Laxson, Binod Bhattarai, Prashnna Gyawali

arxiv logopreprintAug 19 2025
The clinical deployment of deep learning models for high-stakes tasks such as diabetic retinopathy (DR) grading requires demonstrable reliability. While models achieve high accuracy, their clinical utility is limited by a lack of robust uncertainty quantification. Conformal prediction (CP) offers a distribution-free framework to generate prediction sets with statistical guarantees of coverage. However, the interaction between standard training practices like data augmentation and the validity of these guarantees is not well understood. In this study, we systematically investigate how different data augmentation strategies affect the performance of conformal predictors for DR grading. Using the DDR dataset, we evaluate two backbone architectures -- ResNet-50 and a Co-Scale Conv-Attentional Transformer (CoaT) -- trained under five augmentation regimes: no augmentation, standard geometric transforms, CLAHE, Mixup, and CutMix. We analyze the downstream effects on conformal metrics, including empirical coverage, average prediction set size, and correct efficiency. Our results demonstrate that sample-mixing strategies like Mixup and CutMix not only improve predictive accuracy but also yield more reliable and efficient uncertainty estimates. Conversely, methods like CLAHE can negatively impact model certainty. These findings highlight the need to co-design augmentation strategies with downstream uncertainty quantification in mind to build genuinely trustworthy AI systems for medical imaging.

Comparing Conditional Diffusion Models for Synthesizing Contrast-Enhanced Breast MRI from Pre-Contrast Images

Sebastian Ibarra, Javier del Riego, Alessandro Catanese, Julian Cuba, Julian Cardona, Nataly Leon, Jonathan Infante, Karim Lekadir, Oliver Diaz, Richard Osuala

arxiv logopreprintAug 19 2025
Dynamic contrast-enhanced (DCE) MRI is essential for breast cancer diagnosis and treatment. However, its reliance on contrast agents introduces safety concerns, contraindications, increased cost, and workflow complexity. To this end, we present pre-contrast conditioned denoising diffusion probabilistic models to synthesize DCE-MRI, introducing, evaluating, and comparing a total of 22 generative model variants in both single-breast and full breast settings. Towards enhancing lesion fidelity, we introduce both tumor-aware loss functions and explicit tumor segmentation mask conditioning. Using a public multicenter dataset and comparing to respective pre-contrast baselines, we observe that subtraction image-based models consistently outperform post-contrast-based models across five complementary evaluation metrics. Apart from assessing the entire image, we also separately evaluate the region of interest, where both tumor-aware losses and segmentation mask inputs improve evaluation metrics. The latter notably enhance qualitative results capturing contrast uptake, albeit assuming access to tumor localization inputs that are not guaranteed to be available in screening settings. A reader study involving 2 radiologists and 4 MRI technologists confirms the high realism of the synthetic images, indicating an emerging clinical potential of generative contrast-enhancement. We share our codebase at https://github.com/sebastibar/conditional-diffusion-breast-MRI.

Latent Interpolation Learning Using Diffusion Models for Cardiac Volume Reconstruction

Niklas Bubeck, Suprosanna Shit, Chen Chen, Can Zhao, Pengfei Guo, Dong Yang, Georg Zitzlsberger, Daguang Xu, Bernhard Kainz, Daniel Rueckert, Jiazhen Pan

arxiv logopreprintAug 19 2025
Cardiac Magnetic Resonance (CMR) imaging is a critical tool for diagnosing and managing cardiovascular disease, yet its utility is often limited by the sparse acquisition of 2D short-axis slices, resulting in incomplete volumetric information. Accurate 3D reconstruction from these sparse slices is essential for comprehensive cardiac assessment, but existing methods face challenges, including reliance on predefined interpolation schemes (e.g., linear or spherical), computational inefficiency, and dependence on additional semantic inputs such as segmentation labels or motion data. To address these limitations, we propose a novel \textbf{Ca}rdiac \textbf{L}atent \textbf{I}nterpolation \textbf{D}iffusion (CaLID) framework that introduces three key innovations. First, we present a data-driven interpolation scheme based on diffusion models, which can capture complex, non-linear relationships between sparse slices and improves reconstruction accuracy. Second, we design a computationally efficient method that operates in the latent space and speeds up 3D whole-heart upsampling time by a factor of 24, reducing computational overhead compared to previous methods. Third, with only sparse 2D CMR images as input, our method achieves SOTA performance against baseline methods, eliminating the need for auxiliary input such as morphological guidance, thus simplifying workflows. We further extend our method to 2D+T data, enabling the effective modeling of spatiotemporal dynamics and ensuring temporal coherence. Extensive volumetric evaluations and downstream segmentation tasks demonstrate that CaLID achieves superior reconstruction quality and efficiency. By addressing the fundamental limitations of existing approaches, our framework advances the state of the art for spatio and spatiotemporal whole-heart reconstruction, offering a robust and clinically practical solution for cardiovascular imaging.

Automated surgical planning with nnU-Net: delineation of the anatomy in hepatobiliary phase MRI

Karin A. Olthof, Matteo Fusagli, Bianca Güttner, Tiziano Natali, Bram Westerink, Stefanie Speidel, Theo J. M. Ruers, Koert F. D. Kuhlmann, Andrey Zhylka

arxiv logopreprintAug 19 2025
Background: The aim of this study was to develop and evaluate a deep learning-based automated segmentation method for hepatic anatomy (i.e., parenchyma, tumors, portal vein, hepatic vein and biliary tree) from the hepatobiliary phase of gadoxetic acid-enhanced MRI. This method should ease the clinical workflow of preoperative planning. Methods: Manual segmentation was performed on hepatobiliary phase MRI scans from 90 consecutive patients who underwent liver surgery between January 2020 and October 2023. A deep learning network (nnU-Net v1) was trained on 72 patients with an extra focus on thin structures and topography preservation. Performance was evaluated on an 18-patient test set by comparing automated and manual segmentations using Dice similarity coefficient (DSC). Following clinical integration, 10 segmentations (assessment dataset) were generated using the network and manually refined for clinical use to quantify required adjustments using DSC. Results: In the test set, DSCs were 0.97+/-0.01 for liver parenchyma, 0.80+/-0.04 for hepatic vein, 0.79+/-0.07 for biliary tree, 0.77+/-0.17 for tumors, and 0.74+/-0.06 for portal vein. Average tumor detection rate was 76.6+/-24.1%, with a median of one false-positive per patient. The assessment dataset showed minor adjustments were required for clinical use of the 3D models, with high DSCs for parenchyma (1.00+/-0.00), portal vein (0.98+/-0.01) and hepatic vein (0.95+/-0.07). Tumor segmentation exhibited greater variability (DSC 0.80+/-0.27). During prospective clinical use, the model detected three additional tumors initially missed by radiologists. Conclusions: The proposed nnU-Net-based segmentation method enables accurate and automated delineation of hepatic anatomy. This enables 3D planning to be applied efficiently as a standard-of-care for every patient undergoing liver surgery.

Fracture Detection and Localisation in Wrist and Hand Radiographs using Detection Transformer Variants

Aditya Bagri, Vasanthakumar Venugopal, Anandakumar D, Revathi Ezhumalai, Kalyan Sivasailam, Bargava Subramanian, VarshiniPriya, Meenakumari K S, Abi M, Renita S

arxiv logopreprintAug 19 2025
Background: Accurate diagnosis of wrist and hand fractures using radiographs is essential in emergency care, but manual interpretation is slow and prone to errors. Transformer-based models show promise in improving medical image analysis, but their application to extremity fractures is limited. This study addresses this gap by applying object detection transformers to wrist and hand X-rays. Methods: We fine-tuned the RT-DETR and Co-DETR models, pre-trained on COCO, using over 26,000 annotated X-rays from a proprietary clinical dataset. Each image was labeled for fracture presence with bounding boxes. A ResNet-50 classifier was trained on cropped regions to refine abnormality classification. Supervised contrastive learning was used to enhance embedding quality. Performance was evaluated using AP@50, precision, and recall metrics, with additional testing on real-world X-rays. Results: RT-DETR showed moderate results (AP@50 = 0.39), while Co-DETR outperformed it with an AP@50 of 0.615 and faster convergence. The integrated pipeline achieved 83.1% accuracy, 85.1% precision, and 96.4% recall on real-world X-rays, demonstrating strong generalization across 13 fracture types. Visual inspection confirmed accurate localization. Conclusion: Our Co-DETR-based pipeline demonstrated high accuracy and clinical relevance in wrist and hand fracture detection, offering reliable localization and differentiation of fracture types. It is scalable, efficient, and suitable for real-time deployment in hospital workflows, improving diagnostic speed and reliability in musculoskeletal radiology.

Deep Learning-Enhanced Opportunistic Osteoporosis Screening in 100 kV Low-Voltage Chest CT: A Novel Way Toward Bone Mineral Density Measurement and Radiation Dose Reduction.

Li Y, Ye K, Liu S, Zhang Y, Jin D, Jiang C, Ni M, Zhang M, Qian Z, Wu W, Pan X, Yuan H

pubmed logopapersAug 19 2025
To explore the feasibility and accuracy of a deep learning (DL) method for fully automated vertebral body (VB) segmentation, region of interest (ROI) extraction, and bone mineral density (BMD) calculation using 100kV low-voltage chest CT performed for lung cancer screening across various scanners from different manufacturers and hospitals. This study included 1167 patients who underwent 100 kV low-voltage chest and 120 kV lumbar CT from October 2022 to August 2024. Patients were divided into a training set (495 patients), a validation set (169 patients), and three test sets (245, 128, and 130 patients). The DL framework comprised four convolutional neural networks (CNNs): 3D VB-Net and SCN for automated VB segmentation and ROI extraction, and DenseNet and ResNet for BMD calculation of target VBs (T12-L2). The BMD values of 120 kV QCT were identified as reference data. Linear regression and BlandAltman analyses were used to compare the BMD values between 120 kV QCT and 100 kV CNNs and 100 kV QCT. Receiver operating characteristic curve analysis was used to evaluate the diagnostic performance of 100 kV CNNs and 100 kV QCT for osteoporosis and low BMD from normal BMD. For three test sets, linear regression and BlandAltman analyses revealed a stronger correlation (R<sup>2</sup> = 0.970-0.994 and 0.968-0.986, P < .001) and better agreement (mean error, -2.24 to 1.52 and 2.72 to 3.06 mg/cm<sup>3</sup>) for the BMD between the 120 kV QCT and 100 kV CNNs than between the 120 kV and 100 kV QCT. The areas under the curve of the 100 kV CNNs and 100 kV QCT were 1.000 and 0.999-1.000, and 1.000 and 1.000 for detecting osteoporosis and low BMD from normal BMD, respectively. The DL method achieved high accuracy for fully automated osteoporosis screening in 100 kV low-voltage chest CT scans obtained for lung cancer screening and performed well on various scanners from different manufacturers and hospitals.

Objective Task-Based Evaluation of Quantitative Medical Imaging Methods: Emerging Frameworks and Future Directions.

Liu Y, Xia H, Obuchowski NA, Laforest R, Rahmim A, Siegel BA, Jha AK

pubmed logopapersAug 19 2025
Quantitative imaging (QI) holds significant potential across diverse clinical applications. For clinical translation of QI, rigorous evaluation on clinically relevant tasks is essential. This article outlines 4 emerging evaluation frameworks, including virtual imaging trials, evaluation with clinical data in the absence of ground truth, evaluation for joint detection and quantification tasks, and evaluation of QI methods that output multidimensional outputs. These frameworks are presented in the context of recent advancements in PET, such as long axial field of view PET and the development of artificial intelligence algorithms for PET. We conclude by discussing future research directions for evaluating QI methods.

A state-of-the-art new method for diagnosing atrial septal defects with origami technique augmented dataset and a column-based statistical feature extractor.

Yaman I, Kilic I, Yaman O, Poyraz F, Erdem Kaya E, Ozgur Baris V, Ciris S

pubmed logopapersAug 19 2025
Early diagnosis of atrial septal defects (ASDs) from chest X-ray (CXR) images with high accuracy is vital. This study created a dataset from chest X-ray images obtained from different adult subjects. To diagnose atrial septal defects with very high accuracy, which we call state-of-the-art technology, the method known as the Origami paper folding technique, which was used for the first time in the literature on our dataset, was used for data augmentation. Two different augmented data sets were obtained using the Origami technique. The mean, standard deviation, median, variance, and skewness statistical values were obtained column-wise on the images in these data sets. These features were classified with a Support vector machine (SVM). The results obtained using the support vector machine were evaluated according to the k-nearest neighbors (k-NN) and decision tree classifiers for comparison. The results obtained from the classification of the data sets augmented with the Origami technique with the support vector machine (SVM) are state-of-the-art (99.69 %). Our study has provided a clear superiority over deep learning-based artificial intelligence methods.

A Cardiac-specific CT Foundation Model for Heart Transplantation

Xu, H., Woicik, A., Asadian, S., Shen, J., Zhang, Z., Nabipoor, A., Musi, J. P., Keenan, J., Khorsandi, M., Al-Alao, B., Dimarakis, I., Chalian, H., Lin, Y., Fishbein, D., Pal, J., Wang, S., Lin, S.

medrxiv logopreprintAug 19 2025
Heart failure is a major cause of morbitidy and mortality, with the severest forms requiring heart transplantation. Heart size matching between the donor and recipient is a critical step in ensuring a successful transplantation. Currently, a set of equations based on population measures of height, weight, sex and age, viz. predicted heart mass (PHM), are used but can be improved upon by personalized information from recipient and donor chest CT images. Here, we developed GigaHeart, the first heart-specific foundation model pretrained on 180,897 chest CT volumes from 56,607 patients. The key idea of GigaHeart is to direct the foundation models attention towards the heart by contrasting the heart region and the entire chest, thereby encouraging the model to capture fine-grained cardiac features. GigaHeart achieves the best performance on 8 cardiac-specific classification tasks and further, exhibits superior performance on cross-modal tasks by jointly modeling CT images and reports. We similarly developed a thorax-specific foundation model and observed promising performance on 9 thorax-specific tasks, indicating the potential to extend GigaHeart to other organ-specific foundation models. More importantly, GigaHeart addresses the heart sizing problem. It avoids oversizing by correctly segmenting the sizes of hearts of donors and recipients. In regressions against actual heart masses, our AI-segmented total cardiac volumes (TCVs) has a 33.3% R2 improvement when compared to PHM. Meanwhile, GigaHeart also solves the undersizing problem by adding a regression layer to the model. Specifically, GigaHeart reduces the mean squared error by 57% against PHM. In total, we show that GigaHeart increases the acceptable range of donor heart sizes and matches more accurately than the widely used PHM equations. In all, GigaHeart is a state-of-the-art, cardiac-specific foundation model with the key innovation of directing the models attention to the heart. GigaHeart can be finetuned for accomplishing a number of tasks accurately, of which AI-assisted heart sizing is a novel example.

Interpreting convolutional neural network explainability for head-and-neck cancer radiotherapy organ-at-risk segmentation.

Strijbis VIJ, Gurney-Champion OJ, Grama DI, Slotman BJ, Verbakel WFAR

pubmed logopapersAug 19 2025
Convolutional neural networks (CNNs) have emerged to reduce clinical resources and standardize auto-contouring of organs-at-risk (OARs). Although CNNs perform adequately for most patients, understanding when the CNN might fail is critical for effective and safe clinical deployment. However, the limitations of CNNs are poorly understood because of their black-box nature. Explainable artificial intelligence (XAI) can expose CNNs' inner mechanisms for classification. Here, we investigate the inner mechanisms of CNNs for segmentation and explore a novel, computational approach to a-priori flag potentially insufficient parotid gland (PG) contours. First, 3D UNets were trained in three PG segmentation situations using (1) synthetic cases; (2) 1925 clinical computed tomography (CT) scans with typical and (3) more consistent contours curated through a previously validated auto-curation step. Then, we generated attribution maps for seven XAI methods, and qualitatively assessed them for congruency between simulated and clinical contours, and how much XAI agreed with expert reasoning. To objectify observations, we explored persistent homology intensity filtrations to capture essential topological characteristics of XAI attributions. Principal component (PC) eigenvalues of Euler characteristic profiles were correlated with spatial agreement (Dice-Sørensen similarity coefficient; DSC). Evaluation was done using sensitivity, specificity and the area under receiver operating characteristic (AUROC) curve on an external AAPM dataset, where as proof-of-principle, we regard the lowest 15% DSC as insufficient. PatternNet attributions (PNet-A) focused on soft-tissue structures, whereas guided backpropagation (GBP) highlighted both soft-tissue and high-density structures (e.g. mandible bone), which was congruent with synthetic situations. Both methods typically had higher/denser activations in better auto-contoured medial and anterior lobes. Curated models produced "cleaner" gradient class-activation mapping (GCAM) attributions. Quantitative analysis showed that PCλ<sub>1</sub> of guided GCAM's (GGCAM) Euler characteristic (EC) profile had good predictive value (sensitivity>0.85, specificity>0.90) of DSC for AAPM cases, with AUROC = 0.66, 0.74, 0.94, 0.83 for GBP, GCAM, GGCAM and PNet-A. For for λ<sub>1</sub> < -1.8e3 of GGCAM's EC-profile, 87% of cases were insufficient. GBP and PNet-A qualitatively agreed most with expert reasoning on directly (structure borders) and indirectly (proxies used for identifying structure borders) important features for PG segmentation. Additionally, this work investigated as proof-of-principle how topological data analysis could be used for quantitative XAI signal analysis to a-priori mark potentially inadequate CNN-segmentations, using only features from inside the predicted PG. This work used PG as a well-understood segmentation paradigm and may extend to target volumes and other organs-at-risk.
Page 79 of 3493486 results
Show
per page

Ready to Sharpen Your Edge?

Join hundreds of your peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.