Sort by:
Page 29 of 55548 results

SAM-aware Test-time Adaptation for Universal Medical Image Segmentation

Jianghao Wu, Yicheng Wu, Yutong Xie, Wenjia Bai, You Zhang, Feilong Tang, Yulong Li, Yasmeen George, Imran Razzak

arxiv logopreprintJun 5 2025
Universal medical image segmentation using the Segment Anything Model (SAM) remains challenging due to its limited adaptability to medical domains. Existing adaptations, such as MedSAM, enhance SAM's performance in medical imaging but at the cost of reduced generalization to unseen data. Therefore, in this paper, we propose SAM-aware Test-Time Adaptation (SAM-TTA), a fundamentally different pipeline that preserves the generalization of SAM while improving its segmentation performance in medical imaging via a test-time framework. SAM-TTA tackles two key challenges: (1) input-level discrepancies caused by differences in image acquisition between natural and medical images and (2) semantic-level discrepancies due to fundamental differences in object definition between natural and medical domains (e.g., clear boundaries vs. ambiguous structures). Specifically, our SAM-TTA framework comprises (1) Self-adaptive Bezier Curve-based Transformation (SBCT), which adaptively converts single-channel medical images into three-channel SAM-compatible inputs while maintaining structural integrity, to mitigate the input gap between medical and natural images, and (2) Dual-scale Uncertainty-driven Mean Teacher adaptation (DUMT), which employs consistency learning to align SAM's internal representations to medical semantics, enabling efficient adaptation without auxiliary supervision or expensive retraining. Extensive experiments on five public datasets demonstrate that our SAM-TTA outperforms existing TTA approaches and even surpasses fully fine-tuned models such as MedSAM in certain scenarios, establishing a new paradigm for universal medical image segmentation. Code can be found at https://github.com/JianghaoWu/SAM-TTA.

Exploring Adversarial Watermarking in Transformer-Based Models: Transferability and Robustness Against Defense Mechanism for Medical Images

Rifat Sadik, Tanvir Rahman, Arpan Bhattacharjee, Bikash Chandra Halder, Ismail Hossain

arxiv logopreprintJun 5 2025
Deep learning models have shown remarkable success in dermatological image analysis, offering potential for automated skin disease diagnosis. Previously, convolutional neural network(CNN) based architectures have achieved immense popularity and success in computer vision (CV) based task like skin image recognition, generation and video analysis. But with the emergence of transformer based models, CV tasks are now are nowadays carrying out using these models. Vision Transformers (ViTs) is such a transformer-based models that have shown success in computer vision. It uses self-attention mechanisms to achieve state-of-the-art performance across various tasks. However, their reliance on global attention mechanisms makes them susceptible to adversarial perturbations. This paper aims to investigate the susceptibility of ViTs for medical images to adversarial watermarking-a method that adds so-called imperceptible perturbations in order to fool models. By generating adversarial watermarks through Projected Gradient Descent (PGD), we examine the transferability of such attacks to CNNs and analyze the performance defense mechanism -- adversarial training. Results indicate that while performance is not compromised for clean images, ViTs certainly become much more vulnerable to adversarial attacks: an accuracy drop of as low as 27.6%. Nevertheless, adversarial training raises it up to 90.0%.

Prediction of impulse control disorders in Parkinson's disease: a longitudinal machine learning study

Vamvakas, A., van Balkom, T., van Wingen, G., Booij, J., Weintraub, D., Berendse, H. W., van den Heuvel, O. A., Vriend, C.

medrxiv logopreprintJun 5 2025
BackgroundImpulse control disorders (ICD) in Parkinsons disease (PD) patients mainly occur as adverse effects of dopamine replacement therapy. Despite several known risk factors associated with ICD development, this cannot yet be accurately predicted at PD diagnosis. ObjectivesWe aimed to investigate the predictability of incident ICD by baseline measures of demographic, clinical, dopamine transporter single photon emission computed tomography (DAT-SPECT), and genetic variables. MethodsWe used demographic and clinical data of medication-free PD patients from two longitudinal datasets; Parkinsons Progression Markers Initiative (PPMI) (n=311) and Amsterdam UMC (n=72). We extracted radiomic and latent features from DAT-SPECT. We used single nucleotic polymorphisms (SNPs) from PPMIs NeuroX and Exome sequencing data. Four machine learning classifiers were trained on combinations of the input feature sets, to predict incident ICD at any follow-up assessment. Classification performance was measured with 10x5-fold cross-validation. ResultsICD prevalence at any follow-up was 0.32. The highest performance in predicting incident ICD (AUC=0.66) was achieved by the models trained on clinical features only. Anxiety severity and age of PD onset were identified as the most important features. Performance did not improve with adding features from DAT-SPECT or SNPs. We observed significantly higher performance (AUC=0.74) when classifying patients who developed ICD within four years from diagnosis compared with those tested negative for seven or more years. ConclusionsPrediction accuracy for later ICD development, at the time of PD diagnosis, is limited; however, it increases for shorter time-to-event predictions. Neither DAT-SPECT nor genetic data improve the predictability obtained using demographic and clinical variables alone.

DWI and Clinical Characteristics Correlations in Acute Ischemic Stroke After Thrombolysis

Li, J., Huang, C., Liu, Y., Li, Y., Zhang, J., Xiao, M., yan, Z., zhao, H., Zeng, X., Mu, J.

medrxiv logopreprintJun 5 2025
ObjectiveMagnetic Resonance Diffusion-Weighted Imaging (DWI) is a crucial tool for diagnosing acute ischemic stroke, yet some patients present as DWI-negative. This study aims to analyze the imaging differences and associated clinical characteristics in acute ischemic stroke patients receiving intravenous thrombolysis, in order to enhance understanding of DWI-negative strokes. MethodsRetrospective collection of clinical data from acute ischemic stroke patients receiving intravenous thrombolysis at the Stroke Center of the First Affiliated Hospital of Chongqing Medical University from January 2017 to June 2023, categorized into DWI-positive and negative groups. Descriptive statistics, univariate analysis, binary logistic regression, and machine learning model were utilized to assess the predictive value of clinical features. Additionally, telephone follow-up was conducted for DWI-negative patients to record medication compliance, stroke recurrence, and mortality, with Fine-Gray competing risk model used to analyze recurrent risk factors. ResultsThe incidence rate of DWI-negative ischemic stroke is 22.74%. Factors positively associated with DWI-positive cases include onset to needle time (ONT), onset to first MRI time (OMT), NIHSS score at 1 week of hospitalization (NIHSS-1w), hyperlipidemia (HLP), and atrial fibrillation (AF) (p<0.05, OR>1). Conversely, recurrent ischemic stroke (RIS) and platelet count (PLT) are negatively correlated with DWI-positive cases (p<0.05, OR<1). Trial of Org 10172 in Acute Stroke Treatment (TOAST) classification significantly influences DWI presentation (p=0.01), but the specific impact of etiological subtypes remains unclear. Machine learning models suggest that the features with the highest predictive value, in descending order, are AF, HLP, OMT, ONT, NIHSS difference within 24 hours post-thrombolysis(NIHSS-d(0-24h)PT), and RIS. ConclusionsNIHSS-1w, OMT, ONT, HLP, and AF can predict DWI-positive findings, while platelet count and RIS are associated with DWI-negative cases. AF and HLP demonstrate the highest predictive value. DWI-negative patients have a higher risk of stroke recurrence than mortality in the short term, with a potential correlation between TOAST classification and recurrence risk.

Synthetic multi-inversion time magnetic resonance images for visualization of subcortical structures

Savannah P. Hays, Lianrui Zuo, Anqi Feng, Yihao Liu, Blake E. Dewey, Jiachen Zhuo, Ellen M. Mowry, Scott D. Newsome Jerry L. Prince, Aaron Carass

arxiv logopreprintJun 4 2025
Purpose: Visualization of subcortical gray matter is essential in neuroscience and clinical practice, particularly for disease understanding and surgical planning.While multi-inversion time (multi-TI) T$_1$-weighted (T$_1$-w) magnetic resonance (MR) imaging improves visualization, it is rarely acquired in clinical settings. Approach: We present SyMTIC (Synthetic Multi-TI Contrasts), a deep learning method that generates synthetic multi-TI images using routinely acquired T$_1$-w, T$_2$-weighted (T$_2$-w), and FLAIR images. Our approach combines image translation via deep neural networks with imaging physics to estimate longitudinal relaxation time (T$_1$) and proton density (PD) maps. These maps are then used to compute multi-TI images with arbitrary inversion times. Results: SyMTIC was trained using paired MPRAGE and FGATIR images along with T$_2$-w and FLAIR images. It accurately synthesized multi-TI images from standard clinical inputs, achieving image quality comparable to that from explicitly acquired multi-TI data.The synthetic images, especially for TI values between 400-800 ms, enhanced visualization of subcortical structures and improved segmentation of thalamic nuclei. Conclusion: SyMTIC enables robust generation of high-quality multi-TI images from routine MR contrasts. It generalizes well to varied clinical datasets, including those with missing FLAIR images or unknown parameters, offering a practical solution for improving brain MR image visualization and analysis.

A Comprehensive Study on Medical Image Segmentation using Deep Neural Networks

Loan Dao, Ngoc Quoc Ly

arxiv logopreprintJun 4 2025
Over the past decade, Medical Image Segmentation (MIS) using Deep Neural Networks (DNNs) has achieved significant performance improvements and holds great promise for future developments. This paper presents a comprehensive study on MIS based on DNNs. Intelligent Vision Systems are often evaluated based on their output levels, such as Data, Information, Knowledge, Intelligence, and Wisdom (DIKIW),and the state-of-the-art solutions in MIS at these levels are the focus of research. Additionally, Explainable Artificial Intelligence (XAI) has become an important research direction, as it aims to uncover the "black box" nature of previous DNN architectures to meet the requirements of transparency and ethics. The study emphasizes the importance of MIS in disease diagnosis and early detection, particularly for increasing the survival rate of cancer patients through timely diagnosis. XAI and early prediction are considered two important steps in the journey from "intelligence" to "wisdom." Additionally, the paper addresses existing challenges and proposes potential solutions to enhance the efficiency of implementing DNN-based MIS.

Recent Advances in Medical Image Classification

Loan Dao, Ngoc Quoc Ly

arxiv logopreprintJun 4 2025
Medical image classification is crucial for diagnosis and treatment, benefiting significantly from advancements in artificial intelligence. The paper reviews recent progress in the field, focusing on three levels of solutions: basic, specific, and applied. It highlights advances in traditional methods using deep learning models like Convolutional Neural Networks and Vision Transformers, as well as state-of-the-art approaches with Vision Language Models. These models tackle the issue of limited labeled data, and enhance and explain predictive results through Explainable Artificial Intelligence.

Average Calibration Losses for Reliable Uncertainty in Medical Image Segmentation

Theodore Barfoot, Luis C. Garcia-Peraza-Herrera, Samet Akcay, Ben Glocker, Tom Vercauteren

arxiv logopreprintJun 4 2025
Deep neural networks for medical image segmentation are often overconfident, compromising both reliability and clinical utility. In this work, we propose differentiable formulations of marginal L1 Average Calibration Error (mL1-ACE) as an auxiliary loss that can be computed on a per-image basis. We compare both hard- and soft-binning approaches to directly improve pixel-wise calibration. Our experiments on four datasets (ACDC, AMOS, KiTS, BraTS) demonstrate that incorporating mL1-ACE significantly reduces calibration errors, particularly Average Calibration Error (ACE) and Maximum Calibration Error (MCE), while largely maintaining high Dice Similarity Coefficients (DSCs). We find that the soft-binned variant yields the greatest improvements in calibration, over the Dice plus cross-entropy loss baseline, but often compromises segmentation performance, with hard-binned mL1-ACE maintaining segmentation performance, albeit with weaker calibration improvement. To gain further insight into calibration performance and its variability across an imaging dataset, we introduce dataset reliability histograms, an aggregation of per-image reliability diagrams. The resulting analysis highlights improved alignment between predicted confidences and true accuracies. Overall, our approach not only enhances the trustworthiness of segmentation predictions but also shows potential for safer integration of deep learning methods into clinical workflows. We share our code here: https://github.com/cai4cai/Average-Calibration-Losses

Advancements in Artificial Intelligence Applications for Cardiovascular Disease Research

Yuanlin Mo, Haishan Huang, Bocheng Liang, Weibo Ma

arxiv logopreprintJun 4 2025
Recent advancements in artificial intelligence (AI) have revolutionized cardiovascular medicine, particularly through integration with computed tomography (CT), magnetic resonance imaging (MRI), electrocardiography (ECG) and ultrasound (US). Deep learning architectures, including convolutional neural networks and generative adversarial networks, enable automated analysis of medical imaging and physiological signals, surpassing human capabilities in diagnostic accuracy and workflow efficiency. However, critical challenges persist, including the inability to validate input data accuracy, which may propagate diagnostic errors. This review highlights AI's transformative potential in precision diagnostics while underscoring the need for robust validation protocols to ensure clinical reliability. Future directions emphasize hybrid models integrating multimodal data and adaptive algorithms to refine personalized cardiovascular care.

Personalized MR-Informed Diffusion Models for 3D PET Image Reconstruction

George Webber, Alexander Hammers, Andrew P. King, Andrew J. Reader

arxiv logopreprintJun 4 2025
Recent work has shown improved lesion detectability and flexibility to reconstruction hyperparameters (e.g. scanner geometry or dose level) when PET images are reconstructed by leveraging pre-trained diffusion models. Such methods train a diffusion model (without sinogram data) on high-quality, but still noisy, PET images. In this work, we propose a simple method for generating subject-specific PET images from a dataset of multi-subject PET-MR scans, synthesizing "pseudo-PET" images by transforming between different patients' anatomy using image registration. The images we synthesize retain information from the subject's MR scan, leading to higher resolution and the retention of anatomical features compared to the original set of PET images. With simulated and real [$^{18}$F]FDG datasets, we show that pre-training a personalized diffusion model with subject-specific "pseudo-PET" images improves reconstruction accuracy with low-count data. In particular, the method shows promise in combining information from a guidance MR scan without overly imposing anatomical features, demonstrating an improved trade-off between reconstructing PET-unique image features versus features present in both PET and MR. We believe this approach for generating and utilizing synthetic data has further applications to medical imaging tasks, particularly because patient-specific PET images can be generated without resorting to generative deep learning or large training datasets.
Page 29 of 55548 results
Show
per page

Ready to Sharpen Your Edge?

Join hundreds of your peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.