Sort by:
Page 6 of 26256 results

Ultrafast Multi-tracer Total-body PET Imaging Using a Transformer-Based Deep Learning Model.

Sun H, Sanaat A, Yi W, Salimi Y, Huang Y, Decorads CE, Castarède I, Wu H, Lu L, Zaidi H

pubmed logopapersAug 29 2025
Reducing PET scan acquisition time to minimize motion-related artifacts and improving patient comfort is always demanding. This study proposes a deep-learning framework for synthesizing diagnostic-quality PET images from ultrafast scans in multi-tracer total-body PET imaging. A retrospective analysis was conducted on clinical uEXPLORER PET/CT datasets from a single institution, including [<sup>18</sup>F]FDG (N=50), [<sup>18</sup>F]FAPI (N=45) and [<sup>68</sup>Ga]FAPI (N=60) studies. Standard 300-s acquisitions were performed for each patient, with ultrafast scan PET images (3, 6, 15, 30, and 40 s) generated through list-mode data truncation. We developed two variants of a 3D SwinUNETR-V2 architecture: Model 1 (PET-only input) and Model 2 (PET+CT fusion input). The proposed methodology was trained and tested on all three datasets using 5-fold cross-validation. The proposed Model 1 and Model 2 significantly enhanced subjective image quality and lesion detectability in multi-tracer PET images compared to the original ultrafast scans. Model 1 and Model 2 also improved objective image quality metrics. For the [¹⁸F]FDG datasets, both approaches improved peak signal-to-noise ratio (PSNR) metrics across ultra-short acquisitions: 3 s: 48.169±6.121 (Model 1) vs. 48.123±6.103 (Model 2) vs. 44.092±7.508 (ultrafast), p < 0.001; 6 s: 48.997±5.960 vs. 48.461±5.897 vs. 46.503±7.190, p < 0.001; 15 s: 50.310±5.674 vs. 50.042±5.734 vs. 49.331±6.732, p < 0.001. The proposed Model 1 and Model 2 effectively enhance image quality of multi-tracer total-body PET scans with ultrafast acquisition times. The predicted PET images demonstrate comparable performance in terms of image quality and lesion detectability.

Deep Learning-Based Generation of DSC MRI Parameter Maps Using Dynamic Contrast-Enhanced MRI Data.

Pei H, Lyu Y, Lambrecht S, Lin D, Feng L, Liu F, Nyquist P, van Zijl P, Knutsson L, Xu X

pubmed logopapersAug 28 2025
Perfusion and perfusion-related parameter maps obtained by using DSC MRI and dynamic contrast-enhanced (DCE) MRI are both useful for clinical diagnosis and research. However, using both DSC and DCE MRI in the same scan session requires 2 doses of gadolinium contrast agent. The objective was to develop deep learning-based methods to synthesize DSC-derived parameter maps from DCE MRI data. Independent analysis of data collected in previous studies was performed. The database contained 64 participants, including patients with and without brain tumors. The reference parameter maps were measured from DSC MRI performed after DCE MRI. A conditional generative adversarial network (cGAN) was designed and trained to generate synthetic DSC-derived maps from DCE MRI data. The median parameter values and distributions between synthetic and real maps were compared by using linear regression and Bland-Altman plots. Using cGAN, realistic DSC parameter maps could be synthesized from DCE MRI data. For controls without brain tumors, the synthesized parameters had distributions similar to the ground truth values. For patients with brain tumors, the synthesized parameters in the tumor region correlated linearly with the ground truth values. In addition, areas not visible due to susceptibility artifacts in real DSC maps could be visualized by using DCE-derived DSC maps. DSC-derived parameter maps could be synthesized by using DCE MRI data, including susceptibility-artifact-prone regions. This shows the potential to obtain both DSC and DCE parameter maps from DCE MRI by using a single dose of contrast agent.

Deep learning-based dual-energy subtraction synthesis from single-energy kV x-ray fluoroscopy for markerless tumor tracking.

Wang J, Ichiji K, Zeng Y, Zhang X, Takai Y, Homma N

pubmed logopapersAug 27 2025
Markerless tumor tracking in x-ray fluoroscopic images is an important technique for achieving precise dose delivery for moving lung tumors during radiation therapy. However, accurate tumor tracking is challenging due to the poor visibility of the target tumor overlapped by other organs such as rib bones. Dual-energy (DE) x-ray fluoroscopy can enhance tracking accuracy with improved tumor visibility by suppressing bones. However, DE x-ray imaging requires special hardware, limiting its clinical use. This study presents a deep learning-based DE subtraction (DES) synthesis method to avoid hardware limitations and enhance tracking accuracy. The proposed method employs a residual U-Net model trained on a simulated DES dataset from a digital phantom to synthesize DES from single-energy (SE) fluoroscopy. Experimental results using a digital phantom showed quantitative evaluation results of synthesis quality. Also, experimental results using clinical SE fluoroscopic images of ten lung cancer patients showed improved tumor tracking accuracy using synthesized DES images, reducing errors from 1.80 to 1.68 mm on average. The tracking success rate within a 25% movement range increased from 50.2% (SE) to 54.9% (DES). These findings indicate the feasibility of deep learning-based DES synthesis for markerless tumor tracking, offering a potential alternative to hardware-dependent DE imaging.

Toward Non-Invasive Voice Restoration: A Deep Learning Approach Using Real-Time MRI

Saleh, M. W.

medrxiv logopreprintAug 26 2025
Despite recent advances in brain-computer interfaces (BCIs) for speech restoration, existing systems remain invasive, costly, and inaccessible to individuals with congenital mutism or neurodegenerative disease. We present a proof-of-concept pipeline that synthesizes personalized speech directly from real-time magnetic resonance imaging (rtMRI) of the vocal tract, without requiring acoustic input. Segmented rtMRI frames are mapped to articulatory class representations using a Pix2Pix conditional GAN, which are then transformed into synthetic audio waveforms by a convolutional neural network modeling the articulatory-to-acoustic relationship. The outputs are rendered into audible form and evaluated with speaker-similarity metrics derived from Resemblyzer embeddings. While preliminary, our results suggest that even silent articulatory motion encodes sufficient information to approximate a speakers vocal characteristics, offering a non-invasive direction for future speech restoration in individuals who have lost or never developed voice.

MRExtrap: Longitudinal Aging of Brain MRIs using Linear Modeling in Latent Space

Jaivardhan Kapoor, Jakob H. Macke, Christian F. Baumgartner

arxiv logopreprintAug 26 2025
Simulating aging in 3D brain MRI scans can reveal disease progression patterns in neurological disorders such as Alzheimer's disease. Current deep learning-based generative models typically approach this problem by predicting future scans from a single observed scan. We investigate modeling brain aging via linear models in the latent space of convolutional autoencoders (MRExtrap). Our approach, MRExtrap, is based on our observation that autoencoders trained on brain MRIs create latent spaces where aging trajectories appear approximately linear. We train autoencoders on brain MRIs to create latent spaces, and investigate how these latent spaces allow predicting future MRIs through linear extrapolation based on age, using an estimated latent progression rate $\boldsymbol{\beta}$. For single-scan prediction, we propose using population-averaged and subject-specific priors on linear progression rates. We also demonstrate that predictions in the presence of additional scans can be flexibly updated using Bayesian posterior sampling, providing a mechanism for subject-specific refinement. On the ADNI dataset, MRExtrap predicts aging patterns accurately and beats a GAN-based baseline for single-volume prediction of brain aging. We also demonstrate and analyze multi-scan conditioning to incorporate subject-specific progression rates. Finally, we show that the latent progression rates in MRExtrap's linear framework correlate with disease and age-based aging patterns from previously studied structural atrophy rates. MRExtrap offers a simple and robust method for the age-based generation of 3D brain MRIs, particularly valuable in scenarios with multiple longitudinal observations.

Reducing radiomics errors in nasopharyngeal cancer via deep learning-based synthetic CT generation from CBCT.

Xiao Y, Lin W, Xie F, Liu L, Zheng G, Xiao C

pubmed logopapersAug 25 2025
This study investigates the impact of cone beam computed tomography (CBCT) image quality on radiomic analysis and evaluates the potential of deep learning-based enhancement to improve radiomic feature accuracy in nasopharyngeal cancer (NPC). The CBAMRegGAN model was trained on 114 paired CT and CBCT datasets from 114 nasopharyngeal cancer patients to enhance CBCT images, with CT images as ground truth. The dataset was split into 82 patients for training, 12 for validation, and 20 for testing. The radiomic features in 6 different categories, including first-order, gray-level co-occurrence matrix (GLCM), gray-level run-length matrix (GLRLM), gray-level size-zone matrix(GLSZM), neighbouring gray tone difference matrix (NGTDM), and gray-level dependence matrix (GLDM), were extracted from the gross tumor volume (GTV) of original CBCT, enhanced CBCT, and CT. Comparing feature errors between original and enhanced CBCT showed that deep learning-based enhancement improves radiomic feature accuracy. The CBAMRegGAN model achieved improved image quality with a peak signal-to-noise ratio (PSNR) of 29.52 ± 2.28 dB, normalized mean absolute error (NMAE) of 0.0129 ± 0.004, and structural similarity index (SSIM) of 0.910 ± 0.025 for enhanced CBCT images. This led to reduced errors in most radiomic features, with average reductions across 20 patients of 19.0%, 24.0%, 3.0%, 19%, 15.0%, and 5.0% for first-order, GLCM, GLRLM, GLSZM, NGTDM, and GLDM features. This study demonstrates that CBCT image quality significantly influences radiomic analysis, and deep learning-based enhancement techniques can effectively improve both image quality and the accuracy of radiomic features in NPC.

Deep learning steganography for big data security using squeeze and excitation with inception architectures.

Issac BM, Kumar SN, Zafar S, Shakil KA, Wani MA

pubmed logopapersAug 25 2025
With the exponential growth of big data in domains such as telemedicine and digital forensics, the secure transmission of sensitive medical information has become a critical concern. Conventional steganographic methods often fail to maintain diagnostic integrity or exhibit robustness against noise and transformations. In this study, we propose a novel deep learning-based steganographic framework that combines Squeeze-and-Excitation (SE) blocks, Inception modules, and residual connections to address these challenges. The encoder integrates dilated convolutions and SE attention to embed secret medical images within natural cover images, while the decoder employs residual and multi-scale Inception-based feature extraction for accurate reconstruction. Designed for deployment on NVIDIA Jetson TX2, the model ensures real-time, low-power operation suitable for edge healthcare applications. Experimental evaluation on MRI and OCT datasets demonstrates the model's efficacy, achieving Peak Signal-to-Noise Ratio (PSNR) values of 39.02 and 38.75, and Structural Similarity Index (SSIM) values of 0.9757, confirming minimal visual distortion. This research contributes to advancing secure, high-capacity steganographic systems for practical use in privacy-sensitive environments.

Radiomics-Driven Diffusion Model and Monte Carlo Compression Sampling for Reliable Medical Image Synthesis.

Zhao J, Li S

pubmed logopapersAug 25 2025
Reliable medical image synthesis is crucial for clinical applications and downstream tasks, where high-quality anatomical structure and predictive confidence are essential. Existing studies have made significant progress by embedding prior conditional knowledge, such as conditional images or textual information, to synthesize natural images. However, medical image synthesis remains a challenging task due to: 1) Data scarcity: High-quality medical text prompt are extremely rare and require specialized expertise. 2) Insufficient uncertainty estimation: The uncertainty estimation is critical for evaluating the confidence of reliable medical image synthesis. This paper presents a novel approach for medical image synthesis, driven by radiomics prompts and combined with Monte Carlo Compression Sampling (MCCS) to ensure reliability. For the first time, our method leverages clinically focused radiomics prompts to condition the generation process, guiding the model to produce reliable medical images. Furthermore, the innovative MCCS algorithm employs Monte Carlo methods to randomly select and compress sampling steps within the denoising diffusion implicit models (DDIM), enabling efficient uncertainty quantification. Additionally, we introduce a MambaTrans architecture to model long-range dependencies in medical images and embed prior conditions (e.g., radiomics prompts). Extensive experiments on benchmark medical imaging datasets demonstrate that our approach significantly improves image quality and reliability, outperforming SoTA methods in both qualitative and quantitative evaluations.

FCR: Investigating Generative AI models for Forensic Craniofacial Reconstruction

Ravi Shankar Prasad, Dinesh Singh

arxiv logopreprintAug 25 2025
Craniofacial reconstruction in forensics is one of the processes to identify victims of crime and natural disasters. Identifying an individual from their remains plays a crucial role when all other identification methods fail. Traditional methods for this task, such as clay-based craniofacial reconstruction, require expert domain knowledge and are a time-consuming process. At the same time, other probabilistic generative models like the statistical shape model or the Basel face model fail to capture the skull and face cross-domain attributes. Looking at these limitations, we propose a generic framework for craniofacial reconstruction from 2D X-ray images. Here, we used various generative models (i.e., CycleGANs, cGANs, etc) and fine-tune the generator and discriminator parts to generate more realistic images in two distinct domains, which are the skull and face of an individual. This is the first time where 2D X-rays are being used as a representation of the skull by generative models for craniofacial reconstruction. We have evaluated the quality of generated faces using FID, IS, and SSIM scores. Finally, we have proposed a retrieval framework where the query is the generated face image and the gallery is the database of real faces. By experimental results, we have found that this can be an effective tool for forensic science.

ControlEchoSynth: Boosting Ejection Fraction Estimation Models via Controlled Video Diffusion

Nima Kondori, Hanwen Liang, Hooman Vaseli, Bingyu Xie, Christina Luong, Purang Abolmaesumi, Teresa Tsang, Renjie Liao

arxiv logopreprintAug 25 2025
Synthetic data generation represents a significant advancement in boosting the performance of machine learning (ML) models, particularly in fields where data acquisition is challenging, such as echocardiography. The acquisition and labeling of echocardiograms (echo) for heart assessment, crucial in point-of-care ultrasound (POCUS) settings, often encounter limitations due to the restricted number of echo views available, typically captured by operators with varying levels of experience. This study proposes a novel approach for enhancing clinical diagnosis accuracy by synthetically generating echo views. These views are conditioned on existing, real views of the heart, focusing specifically on the estimation of ejection fraction (EF), a critical parameter traditionally measured from biplane apical views. By integrating a conditional generative model, we demonstrate an improvement in EF estimation accuracy, providing a comparative analysis with traditional methods. Preliminary results indicate that our synthetic echoes, when used to augment existing datasets, not only enhance EF estimation but also show potential in advancing the development of more robust, accurate, and clinically relevant ML models. This approach is anticipated to catalyze further research in synthetic data applications, paving the way for innovative solutions in medical imaging diagnostics.
Page 6 of 26256 results
Show
per page

Ready to Sharpen Your Edge?

Join hundreds of your peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.