Sort by:
Page 46 of 1651650 results

Unsupervised Cardiac Video Translation Via Motion Feature Guided Diffusion Model

Swakshar Deb, Nian Wu, Frederick H. Epstein, Miaomiao Zhang

arxiv logopreprintJul 1 2025
This paper presents a novel motion feature guided diffusion model for unpaired video-to-video translation (MFD-V2V), designed to synthesize dynamic, high-contrast cine cardiac magnetic resonance (CMR) from lower-contrast, artifact-prone displacement encoding with stimulated echoes (DENSE) CMR sequences. To achieve this, we first introduce a Latent Temporal Multi-Attention (LTMA) registration network that effectively learns more accurate and consistent cardiac motions from cine CMR image videos. A multi-level motion feature guided diffusion model, equipped with a specialized Spatio-Temporal Motion Encoder (STME) to extract fine-grained motion conditioning, is then developed to improve synthesis quality and fidelity. We evaluate our method, MFD-V2V, on a comprehensive cardiac dataset, demonstrating superior performance over the state-of-the-art in both quantitative metrics and qualitative assessments. Furthermore, we show the benefits of our synthesized cine CMRs improving downstream clinical and analytical tasks, underscoring the broader impact of our approach. Our code is publicly available at https://github.com/SwaksharDeb/MFD-V2V.

Leveraging multithreading on edge computing for smart healthcare based on intelligent multimodal classification approach.

Alghareb FS, Hasan BT

pubmed logopapersJul 1 2025
Medical digitization has been intensively developed in the last decade, leading to paving the path for computer-aided medical diagnosis research. Thus, anomaly detection based on machine and deep learning techniques has been extensively employed in healthcare applications, such as medical imaging classification and monitoring of patients' vital signs. To effectively leverage digitized medical records for identifying challenges in healthcare, this manuscript presents a smart Clinical Decision Support System (CDSS) dedicated for medical multimodal data automated diagnosis. A smart healthcare system necessitating medical data management and decision-making is proposed. To deliver timely rapid diagnosis, thread-level parallelism (TLP) is utilized for parallel distribution of classification tasks on three edge computing devices, each employing an AI module for on-device AI classifications. In comparison to existing machine and deep learning classification techniques, the proposed multithreaded architecture realizes a hybrid (ML and DL) processing module on each edge node. In this context, the presented edge computing-based parallel architecture captures a high level of parallelism, tailored for dealing with multiple categories of medical records. The cluster of the proposed architecture encompasses three edge computing Raspberry Pi devices and an edge server. Furthermore, lightweight neural networks, such as MobileNet, EfficientNet, and ResNet18, are trained and optimized based on genetic algorithms to provide classification of brain tumor, pneumonia, and colon cancer. Model deployment was conducted based on Python programming, where PyCharm is run on the edge server whereas Thonny is installed on edge nodes. In terms of accuracy, the proposed GA-based optimized ResNet18 for pneumonia diagnosis achieves 93.59% predictive accuracy and reduces the classifier computation complexity by 33.59%, whereas an outstanding accuracy of 99.78% and 100% were achieved with EfficientNet-v2 for brain tumor and colon cancer prediction, respectively, while both models preserving a reduction of 25% in the model's classifier. More importantly, an inference speedup of 28.61% and 29.08% was obtained by implementing parallel 2 DL and 3 DL threads configurations compared to the sequential implementation, respectively. Thus, the proposed multimodal-multithreaded architecture offers promising prospects for comprehensive and accurate anomaly detection of patients' medical imaging and vital signs. To summarize, our proposed architecture contributes to the advancement of healthcare services, aiming to improve patient medical diagnosis and therapy outcomes.

Dynamic frame-by-frame motion correction for 18F-flurpiridaz PET-MPI using convolution neural network

Urs, M., Killekar, A., Builoff, V., Lemley, M., Wei, C.-C., Ramirez, G., Kavanagh, P., Buckley, C., Slomka, P. J.

medrxiv logopreprintJul 1 2025
PurposePrecise quantification of myocardial blood flow (MBF) and flow reserve (MFR) in 18F-flurpiridaz PET significantly relies on motion correction (MC). However, the manual frame-by-frame correction leads to significant inter-observer variability, time-consuming, and requires significant experience. We propose a deep learning (DL) framework for automatic MC of 18F-flurpiridaz PET. MethodsThe method employs a 3D ResNet based architecture that takes 3D PET volumes and outputs motion vectors. It was validated using 5-fold cross-validation on data from 32 sites of a Phase III clinical trial (NCT01347710). Manual corrections from two experienced operators served as ground truth, and data augmentation using simulated vectors enhanced training robustness. The study compared the DL approach to both manual and standard non-AI automatic MC methods, assessing agreement and diagnostic accuracy using minimal segmental MBF and MFR. ResultsThe area under the receiver operating characteristic curves (AUC) for significant CAD were comparable between DL-MC MBF, manual-MC MBF from Operators (AUC=0.897,0.892 and 0.889, respectively; p>0.05), standard non-AI automatic MC (AUC=0.877; p>0.05) and significantly higher than No-MC (AUC=0.835; p<0.05). Similar findings were observed with MFR. The 95% confidence limits for agreement with the operator were {+/-}0.49ml/g/min (mean difference = 0.00) for MFR and {+/-}0.24ml/g/min (mean difference = 0.00) for MBF. ConclusionDL-MC is significantly faster but diagnostically comparable to manual-MC. The quantitative results obtained with DL-MC for MBF and MFR are in excellent agreement with those manually corrected by experienced operators compared to standard non-AI automatic MC in patients undergoing 18F-flurpiridaz PET-MPI.

Unsupervised Cardiac Video Translation Via Motion Feature Guided Diffusion Model

Swakshar Deb, Nian Wu, Frederick H. Epstein, Miaomiao Zhang

arxiv logopreprintJul 1 2025
This paper presents a novel motion feature guided diffusion model for unpaired video-to-video translation (MFD-V2V), designed to synthesize dynamic, high-contrast cine cardiac magnetic resonance (CMR) from lower-contrast, artifact-prone displacement encoding with stimulated echoes (DENSE) CMR sequences. To achieve this, we first introduce a Latent Temporal Multi-Attention (LTMA) registration network that effectively learns more accurate and consistent cardiac motions from cine CMR image videos. A multi-level motion feature guided diffusion model, equipped with a specialized Spatio-Temporal Motion Encoder (STME) to extract fine-grained motion conditioning, is then developed to improve synthesis quality and fidelity. We evaluate our method, MFD-V2V, on a comprehensive cardiac dataset, demonstrating superior performance over the state-of-the-art in both quantitative metrics and qualitative assessments. Furthermore, we show the benefits of our synthesized cine CMRs improving downstream clinical and analytical tasks, underscoring the broader impact of our approach. Our code is publicly available at https://github.com/SwaksharDeb/MFD-V2V.

MedGround-R1: Advancing Medical Image Grounding via Spatial-Semantic Rewarded Group Relative Policy Optimization

Huihui Xu, Yuanpeng Nie, Hualiang Wang, Ying Chen, Wei Li, Junzhi Ning, Lihao Liu, Hongqiu Wang, Lei Zhu, Jiyao Liu, Xiaomeng Li, Junjun He

arxiv logopreprintJul 1 2025
Medical Image Grounding (MIG), which involves localizing specific regions in medical images based on textual descriptions, requires models to not only perceive regions but also deduce spatial relationships of these regions. Existing Vision-Language Models (VLMs) for MIG often rely on Supervised Fine-Tuning (SFT) with large amounts of Chain-of-Thought (CoT) reasoning annotations, which are expensive and time-consuming to acquire. Recently, DeepSeek-R1 demonstrated that Large Language Models (LLMs) can acquire reasoning abilities through Group Relative Policy Optimization (GRPO) without requiring CoT annotations. In this paper, we adapt the GRPO reinforcement learning framework to VLMs for Medical Image Grounding. We propose the Spatial-Semantic Rewarded Group Relative Policy Optimization to train the model without CoT reasoning annotations. Specifically, we introduce Spatial-Semantic Rewards, which combine spatial accuracy reward and semantic consistency reward to provide nuanced feedback for both spatially positive and negative completions. Additionally, we propose to use the Chain-of-Box template, which integrates visual information of referring bounding boxes into the <think> reasoning process, enabling the model to explicitly reason about spatial regions during intermediate steps. Experiments on three datasets MS-CXR, ChestX-ray8, and M3D-RefSeg demonstrate that our method achieves state-of-the-art performance in Medical Image Grounding. Ablation studies further validate the effectiveness of each component in our approach. Code, checkpoints, and datasets are available at https://github.com/bio-mlhui/MedGround-R1

Prediction Crohn's Disease Activity Using Computed Tomography Enterography-Based Radiomics and Serum Markers.

Wang P, Liu Y, Wang Y

pubmed logopapersJun 30 2025
Accurate stratification of the activity index of Crohn's disease (CD) using computed tomography enterography (CTE) radiomics and serum markers can aid in predicting disease progression and assist physicians in personalizing therapeutic regimens for patients with CD. This retrospective study enrolled 233 patients diagnosed with CD between January 2019 and August 2024. Patients were divided into training and testing cohorts at a ratio of 7:3 and further categorized into remission, mild active phase, and moderate-severe active phase groups based on simple endoscopic score for CD (SEC-CD). Radiomics features were extracted from CTE venous images, and T-test and least absolute shrinkage and selection operator (LASSO) regression were applied for feature selection. The serum markers were selected based on the variance analysis. We also developed a random forest (RF) model for multi-class stratification of CD. The model performance was evaluated by the area under the receiver operating characteristic curve (AUC) and quantified the contribution of each feature in the dataset to CD activity via Shapley additive exPlanations (SHAP) values. Finally, we enrolled gender, radiomics scores, and serum scores to develop a nomogram model to verify the effectiveness of feature extraction. 14 non-zero coefficient radiomics features and six serum markers with significant differences (P<0.01) were ultimately selected to predict CD activity. The AUC (micro/macro) for the ensemble machine learning model combining the radiomics features and serum markers is 0.931/0.928 for three-class. The AUC for the remission phase, the mild active phase, and the moderate-severe active phase were 0.983, 0.852, and 0.917, respectively. The mean AUC for the nomogram model was 0.940. A radiomics model was developed by integrating radiomics and serum markers of CD patients, achieving enhanced consistency with SEC-CD in grade CD. This model has the potential to assist clinicians in accurate diagnosis and treatment.

Machine learning methods for sex estimation of sub-adults using cranial computed tomography images.

Syed Mohd Hamdan SN, Faizal Abdullah ERM, Wen KJ, Al-Adawiyah Rahmat R, Wan Ibrahim WI, Abd Kadir KA, Ibrahim N

pubmed logopapersJun 30 2025
This research aimed to compare the classification accuracy of three machine learning (ML) methods (random forest (RF), support vector machines (SVM), linear discriminant analysis (LDA)) for sex estimation of sub-adults using cranial computed tomography (CCT) images. A total of 521 CCT scans from sub-adult Malaysians aged 0 to 20 were analysed using Mimics software (Materialise Mimics Ver. 21). Plane-to-plane (PTP) protocol was used for measuring 14 chosen craniometric parameters. A trio of machine learning algorithms RF, SVM, and LDA with GridSearchCV was used to produce classification models for sex estimation. In addition, performance was measured in the form of accuracy, precision, recall, and F1-score, among others. RF produced testing accuracy of 73%, with the best hyperparameters of max_depth = 6, max_samples = 40, and n_estimators = 45. SVM obtained an accuracy of 67% with the best hyperparameters: learning rate (C) = 10, gamma = 0.01, and kernel = radial basis function (RBF). LDA obtained the lowest accuracy of 65% with shrinkage of 0.02. Among the tested ML methods, RF showed the highest testing accuracy in comparison to SVM and LDA. This is the first AI-based classification model that can be used for estimating sex in sub-adults using CCT scans.

Automatic Multiclass Tissue Segmentation Using Deep Learning in Brain MR Images of Tumor Patients.

Kandpal A, Kumar P, Gupta RK, Singh A

pubmed logopapersJun 30 2025
Precise delineation of brain tissues, including lesions, in MR images is crucial for data analysis and objectively assessing conditions like neurological disorders and brain tumors. Existing methods for tissue segmentation often fall short in addressing patients with lesions, particularly those with brain tumors. This study aimed to develop and evaluate a robust pipeline utilizing convolutional neural networks for rapid and automatic segmentation of whole brain tissues, including tumor lesions. The proposed pipeline was developed using BraTS'21 data (1251 patients) and tested on local hospital data (100 patients). Ground truth masks for lesions as well as brain tissues were generated. Two convolutional neural networks based on deep residual U-Net framework were trained for segmenting brain tissues and tumor lesions. The performance of the pipeline was evaluated on independent test data using dice similarity coefficient (DSC) and volume similarity (VS). The proposed pipeline achieved a mean DSC of 0.84 and a mean VS of 0.93 on the BraTS'21 test data set. On the local hospital test data set, it attained a mean DSC of 0.78 and a mean VS of 0.91. The proposed pipeline also generated satisfactory masks in cases where the SPM12 software performed inadequately. The proposed pipeline offers a reliable and automatic solution for segmenting brain tissues and tumor lesions in MR images. Its adaptability makes it a valuable tool for both research and clinical applications, potentially streamlining workflows and enhancing the precision of analyses in neurological and oncological studies.

Statistical Toolkit for Analysis of Radiotherapy DICOM Data.

Kinz M, Molodowitch C, Killoran J, Hesser JW, Zygmanski P

pubmed logopapersJun 30 2025
&#xD;Radiotherapy (RT) has become increasingly sophisticated, necessitating advanced tools for analyzing extensive treatment data in hospital databases. Such analyses can enhance future treatments, particularly through Knowledge-Based Planning, and aid in developing new treatment modalities like convergent kV RT.&#xD;Purpose: The objective is to develop automated software tools for large-scale retrospective analysis of over 10,000 MeV x-ray radiotherapy plans. This aims to identify trends and references in plans delivered at our institution across all treatment sites, focusing on: (A) Planning-Target-Volume, Clinical-Target-Volume, Gross-Tumor-Volume, and Organ-At-Risk (PTV/CTV/GTV/OAR) topology, morphology, and dosimetry, and (B) RT plan efficiency and complexity.&#xD;Methods:&#xD;The software tools are coded in Python. Topological metrics are evaluated using principal component analysis, including center of mass, volume, size, and depth. Morphology is quantified using Hounsfield Units, while dose distribution is characterized by conformity and homogeneity indexes. The total dose within the target versus the body is defined as the Dose Balance Index. &#xD;Results:&#xD;The primary outcome of this study is the toolkit and an analysis of our database. For example, the mean minimum and maximum PTV depths are about 2.5±2.3 cm and 9±3 cm, respectively.&#xD;Conclusions:&#xD;This study provides a statistical basis for RT plans and the necessary tools to generate them. It aids in selecting plans for knowledge-based models and deep-learning networks. The site-specific volume and depth results help identify the limitations and opportunities of current and future treatment modalities, in our case convergent kV RT. The compiled statistics and tools are versatile for training, quality assurance, comparing plans from different periods or institutions, and establishing guidelines. The toolkit is publicly available at https://github.com/m-kinz/STAR.

Scout-Dose-TCM: Direct and Prospective Scout-Based Estimation of Personalized Organ Doses from Tube Current Modulated CT Exams

Maria Jose Medrano, Sen Wang, Liyan Sun, Abdullah-Al-Zubaer Imran, Jennie Cao, Grant Stevens, Justin Ruey Tse, Adam S. Wang

arxiv logopreprintJun 30 2025
This study proposes Scout-Dose-TCM for direct, prospective estimation of organ-level doses under tube current modulation (TCM) and compares its performance to two established methods. We analyzed contrast-enhanced chest-abdomen-pelvis CT scans from 130 adults (120 kVp, TCM). Reference doses for six organs (lungs, kidneys, liver, pancreas, bladder, spleen) were calculated using MC-GPU and TotalSegmentator. Based on these, we trained Scout-Dose-TCM, a deep learning model that predicts organ doses corresponding to discrete cosine transform (DCT) basis functions, enabling real-time estimates for any TCM profile. The model combines a feature learning module that extracts contextual information from lateral and frontal scouts and scan range with a dose learning module that output DCT-based dose estimates. A customized loss function incorporated the DCT formulation during training. For comparison, we implemented size-specific dose estimation per AAPM TG 204 (Global CTDIvol) and its organ-level TCM-adapted version (Organ CTDIvol). A 5-fold cross-validation assessed generalizability by comparing mean absolute percentage dose errors and r-squared correlations with benchmark doses. Average absolute percentage errors were 13% (Global CTDIvol), 9% (Organ CTDIvol), and 7% (Scout-Dose-TCM), with bladder showing the largest discrepancies (15%, 13%, and 9%). Statistical tests confirmed Scout-Dose-TCM significantly reduced errors vs. Global CTDIvol across most organs and improved over Organ CTDIvol for the liver, bladder, and pancreas. It also achieved higher r-squared values, indicating stronger agreement with Monte Carlo benchmarks. Scout-Dose-TCM outperformed Global CTDIvol and was comparable to or better than Organ CTDIvol, without requiring organ segmentations at inference, demonstrating its promise as a tool for prospective organ-level dose estimation in CT.
Page 46 of 1651650 results
Show
per page

Ready to Sharpen Your Edge?

Join hundreds of your peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.