Sort by:
Page 8 of 38373 results

HarmonicEchoNet: Leveraging harmonic convolutions for automated standard plane detection in fetal heart ultrasound videos.

Sarker MMK, Mishra D, Alsharid M, Hernandez-Cruz N, Ahuja R, Patey O, Papageorghiou AT, Noble JA

pubmed logopapersAug 26 2025
Fetal echocardiography offers non-invasive and real-time imaging acquisition of fetal heart images to identify congenital heart conditions. Manual acquisition of standard heart views is time-consuming, whereas automated detection remains challenging due to high spatial similarity across anatomical views with subtle local image appearance variations. To address these challenges, we introduce a very lightweight frequency-guided deep learning-based model named HarmonicEchoNet that can automatically detect heart standard views in a transverse sweep or freehand ultrasound scan of the fetal heart. HarmonicEchoNet uses harmonic convolution blocks (HCBs) and a harmonic spatial and channel squeeze-and-excitation (hscSE) module. The HCBs apply a Discrete Cosine Transform (DCT)-based harmonic decomposition to input features, which are then combined using learned weights. The hscSE module identifies significant regions in the spatial domain to improve feature extraction of the fetal heart anatomical structures, capturing both spatial and channel-wise dependencies in an ultrasound image. The combination of these modules improves model performance relative to recent CNN-based, transformer-based, and CNN+transformer-based image classification models. We use four datasets from two private studies, PULSE (Perception Ultrasound by Learning Sonographic Experience) and CAIFE (Clinical Artificial Intelligence in Fetal Echocardiography), to develop and evaluate HarmonicEchoNet models. Experimental results show that HarmonicEchoNet is 10-15 times faster than ConvNeXt, DeiT, and VOLO, with an inference time of just 3.9 ms. It also achieves 2%-7% accuracy improvement in classifying fetal heart standard planes compared to these baselines. Furthermore, with just 19.9 million parameters compared to ConvNeXt's 196.24 million, HarmonicEchoNet is nearly ten times more parameter-efficient.

PRISM: A Framework Harnessing Unsupervised Visual Representations and Textual Prompts for Explainable MACE Survival Prediction from Cardiac Cine MRI

Haoyang Su, Jin-Yi Xiang, Shaohao Rui, Yifan Gao, Xingyu Chen, Tingxuan Yin, Xiaosong Wang, Lian-Ming Wu

arxiv logopreprintAug 26 2025
Accurate prediction of major adverse cardiac events (MACE) remains a central challenge in cardiovascular prognosis. We present PRISM (Prompt-guided Representation Integration for Survival Modeling), a self-supervised framework that integrates visual representations from non-contrast cardiac cine magnetic resonance imaging with structured electronic health records (EHRs) for survival analysis. PRISM extracts temporally synchronized imaging features through motion-aware multi-view distillation and modulates them using medically informed textual prompts to enable fine-grained risk prediction. Across four independent clinical cohorts, PRISM consistently surpasses classical survival prediction models and state-of-the-art (SOTA) deep learning baselines under internal and external validation. Further clinical findings demonstrate that the combined imaging and EHR representations derived from PRISM provide valuable insights into cardiac risk across diverse cohorts. Three distinct imaging signatures associated with elevated MACE risk are uncovered, including lateral wall dyssynchrony, inferior wall hypersensitivity, and anterior elevated focus during diastole. Prompt-guided attribution further identifies hypertension, diabetes, and smoking as dominant contributors among clinical and physiological EHR factors.

Bi-directional semi-3D network for accurate epicardial fat segmentation and quantification using reflection equivariant quantum neural networks.

S J, Perumalsamy M

pubmed logopapersAug 25 2025
The process of detecting and measuring the fat layer surrounding the heart from medical images is referred to as epicardial fat segmentation. Accurate segmentation is essential for assessing heart health and associated risk factors. It plays a critical role in evaluating cardiovascular disease, requiring advanced techniques to enhance precision and effectiveness. However, there is currently a shortage of resources made for fat mass measurement. The Visual Lab's cardiac fat database addresses this limitation by providing a comprehensive set of high-resolution images crucial for reliable fat analysis. This study proposes a novel method for epicardial fat segmentation, involving a multi-stage framework. In the preprocessing phase, window-aware guided bilateral filtering (WGBR) is applied to reduce noise while preserving structural features. For region-of-interest (ROI) selection, the White Shark Optimizer (WSO) is employed to improve exploration and exploitation accuracy. The segmentation task is handled using a bidirectional guided semi-3D network (BGSNet), which enhances robustness by extracting features in both forward and backward directions. Following segmentation, quantification is performed to estimate the epicardial fat volume. This is achieved using reflection-equivariant quantum neural networks (REQNN), which are well-suited for modelling complex visual patterns. The Parrot optimizer is further utilized to fine-tune hyperparameters, ensuring optimal performance. The experimental results confirm the effectiveness of the suggested BGSNet with REQNN approach, achieving a Dice score of 99.50 %, an accuracy of 99.50 %, and an execution time of 1.022 s per slice. Furthermore, the Spearman correlation coefficient for fat quantification yielded an R<sup>2</sup> value of 0.9867, indicating a strong agreement with the reference measurements. This integrated approach offers a reliable solution for epicardial fat segmentation and quantification, thereby supporting improved cardiovascular risk assessment and monitoring.

ControlEchoSynth: Boosting Ejection Fraction Estimation Models via Controlled Video Diffusion

Nima Kondori, Hanwen Liang, Hooman Vaseli, Bingyu Xie, Christina Luong, Purang Abolmaesumi, Teresa Tsang, Renjie Liao

arxiv logopreprintAug 25 2025
Synthetic data generation represents a significant advancement in boosting the performance of machine learning (ML) models, particularly in fields where data acquisition is challenging, such as echocardiography. The acquisition and labeling of echocardiograms (echo) for heart assessment, crucial in point-of-care ultrasound (POCUS) settings, often encounter limitations due to the restricted number of echo views available, typically captured by operators with varying levels of experience. This study proposes a novel approach for enhancing clinical diagnosis accuracy by synthetically generating echo views. These views are conditioned on existing, real views of the heart, focusing specifically on the estimation of ejection fraction (EF), a critical parameter traditionally measured from biplane apical views. By integrating a conditional generative model, we demonstrate an improvement in EF estimation accuracy, providing a comparative analysis with traditional methods. Preliminary results indicate that our synthetic echoes, when used to augment existing datasets, not only enhance EF estimation but also show potential in advancing the development of more robust, accurate, and clinically relevant ML models. This approach is anticipated to catalyze further research in synthetic data applications, paving the way for innovative solutions in medical imaging diagnostics.

Pushing the limits of cardiac MRI: deep-learning based real-time cine imaging in free breathing vs breath hold.

Klemenz AC, Watzke LM, Deyerberg KK, Böttcher B, Gorodezky M, Manzke M, Dalmer A, Lorbeer R, Weber MA, Meinel FG

pubmed logopapersAug 23 2025
To evaluate deep-learning (DL) based real-time cardiac cine sequences acquired in free breathing (FB) vs breath hold (BH). In this prospective single-centre cohort study, 56 healthy adult volunteers were investigated on a 1.5-T MRI scanner. A set of real-time cine sequences, including a short-axis stack, 2-, 3-, and 4-chamber views, was acquired in FB and with BH. A validated DL-based cine sequence acquired over three cardiac cycles served as the reference standard for volumetric results. Subjective image quality (sIQ) was rated by two blinded readers. Volumetric analysis of both ventricles was performed. sIQ was rated as good to excellent for FB real-time cine images, slightly inferior to BH real-time cine images (p < 0.0001). Overall acquisition time for one set of cine sequences was 50% shorter with FB (median 90 vs 180 s, p < 0.0001). There were significant differences between the real-time sequences and the reference in left ventricular (LV) end-diastolic volume, LV end-systolic volume, LV stroke volume and LV mass. Nevertheless, BH cine imaging showed excellent correlation with the reference standard, with an intra-class correlation coefficient (ICC) > 0.90 for all parameters except right ventricular ejection fraction (RV EF, ICC = 0.887). With FB cine imaging, correlation with the reference standard was good for LV ejection fraction (LV EF, ICC = 0.825) and RV EF (ICC = 0.824) and excellent (ICC > 0.90) for all other parameters. DL-based real-time cine imaging is feasible even in FB with good to excellent image quality and acceptable volumetric results in healthy volunteers. Question Conventional cardiac MR (CMR) cine imaging is challenged by arrhythmias and patients unable to hold their breath, since data is acquired over several heartbeats. Findings DL-based real-time cine imaging is feasible in FB with acceptable volumetric results and reduced acquisition time by 50% compared to real-time breath-hold sequences. Clinical relevance This study fits into the wider goal of increasing the availability of CMR by reducing the complexity, duration of the examination and improving patient comfort and making CMR available even for patients who are unable to hold their breath.

Utility of machine learning for predicting severe chronic thromboembolic pulmonary hypertension based on CT metrics in a surgical cohort.

Grubert Van Iderstine M, Kim S, Karur GR, Granton J, de Perrot M, McIntosh C, McInnis M

pubmed logopapersAug 23 2025
The aim of this study was to develop machine learning (ML) models to explore the relationship between chronic pulmonary embolism (PE) burden and severe pulmonary hypertension (PH) in surgical chronic thromboembolic pulmonary hypertension (CTEPH). CTEPH patients with a preoperative CT pulmonary angiogram and pulmonary endarterectomy between 01/2017 and 06/2022 were included. A mean pulmonary artery pressure of > 50 mmHg was classified as severe. CTs were scored by a blinded radiologist who recorded chronic pulmonary embolism extent in detail, and measured the right ventricle (RV), left ventricle (LV), main pulmonary artery (PA) and ascending aorta (Ao) diameters. XGBoost models were developed to identify CTEPH feature importance and compared to a logistic regression model. There were 184 patients included; 54.9% were female, and 21.7% had severe PH. The average age was 57 ± 15 years. PE burden alone was not helpful in identifying severe PH. The RV/LV ratio logistic regression model performed well (AUC 0.76) with a cutoff of 1.4. A baseline ML model (Model 1) including only the RV, LV, Pa and Ao measures and their ratios yielded an average AUC of 0.66 ± 0.10. The addition of demographics and statistics summarizing the CT findings raised the AUC to 0.75 ± 0.08 (F1 score 0.41). While measures of PE burden had little bearing on PH severity independently, the RV/LV ratio, extent of disease in various segments, total webs observed, and patient demographics improved performance of machine learning models in identifying severe PH. Question Can machine learning methods applied to CT-based cardiac measurements and detailed maps of chronic thromboembolism type and distribution predict pulmonary hypertension (PH) severity? Findings The right-to-left ventricle (RV/LV) ratio was predictive of PH severity with an optimal cutoff of 1.4, and detailed accounts of chronic thromboembolic burden improved model performance. Clinical relevance The identification of a CT-based RV/LV ratio cutoff of 1.4 gives radiologists, clinicians, and patients a point of reference for chronic thromboembolic PH severity. Detailed chronic thromboembolic burden data are useful but cannot be used alone to predict PH severity.

Epicardial and paracardial adipose tissue quantification in short-axis cardiac cine MRI using deep learning.

Zhang R, Wang X, Zhou Z, Ni L, Jiang M, Hu P

pubmed logopapersAug 23 2025
Epicardial and paracardial adipose tissues (EAT and PAT) are two types of fat depots around the heart and they have important roles in cardiac physiology. Manual quantification of EAT and PAT from cardiac MR (CMR) is time-consuming and prone to human bias. Leveraging the cardiac motion, we aimed to develop deep learning neural networks for automated segmentation and quantification of EAT and PAT in short-axis cine CMR. A modified U-Net equipped with modules of multi-resolution convolution, motion information extraction, feature fusion, and dual attention mechanisms, was developed. Multiple steps of ablation studies were performed to verify the efficacy of each module. The performance of different networks was also compared. The final network incorporating all modules achieved segmentation Dice indices of 77.72% ± 2.53% and 77.18% ± 3.54% for EAT and PAT, respectively, which were significantly higher than the baseline U-Net. It also achieved the highest performance compared to other networks. With our model, the determination coefficients of EAT and PAT volumes to the reference were 0.8550 and 0.8025, respectively. Our proposed network can provide accurate and quick quantification of EAT and PAT on routine short-axis cine CMR, which can potentially aid cardiologists in clinical settings.

Deep Learning-based Automated Coronary Plaque Quantification: First Demonstration With Ultra-high Resolution Photon-counting Detector CT at Different Temporal Resolutions.

Klambauer K, Burger SD, Demmert TT, Mergen V, Moser LJ, Gulsun MA, Schöbinger M, Schwemmer C, Wels M, Allmendinger T, Eberhard M, Alkadhi H, Schmidt B

pubmed logopapersAug 22 2025
The aim of this study was to evaluate the feasibility and reproducibility of a novel deep learning (DL)-based coronary plaque quantification tool with automatic case preparation in patients undergoing ultra-high resolution (UHR) photon-counting detector CT coronary angiography (CCTA), and to assess the influence of temporal resolution on plaque quantification. In this retrospective single-center study, 45 patients undergoing clinically indicated UHR CCTA were included. In each scan, 2 image data sets were reconstructed: one in the dual-source mode with 66 ms temporal resolution and one simulating a single-source mode with 125 ms temporal resolution. A novel, DL-based algorithm for fully automated coronary segmentation and intensity-based plaque quantification was applied to both data sets in each patient. Plaque volume quantification was performed at the vessel-level for the entire left anterior descending artery (LAD), left circumflex artery (CX), and right coronary artery (RCA), as well as at the lesion-level for the largest coronary plaque in each vessel. Diameter stenosis grade was quantified for the coronary lesion with the greatest longitudinal extent in each vessel. To assess reproducibility, the algorithm was rerun 3 times in 10 randomly selected patients, and all outputs were visually reviewed and confirmed by an expert reader. Paired Wilcoxon signed-rank tests with Benjamini-Hochberg correction were used for statistical comparisons. One hundred nineteen out of 135 (88.1%) coronary arteries showed atherosclerotic plaques and were included in the analysis. In the reproducibility analysis, repeated runs of the algorithm yielded identical results across all plaque and lumen measurements (P > 0.999). All outputs were confirmed to be anatomically correct, visually consistent, and did not require manual correction. At the vessel level, total plaque volumes were higher in the 125 ms reconstructions compared with the 66 ms reconstructions in 28 of 45 patients (62%), with both calcified and noncalcified plaque volumes being higher in 32 (71%) and 28 (62%) patients, respectively. Total plaque volumes in the LAD, CX, and RCA were significantly higher in the 125 ms reconstructions (681.3 vs. 647.8  mm3, P < 0.05). At the lesion level, total plaque volumes were higher in the 125 ms reconstructions in 44 of 45 patients (98%; 447.3 vs. 414.9  mm3, P < 0.001), with both calcified and noncalcified plaque volumes being higher in 42 of 45 patients (93%). The median diameter stenosis grades for all vessels were significantly higher in the 125 ms reconstructions (35.4% vs. 28.1%, P < 0.01). This study evaluated a novel DL-based tool with automatic case preparation for quantitative coronary plaque in UHR CCTA data sets. The algorithm was technically robust and reproducible, delivering anatomically consistent outputs not requiring manual correction. Reconstructions with lower temporal resolution (125 ms) systematically overestimated plaque burden compared with higher temporal resolution (66 ms), underscoring that protocol standardization is essential for reliable DL-based plaque quantification.

Integrating Imaging-Derived Clinical Endotypes with Plasma Proteomics and External Polygenic Risk Scores Enhances Coronary Microvascular Disease Risk Prediction

Venkatesh, R., Cherlin, T., Penn Medicine BioBank,, Ritchie, M. D., Guerraty, M., Verma, S. S.

medrxiv logopreprintAug 21 2025
Coronary microvascular disease (CMVD) is an underdiagnosed but significant contributor to the burden of ischemic heart disease, characterized by angina and myocardial infarction. The development of risk prediction models such as polygenic risk scores (PRS) for CMVD has been limited by a lack of large-scale genome-wide association studies (GWAS). However, there is significant overlap between CMVD and enrollment criteria for coronary artery disease (CAD) GWAS. In this study, we developed CMVD PRS models by selecting variants identified in a CMVD GWAS and applying weights from an external CAD GWAS, using CMVD-associated loci as proxies for the genetic risk. We integrated plasma proteomics, clinical measures from perfusion PET imaging, and PRS to evaluate their contributions to CMVD risk prediction in comprehensive machine and deep learning models. We then developed a novel unsupervised endotyping framework for CMVD from perfusion PET-derived myocardial blood flow data, revealing distinct patient subgroups beyond traditional case-control definitions. This imaging-based stratification substantially improved classification performance alongside plasma proteomics and PRS, achieving AUROCs between 0.65 and 0.73 per class, significantly outperforming binary classifiers and existing clinical models, highlighting the potential of this stratification approach to enable more precise and personalized diagnosis by capturing the underlying heterogeneity of CMVD. This work represents the first application of imaging-based endotyping and the integration of genetic and proteomic data for CMVD risk prediction, establishing a framework for multimodal modeling in complex diseases.

TPA: Temporal Prompt Alignment for Fetal Congenital Heart Defect Classification

Darya Taratynova, Alya Almsouti, Beknur Kalmakhanbet, Numan Saeed, Mohammad Yaqub

arxiv logopreprintAug 21 2025
Congenital heart defect (CHD) detection in ultrasound videos is hindered by image noise and probe positioning variability. While automated methods can reduce operator dependence, current machine learning approaches often neglect temporal information, limit themselves to binary classification, and do not account for prediction calibration. We propose Temporal Prompt Alignment (TPA), a method leveraging foundation image-text model and prompt-aware contrastive learning to classify fetal CHD on cardiac ultrasound videos. TPA extracts features from each frame of video subclips using an image encoder, aggregates them with a trainable temporal extractor to capture heart motion, and aligns the video representation with class-specific text prompts via a margin-hinge contrastive loss. To enhance calibration for clinical reliability, we introduce a Conditional Variational Autoencoder Style Modulation (CVAESM) module, which learns a latent style vector to modulate embeddings and quantifies classification uncertainty. Evaluated on a private dataset for CHD detection and on a large public dataset, EchoNet-Dynamic, for systolic dysfunction, TPA achieves state-of-the-art macro F1 scores of 85.40% for CHD diagnosis, while also reducing expected calibration error by 5.38% and adaptive ECE by 6.8%. On EchoNet-Dynamic's three-class task, it boosts macro F1 by 4.73% (from 53.89% to 58.62%). Temporal Prompt Alignment (TPA) is a framework for fetal congenital heart defect (CHD) classification in ultrasound videos that integrates temporal modeling, prompt-aware contrastive learning, and uncertainty quantification.
Page 8 of 38373 results
Show
per page

Ready to Sharpen Your Edge?

Join hundreds of your peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.