Sort by:
Page 324 of 3313305 results

Artificial Intelligence based radiomic model in Craniopharyngiomas: A Systematic Review and Meta-Analysis on Diagnosis, Segmentation, and Classification.

Mohammadzadeh I, Hajikarimloo B, Niroomand B, Faizi N, Faizi N, Habibi MA, Mohammadzadeh S, Soltani R

pubmed logopapersMay 7 2025
Craniopharyngiomas (CPs) are rare, benign brain tumors originating from Rathke's pouch remnants, typically located in the sellar/parasellar region. Accurate differentiation is crucial due to varying prognoses, with ACPs having higher recurrence and worse outcomes. MRI struggles with overlapping features, complicating diagnosis. this study evaluates the role of Artificial Intelligence (AI) in diagnosing, segmenting, and classifying CPs, emphasizing its potential to improve clinical decision-making, particularly for radiologists and neurosurgeons. This systematic review and meta-analysis assess AI applications in diagnosing, segmenting, and classifying on CPs patients. a comprehensive search was conducted across PubMed, Scopus, Embase and Web of Science for studies employing AI models in patients with CP. Performance metrics such as sensitivity, specificity, accuracy, and area under the curve (AUC) were extracted and synthesized. Eleven studies involving 1916 patients were included in the analysis. The pooled results revealed a sensitivity of 0.740 (95% CI: 0.673-0.808), specificity of 0.813 (95% CI: 0.729-0.898), and accuracy of 0.746 (95% CI: 0.679-0.813). The area under the curve (AUC) for diagnosis was 0.793 (95% CI: 0.719-0.866), and for classification, it was 0.899 (95% CI: 0.846-0.951). The sensitivity for segmentation was found to be 0.755 (95% CI: 0.704-0.805). AI-based models show strong potential in enhancing the diagnostic accuracy and clinical decision-making process for CPs. These findings support the use of AI tools for more reliable preoperative assessment, leading to better treatment planning and patient outcomes. Further research with larger datasets is needed to optimize and validate AI applications in clinical practice.

A Vision-Language Model for Focal Liver Lesion Classification

Song Jian, Hu Yuchang, Wang Hui, Chen Yen-Wei

arxiv logopreprintMay 6 2025
Accurate classification of focal liver lesions is crucial for diagnosis and treatment in hepatology. However, traditional supervised deep learning models depend on large-scale annotated datasets, which are often limited in medical imaging. Recently, Vision-Language models (VLMs) such as Contrastive Language-Image Pre-training model (CLIP) has been applied to image classifications. Compared to the conventional convolutional neural network (CNN), which classifiers image based on visual information only, VLM leverages multimodal learning with text and images, allowing it to learn effectively even with a limited amount of labeled data. Inspired by CLIP, we pro-pose a Liver-VLM, a model specifically designed for focal liver lesions (FLLs) classification. First, Liver-VLM incorporates class information into the text encoder without introducing additional inference overhead. Second, by calculating the pairwise cosine similarities between image and text embeddings and optimizing the model with a cross-entropy loss, Liver-VLM ef-fectively aligns image features with class-level text features. Experimental results on MPCT-FLLs dataset demonstrate that the Liver-VLM model out-performs both the standard CLIP and MedCLIP models in terms of accuracy and Area Under the Curve (AUC). Further analysis shows that using a lightweight ResNet18 backbone enhances classification performance, particularly under data-constrained conditions.

Physics-informed neural network estimation of active material properties in time-dependent cardiac biomechanical models

Matthias Höfler, Francesco Regazzoni, Stefano Pagani, Elias Karabelas, Christoph Augustin, Gundolf Haase, Gernot Plank, Federica Caforio

arxiv logopreprintMay 6 2025
Active stress models in cardiac biomechanics account for the mechanical deformation caused by muscle activity, thus providing a link between the electrophysiological and mechanical properties of the tissue. The accurate assessment of active stress parameters is fundamental for a precise understanding of myocardial function but remains difficult to achieve in a clinical setting, especially when only displacement and strain data from medical imaging modalities are available. This work investigates, through an in-silico study, the application of physics-informed neural networks (PINNs) for inferring active contractility parameters in time-dependent cardiac biomechanical models from these types of imaging data. In particular, by parametrising the sought state and parameter field with two neural networks, respectively, and formulating an energy minimisation problem to search for the optimal network parameters, we are able to reconstruct in various settings active stress fields in the presence of noise and with a high spatial resolution. To this end, we also advance the vanilla PINN learning algorithm with the use of adaptive weighting schemes, ad-hoc regularisation strategies, Fourier features, and suitable network architectures. In addition, we thoroughly analyse the influence of the loss weights in the reconstruction of active stress parameters. Finally, we apply the method to the characterisation of tissue inhomogeneities and detection of fibrotic scars in myocardial tissue. This approach opens a new pathway to significantly improve the diagnosis, treatment planning, and management of heart conditions associated with cardiac fibrosis.

Nonperiodic dynamic CT reconstruction using backward-warping INR with regularization of diffeomorphism (BIRD)

Muge Du, Zhuozhao Zheng, Wenying Wang, Guotao Quan, Wuliang Shi, Le Shen, Li Zhang, Liang Li, Yinong Liu, Yuxiang Xing

arxiv logopreprintMay 6 2025
Dynamic computed tomography (CT) reconstruction faces significant challenges in addressing motion artifacts, particularly for nonperiodic rapid movements such as cardiac imaging with fast heart rates. Traditional methods struggle with the extreme limited-angle problems inherent in nonperiodic cases. Deep learning methods have improved performance but face generalization challenges. Recent implicit neural representation (INR) techniques show promise through self-supervised deep learning, but have critical limitations: computational inefficiency due to forward-warping modeling, difficulty balancing DVF complexity with anatomical plausibility, and challenges in preserving fine details without additional patient-specific pre-scans. This paper presents a novel INR-based framework, BIRD, for nonperiodic dynamic CT reconstruction. It addresses these challenges through four key contributions: (1) backward-warping deformation that enables direct computation of each dynamic voxel with significantly reduced computational cost, (2) diffeomorphism-based DVF regularization that ensures anatomically plausible deformations while maintaining representational capacity, (3) motion-compensated analytical reconstruction that enhances fine details without requiring additional pre-scans, and (4) dimensional-reduction design for efficient 4D coordinate encoding. Through various simulations and practical studies, including digital and physical phantoms and retrospective patient data, we demonstrate the effectiveness of our approach for nonperiodic dynamic CT reconstruction with enhanced details and reduced motion artifacts. The proposed framework enables more accurate dynamic CT reconstruction with potential clinical applications, such as one-beat cardiac reconstruction, cinematic image sequences for functional imaging, and motion artifact reduction in conventional CT scans.

Rethinking Boundary Detection in Deep Learning-Based Medical Image Segmentation

Yi Lin, Dong Zhang, Xiao Fang, Yufan Chen, Kwang-Ting Cheng, Hao Chen

arxiv logopreprintMay 6 2025
Medical image segmentation is a pivotal task within the realms of medical image analysis and computer vision. While current methods have shown promise in accurately segmenting major regions of interest, the precise segmentation of boundary areas remains challenging. In this study, we propose a novel network architecture named CTO, which combines Convolutional Neural Networks (CNNs), Vision Transformer (ViT) models, and explicit edge detection operators to tackle this challenge. CTO surpasses existing methods in terms of segmentation accuracy and strikes a better balance between accuracy and efficiency, without the need for additional data inputs or label injections. Specifically, CTO adheres to the canonical encoder-decoder network paradigm, with a dual-stream encoder network comprising a mainstream CNN stream for capturing local features and an auxiliary StitchViT stream for integrating long-range dependencies. Furthermore, to enhance the model's ability to learn boundary areas, we introduce a boundary-guided decoder network that employs binary boundary masks generated by dedicated edge detection operators to provide explicit guidance during the decoding process. We validate the performance of CTO through extensive experiments conducted on seven challenging medical image segmentation datasets, namely ISIC 2016, PH2, ISIC 2018, CoNIC, LiTS17, and BTCV. Our experimental results unequivocally demonstrate that CTO achieves state-of-the-art accuracy on these datasets while maintaining competitive model complexity. The codes have been released at: https://github.com/xiaofang007/CTO.

Path and Bone-Contour Regularized Unpaired MRI-to-CT Translation

Teng Zhou, Jax Luo, Yuping Sun, Yiheng Tan, Shun Yao, Nazim Haouchine, Scott Raymond

arxiv logopreprintMay 6 2025
Accurate MRI-to-CT translation promises the integration of complementary imaging information without the need for additional imaging sessions. Given the practical challenges associated with acquiring paired MRI and CT scans, the development of robust methods capable of leveraging unpaired datasets is essential for advancing the MRI-to-CT translation. Current unpaired MRI-to-CT translation methods, which predominantly rely on cycle consistency and contrastive learning frameworks, frequently encounter challenges in accurately translating anatomical features that are highly discernible on CT but less distinguishable on MRI, such as bone structures. This limitation renders these approaches less suitable for applications in radiation therapy, where precise bone representation is essential for accurate treatment planning. To address this challenge, we propose a path- and bone-contour regularized approach for unpaired MRI-to-CT translation. In our method, MRI and CT images are projected to a shared latent space, where the MRI-to-CT mapping is modeled as a continuous flow governed by neural ordinary differential equations. The optimal mapping is obtained by minimizing the transition path length of the flow. To enhance the accuracy of translated bone structures, we introduce a trainable neural network to generate bone contours from MRI and implement mechanisms to directly and indirectly encourage the model to focus on bone contours and their adjacent regions. Evaluations conducted on three datasets demonstrate that our method outperforms existing unpaired MRI-to-CT translation approaches, achieving lower overall error rates. Moreover, in a downstream bone segmentation task, our approach exhibits superior performance in preserving the fidelity of bone structures. Our code is available at: https://github.com/kennysyp/PaBoT.

Phenotype-Guided Generative Model for High-Fidelity Cardiac MRI Synthesis: Advancing Pretraining and Clinical Applications

Ziyu Li, Yujian Hu, Zhengyao Ding, Yiheng Mao, Haitao Li, Fan Yi, Hongkun Zhang, Zhengxing Huang

arxiv logopreprintMay 6 2025
Cardiac Magnetic Resonance (CMR) imaging is a vital non-invasive tool for diagnosing heart diseases and evaluating cardiac health. However, the limited availability of large-scale, high-quality CMR datasets poses a major challenge to the effective application of artificial intelligence (AI) in this domain. Even the amount of unlabeled data and the health status it covers are difficult to meet the needs of model pretraining, which hinders the performance of AI models on downstream tasks. In this study, we present Cardiac Phenotype-Guided CMR Generation (CPGG), a novel approach for generating diverse CMR data that covers a wide spectrum of cardiac health status. The CPGG framework consists of two stages: in the first stage, a generative model is trained using cardiac phenotypes derived from CMR data; in the second stage, a masked autoregressive diffusion model, conditioned on these phenotypes, generates high-fidelity CMR cine sequences that capture both structural and functional features of the heart in a fine-grained manner. We synthesized a massive amount of CMR to expand the pretraining data. Experimental results show that CPGG generates high-quality synthetic CMR data, significantly improving performance on various downstream tasks, including diagnosis and cardiac phenotypes prediction. These gains are demonstrated across both public and private datasets, highlighting the effectiveness of our approach. Code is availabel at https://anonymous.4open.science/r/CPGG.

Multi-task learning for joint prediction of breast cancer histological indicators in dynamic contrast-enhanced magnetic resonance imaging.

Sun R, Li X, Han B, Xie Y, Nie S

pubmed logopapersMay 6 2025
Achieving efficient analysis of multiple pathological indicators has great significance for breast cancer prognosis and therapeutic decision-making. In this study, we aim to explore a deep multi-task learning (MTL) framework for collaborative prediction of histological grade and proliferation marker (Ki-67) status in breast cancer using multi-phase dynamic contrast-enhanced magnetic resonance imaging (DCE-MRI). In the novel design of hybrid multi-task architecture (HMT-Net), co-representative features are explicitly distilled using a feature extraction backbone. A customized prediction network is then introduced to perform soft-parameter sharing between two correlated tasks. Specifically, task-common and task-specific knowledge is transmitted into tower layers for informative interactions. Furthermore, low-level feature maps containing tumor edges and texture details are recaptured by a hard-parameter sharing branch, which are then incorporated into the tower layer for each subtask. Finally, the probabilities of two histological indicators, predicted in the multi-phase DCE-MRI, are separately fused using a decision-level fusion strategy. Experimental results demonstrate that the proposed HMT-Net achieves optimal discriminative performance over other recent MTL architectures and deep models based on single image series, with the area under the receiver operating characteristic curve of 0.908 for tumor grade and 0.694 for Ki-67 status. Benefiting from the innovative HMT-Net, our proposed method elucidates its strong robustness and flexibility in the collaborative prediction task of breast biomarkers. Multi-phase DCE-MRI is expected to contribute valuable dynamic information for breast cancer pathological assessment in a non-invasive manner.

A Deep Learning Approach for Mandibular Condyle Segmentation on Ultrasonography.

Keser G, Yülek H, Öner Talmaç AG, Bayrakdar İŞ, Namdar Pekiner F, Çelik Ö

pubmed logopapersMay 6 2025
Deep learning techniques have demonstrated potential in various fields, including segmentation, and have recently been applied to medical image processing. This study aims to develop and evaluate computer-based diagnostic software designed to assess the segmentation of the mandibular condyle in ultrasound images. A total of 668 retrospective ultrasound images of anonymous adult mandibular condyles were analyzed. The CranioCatch labeling program (CranioCatch, Eskişehir, Turkey) was utilized to annotate the mandibular condyle using a polygonal labeling method. These annotations were subsequently reviewed and validated by experts in oral and maxillofacial radiology. In this study, all test images were detected and segmented using the YOLOv8 deep learning artificial intelligence (AI) model. When evaluating the model's performance in image estimation, it achieved an F1 score of 0.93, a sensitivity of 0.90, and a precision of 0.96. The automatic segmentation of the mandibular condyle from ultrasound images presents a promising application of artificial intelligence. This approach can help surgeons, radiologists, and other specialists save time in the diagnostic process.

Machine learning algorithms integrating positron emission tomography/computed tomography features to predict pathological complete response after neoadjuvant chemoimmunotherapy in lung cancer.

Sheng Z, Ji S, Chen Y, Mi Z, Yu H, Zhang L, Wan S, Song N, Shen Z, Zhang P

pubmed logopapersMay 6 2025
Reliable methods for predicting pathological complete response (pCR) in non-small cell lung cancer (NSCLC) patients undergoing neoadjuvant chemoimmunotherapy are still under exploration. Although Fluorine-18 fluorodeoxyglucose-positron emission tomography/computed tomography (18F-FDG PET/CT) features reflect tumour response, their utility in predicting pCR remains controversial. This retrospective analysis included NSCLC patients who received neoadjuvant chemoimmunotherapy followed by 18F-FDG PET/CT imaging at Shanghai Pulmonary Hospital from October 2019 to August 2024. Eligible patients were randomly divided into training and validation cohort at a 7:3 ratio. Relevant 18F-FDG PET/CT features were evaluated as individual predictors and incorporated into 5 machine learning (ML) models. Model performance was assessed using the area under the receiver operating characteristic curve (AUC), and Shapley additive explanation was applied for model interpretation. A total of 205 patients were included, with 91 (44.4%) achieving pCR. Post-treatment tumour maximum standardized uptake value (SUVmax) demonstrated the highest predictive performance among individual predictors, achieving an AUC of 0.72 (95% CI 0.65-0.79), while ΔT SUVmax achieved an AUC of 0.65 (95% CI 0.53-0.77). The Light Gradient Boosting Machine algorithm outperformed other models and individual predictors, achieving an average AUC of 0.87 (95% CI 0.78-0.97) in training cohort and 0.83 (95% CI 0.72-0.94) in validation cohort. Shapley additive explanation analysis identified post-treatment tumour SUVmax and post-treatment nodal volume as key contributors. This ML models offer a non-invasive and effective approach for predicting pCR after neoadjuvant chemoimmunotherapy in NSCLC.
Page 324 of 3313305 results
Show
per page

Ready to Sharpen Your Edge?

Join hundreds of your peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.