Sort by:
Page 30 of 3503499 results

Comparative analysis of tumor and mesorectum radiomics in predicting neoadjuvant chemoradiotherapy response in locally advanced rectal cancer.

Cantürk A, Yarol RC, Tasak AS, Gülmez H, Kadirli K, Bişgin T, Manoğlu B, Sökmen S, Öztop İ, Görken Bilkay İ, Sağol Ö, Sarıoğlu S, Barlık F

pubmed logopapersAug 12 2025
Neoadjuvant chemoradiotherapy (CRT) is known to increase sphincter preservation rates and decrease the risk of postoperative recurrence in patients with locally advanced rectal tumors. However, the response to CRT in patients with locally advanced rectal cancer (LARC) varies significantly. The objective of this study was to compare the performance of models based on radiomics features of the tumor alone, the mesorectum alone, and a combination of both in predicting tumor response to neoadjuvant CRT in LARC. This retrospective study included 101 patients with LARC. Patients were categorized as responders (modified Ryan score 0-1) and non-responders (modified Ryan score 2-3). Pre-CRT magnetic resonance imaging evaluations included tumor-T2 weighted imaging (T2WI), tumor-diffusion weighted imaging (DWI), tumor-apparent diffusion coefficient (ADC) maps, and mesorectum-T2WI. The first radiologist segmented the tumor and mesorectum from T2-weighted images, and the second radiologist performed tumor segmentation using DWI and ADC maps. Feature reproducibility was assessed by calculating the intraclass correlation coefficient (ICC) using a two-way mixed-effects model with absolute agreement for single measurements [ICC(3,1)]. Radiomic features with ICC values <0.60 were excluded from further analysis. Subsequently, the least absolute shrinkage and selection operator method was applied to select the most relevant radiomic features. The top five features with the highest coefficients were selected for model training. To address class imbalance between groups, the synthetic minority over-sampling technique was applied exclusively to the training folds during cross-validation. Thereafter, classification learner models were developed using 10-fold cross-validation to achieve the highest performance. The performance metrics of the final models, including accuracy, precision, recall, F1-score, and area under the receiver operating characteristic curve (AUC), were calculated to evaluate the classification performance. Among the 101 patients, 36 were classified as responders and 65 as non-responders. A total of 25 radiomic features from the tumor and 20 from the mesorectum were found to be statistically significant (<i>P</i> < 0.05). The AUC values for predicting treatment response were 0.781 for the tumor-only model (random forest), 0.726 for the mesorectum-only model (logistic regression), and 0.837 for the combined model (logistic regression). Radiomic features derived from both the tumor and mesorectum demonstrated complementary prognostic value in predicting treatment response. The inclusion of mesorectal features substantially improved model performance, with the combined model achieving the highest AUC value. These findings highlight the added predictive contribution of the mesorectum as a key peritumoral structure in radiomics-based assessment. Currently, the response of locally advanced rectal tumors to neoadjuvant therapy cannot be reliably predicted using conventional methods. Recently, the significance of the mesorectum in predicting treatment response has gained attention, although the number of studies focusing on this area remains limited. In our study, we performed radiomics analyses of both the tumor tissue and the mesorectum to predict neoadjuvant treatment response.

The association of symptoms, pulmonary function test and computed tomography in interstitial lung disease at the onset of connective tissue disease: an observational study with artificial intelligence analysis of high-resolution computed tomography.

Hoffmann T, Teichgräber U, Brüheim LB, Lassen-Schmidt B, Renz D, Weise T, Krämer M, Oelzner P, Böttcher J, Güttler F, Wolf G, Pfeil A

pubmed logopapersAug 12 2025
Interstitial lung disease (ILD) is a common and serious organ manifestation in patients with connective tissue disease (CTD), but it is uncertain whether there is a difference in ILD between symptomatic and asymptomatic patients. Therefore, we conducted a study to evaluate differences in the extent of ILD based on radiological findings between symptomatic/asymptomatic patients, using an artificial intelligence (AI)-based quantification of pulmonary high-resolution computed tomography (AIpqHRCT). Within the study, 67 cross-sectional HRCT datasets and clinical data (including pulmonary function test) of consecutively patients (mean age: 57.1 ± 14.7 years, woman n = 45; 67.2%) with both, initial diagnosis of CTD, with systemic sclerosis being the most frequent (n = 21, 31.3%), and ILD (all without immunosuppressive therapy), were analysed using AIqpHRCT. 25.4% (n = 17) of the patients with ILD at initial diagnosis of CTD had no pulmonary symptoms. Regarding the baseline characteristics (age, gender, disease), there were no significant difference between the symptomatic and asymptomatic group. The pulmonary function test (PFT) revealed the following mean values (%predicted) in the symptomatic and asymptomatic group, respectively: Forced vital capacity (FVC) 69.4 ± 17.4% versus 86.1 ± 15.8% (p = 0.001), and diffusing capacity of the lung for carbon monoxide (DLCO) 49.7 ± 17.9% versus 60.0 ± 15.8% (p = 0.043). AIqpHRCT data showed a significant higher amount of high attenuated volume (HAV) (14.8 ± 11.0% versus 8.9 ± 3.9%; p = 0.021) and reticulations (5.4 ± 8.7% versus 1.4 ± 1.5%; p = 0.035) in symptomatic patients. A quarter of patients with ILD at the time of initial CTD diagnosis had no pulmonary symptoms, showing DLCO were reduced in both groups. Also, AIqpHRCT demonstrated clinically relevant ILD in asymptomatic patients. These results underline the importance of an early risk adapted screening for ILD also in asymptomatic CTD patients, as ILD is associated with increased mortality.

Development and validation of machine learning models to predict vertebral artery injury by C2 pedicle screws.

Ye B, Sun Y, Chen G, Wang B, Meng H, Shan L

pubmed logopapersAug 12 2025
Cervical 2 pedicle screw (C2PS) fixation is widely used in posterior cervical surgery but carries risks of vertebral artery injury (VAI), a rare yet severe complication. This study aimed to identify risk factors for VAI during C2PS placement and develop a machine learning (ML)-based predictive model to enhance preoperative risk assessment. Clinical and radiological data from 280 patients undergoing head and neck CT angiography were retrospectively analyzed. Three-dimensional reconstructed images simulated C2PS placement, classifying patients into injury (n = 98) and non-injury (n = 182) groups. Fifteen variables, including characteristic of patients and anatomic variables were evaluated. Eight ML algorithms were trained (70% training cohort) and validated (30% validation cohort). Model performance was assessed using AUC, sensitivity, specificity, and SHAP (SHapley Additive exPlanations) for interpretability. Six key risk factors were identified: pedicle diameter, high-riding vertebral artery (HRVA), intra-axial vertebral artery (IAVA), vertebral artery diameter (VAD), distance between the transverse foramen and the posterior end of the vertebral body (TFPEVB) and distance between the vertebral artery and the vertebral body (VAVB). The neural network model (NNet) demonstrated optimal predictive performance, achieving AUCs of 0.929 (training) and 0.936 (validation). SHAP analysis confirmed these variables as primary contributors to VAI risk. This study established an ML-driven predictive model for VAI during C2PS placement, highlighting six critical anatomical and radiological risk factors. Integrating this model into clinical workflows may optimize preoperative planning, reduce complications, and improve surgical outcomes. External validation in multicenter cohorts is warranted to enhance generalizability.

CRCFound: A Colorectal Cancer CT Image Foundation Model Based on Self-Supervised Learning.

Yang J, Cai D, Liu J, Zhuang Z, Zhao Y, Wang FA, Li C, Hu C, Gai B, Chen Y, Li Y, Wang L, Gao F, Wu X

pubmed logopapersAug 12 2025
Accurate risk stratification is crucial for determining the optimal treatment plan for patients with colorectal cancer (CRC). However, existing deep learning models perform poorly in the preoperative diagnosis of CRC and exhibit limited generalizability, primarily due to insufficient annotated data. To address these issues, CRCFound, a self-supervised learning-based CT image foundation model for CRC is proposed. After pretraining on 5137 unlabeled CRC CT images, CRCFound can learn universal feature representations and provide efficient and reliable adaptability for various clinical applications. Comprehensive benchmark tests are conducted on six different diagnostic tasks and two prognosis tasks to validate the performance of the pretrained model. Experimental results demonstrate that CRCFound can easily transfer to most CRC tasks and exhibit outstanding performance and generalization ability. Overall, CRCFound can solve the problem of insufficient annotated data and perform well in a wide range of downstream tasks of CRC, making it a promising solution for accurate diagnosis and personalized treatment of CRC patients.

Unified and Semantically Grounded Domain Adaptation for Medical Image Segmentation

Xin Wang, Yin Guo, Jiamin Xia, Kaiyu Zhang, Niranjan Balu, Mahmud Mossa-Basha, Linda Shapiro, Chun Yuan

arxiv logopreprintAug 12 2025
Most prior unsupervised domain adaptation approaches for medical image segmentation are narrowly tailored to either the source-accessible setting, where adaptation is guided by source-target alignment, or the source-free setting, which typically resorts to implicit supervision mechanisms such as pseudo-labeling and model distillation. This substantial divergence in methodological designs between the two settings reveals an inherent flaw: the lack of an explicit, structured construction of anatomical knowledge that naturally generalizes across domains and settings. To bridge this longstanding divide, we introduce a unified, semantically grounded framework that supports both source-accessible and source-free adaptation. Fundamentally distinct from all prior works, our framework's adaptability emerges naturally as a direct consequence of the model architecture, without the need for any handcrafted adaptation strategies. Specifically, our model learns a domain-agnostic probabilistic manifold as a global space of anatomical regularities, mirroring how humans establish visual understanding. Thus, the structural content in each image can be interpreted as a canonical anatomy retrieved from the manifold and a spatial transformation capturing individual-specific geometry. This disentangled, interpretable formulation enables semantically meaningful prediction with intrinsic adaptability. Extensive experiments on challenging cardiac and abdominal datasets show that our framework achieves state-of-the-art results in both settings, with source-free performance closely approaching its source-accessible counterpart, a level of consistency rarely observed in prior works. Beyond quantitative improvement, we demonstrate strong interpretability of the proposed framework via manifold traversal for smooth shape manipulation.

Hierarchical Variable Importance with Statistical Control for Medical Data-Based Prediction

Joseph Paillard, Antoine Collas, Denis A. Engemann, Bertrand Thirion

arxiv logopreprintAug 12 2025
Recent advances in machine learning have greatly expanded the repertoire of predictive methods for medical imaging. However, the interpretability of complex models remains a challenge, which limits their utility in medical applications. Recently, model-agnostic methods have been proposed to measure conditional variable importance and accommodate complex non-linear models. However, they often lack power when dealing with highly correlated data, a common problem in medical imaging. We introduce Hierarchical-CPI, a model-agnostic variable importance measure that frames the inference problem as the discovery of groups of variables that are jointly predictive of the outcome. By exploring subgroups along a hierarchical tree, it remains computationally tractable, yet also enjoys explicit family-wise error rate control. Moreover, we address the issue of vanishing conditional importance under high correlation with a tree-based importance allocation mechanism. We benchmarked Hierarchical-CPI against state-of-the-art variable importance methods. Its effectiveness is demonstrated in two neuroimaging datasets: classifying dementia diagnoses from MRI data (ADNI dataset) and analyzing the Berger effect on EEG data (TDBRAIN dataset), identifying biologically plausible variables.

PADReg: Physics-Aware Deformable Registration Guided by Contact Force for Ultrasound Sequences

Yimeng Geng, Mingyang Zhao, Fan Xu, Guanglin Cao, Gaofeng Meng, Hongbin Liu

arxiv logopreprintAug 12 2025
Ultrasound deformable registration estimates spatial transformations between pairs of deformed ultrasound images, which is crucial for capturing biomechanical properties and enhancing diagnostic accuracy in diseases such as thyroid nodules and breast cancer. However, ultrasound deformable registration remains highly challenging, especially under large deformation. The inherently low contrast, heavy noise and ambiguous tissue boundaries in ultrasound images severely hinder reliable feature extraction and correspondence matching. Existing methods often suffer from poor anatomical alignment and lack physical interpretability. To address the problem, we propose PADReg, a physics-aware deformable registration framework guided by contact force. PADReg leverages synchronized contact force measured by robotic ultrasound systems as a physical prior to constrain the registration. Specifically, instead of directly predicting deformation fields, we first construct a pixel-wise stiffness map utilizing the multi-modal information from contact force and ultrasound images. The stiffness map is then combined with force data to estimate a dense deformation field, through a lightweight physics-aware module inspired by Hooke's law. This design enables PADReg to achieve physically plausible registration with better anatomical alignment than previous methods relying solely on image similarity. Experiments on in-vivo datasets demonstrate that it attains a HD95 of 12.90, which is 21.34\% better than state-of-the-art methods. The source code is available at https://github.com/evelynskip/PADReg.

MMIF-AMIN: Adaptive Loss-Driven Multi-Scale Invertible Dense Network for Multimodal Medical Image Fusion

Tao Luo, Weihua Xu

arxiv logopreprintAug 12 2025
Multimodal medical image fusion (MMIF) aims to integrate images from different modalities to produce a comprehensive image that enhances medical diagnosis by accurately depicting organ structures, tissue textures, and metabolic information. Capturing both the unique and complementary information across multiple modalities simultaneously is a key research challenge in MMIF. To address this challenge, this paper proposes a novel image fusion method, MMIF-AMIN, which features a new architecture that can effectively extract these unique and complementary features. Specifically, an Invertible Dense Network (IDN) is employed for lossless feature extraction from individual modalities. To extract complementary information between modalities, a Multi-scale Complementary Feature Extraction Module (MCFEM) is designed, which incorporates a hybrid attention mechanism, convolutional layers of varying sizes, and Transformers. An adaptive loss function is introduced to guide model learning, addressing the limitations of traditional manually-designed loss functions and enhancing the depth of data mining. Extensive experiments demonstrate that MMIF-AMIN outperforms nine state-of-the-art MMIF methods, delivering superior results in both quantitative and qualitative analyses. Ablation experiments confirm the effectiveness of each component of the proposed method. Additionally, extending MMIF-AMIN to other image fusion tasks also achieves promising performance.

Machine learning models for diagnosing lymph node recurrence in postoperative PTC patients: a radiomic analysis.

Pang F, Wu L, Qiu J, Guo Y, Xie L, Zhuang S, Du M, Liu D, Tan C, Liu T

pubmed logopapersAug 12 2025
Postoperative papillary thyroid cancer (PTC) patients often have enlarged cervical lymph nodes due to inflammation or hyperplasia, which complicates the assessment of recurrence or metastasis. This study aimed to explore the diagnostic capabilities of computed tomography (CT) imaging and radiomic analysis to distinguish the recurrence of cervical lymph nodes in patients with PTC postoperatively. A retrospective analysis of 194 PTC patients who underwent total thyroidectomy was conducted, with 98 cases of cervical lymph node recurrence and 96 cases without recurrence. Using 3D Slicer software, Regions of Interest (ROI) were delineated on enhanced venous phase CT images, analyzing 302 positive and 391 negative lymph nodes. These nodes were randomly divided into training and validation sets in a 3:2 ratio. Python was used to extract radiomic features from the ROIs and to develop radiomic models. Univariate and multivariate analyses identified statistically significant risk factors for cervical lymph node recurrence from clinical data, which, when combined with radiomic scores, formed a nomogram to predict recurrence risk. The diagnostic efficacy and clinical utility of the models were assessed using ROC curves, calibration curves, and Decision Curve Analysis (DCA). This study analyzed 693 lymph nodes (302 positive and 391 negative) and identified 35 significant radiomic features through dimensionality reduction and selection. The three machine learning models, including the Lasso regression, Support Vector Machine (SVM), and RF radiomics models, showed.

Dynamic Survival Prediction using Longitudinal Images based on Transformer

Bingfan Liu, Haolun Shi, Jiguo Cao

arxiv logopreprintAug 12 2025
Survival analysis utilizing multiple longitudinal medical images plays a pivotal role in the early detection and prognosis of diseases by providing insight beyond single-image evaluations. However, current methodologies often inadequately utilize censored data, overlook correlations among longitudinal images measured over multiple time points, and lack interpretability. We introduce SurLonFormer, a novel Transformer-based neural network that integrates longitudinal medical imaging with structured data for survival prediction. Our architecture comprises three key components: a Vision Encoder for extracting spatial features, a Sequence Encoder for aggregating temporal information, and a Survival Encoder based on the Cox proportional hazards model. This framework effectively incorporates censored data, addresses scalability issues, and enhances interpretability through occlusion sensitivity analysis and dynamic survival prediction. Extensive simulations and a real-world application in Alzheimer's disease analysis demonstrate that SurLonFormer achieves superior predictive performance and successfully identifies disease-related imaging biomarkers.
Page 30 of 3503499 results
Show
per page

Ready to Sharpen Your Edge?

Join hundreds of your peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.