Sort by:
Page 16 of 42416 results

Predicting Primary Graft Dysfunction in Systemic Sclerosis Lung Transplantation Using Machine-Learning and CT Features.

Singh J, Meng X, Leader JK, Ryan J, Pu L, Deitz R, Chan EG, Shigemura N, Hage CA, Sanchez PG, Pu J

pubmed logopapersJul 1 2025
Primary graft dysfunction (PGD) is a significant barrier to survival in lung transplant (LTx) recipients. PGD in patients with systemic sclerosis (SSc) remains especially underrepresented in research. We investigated 92 SSc recipients (mean age 51 years ± 10) who underwent bilateral LTx between 2007 and 2020. PGD was defined as grade 3 PGD at 72 h post-LTx. A comprehensive set of CT image features was automatically computed from recipient chest CT scans using deep learning algorithms. Volumetric analysis of recipients' lungs and chest cavity was used to estimate lung-size matching. Four machine learning (ML) algorithms were developed to predict PGD, including multivariate logistic regression, support vector machine (SVM), random forest classifier (RFC), and multilayer perceptron (MLP). PGD was significantly associated with BMI >30 kg/m<sup>2</sup> (p = 0.009), African American race (p = 0.011), lower Preop FEV1 (p = 0.002) and FVC (p = 0.004), longer waitlist time (p = 0.014), higher lung allocation score (LAS) (p = 0.028), and interstitial lung disease (p = 0.050). From CT analysis, PGD was significantly associated with decreased lung volume (p < 0.001), increased heart-chest cavity volume ratio (p < 0.001), epicardial (p = 0.033) and total heart (p = 0.049) adipose tissue, and five cardiopulmonary features (p < 0.050). Oversized donor allografts estimated using CT analysis were significantly associated with PGD (p < 0.050). The MLP model achieved a maximum AUROC of 0.85 (95% CI: 0.81-0.88) in predicting PGD with four features: Preop FEV1, heart-chest cavity volume ratio, waitlist time, and donor to recipient chest cavity volume ratio. CT-derived features are significantly associated with PGD, and models incorporating these features can predict PGD in SSc recipients.

CAD-Unet: A capsule network-enhanced Unet architecture for accurate segmentation of COVID-19 lung infections from CT images.

Dang Y, Ma W, Luo X, Wang H

pubmed logopapersJul 1 2025
Since the outbreak of the COVID-19 pandemic in 2019, medical imaging has emerged as a primary modality for diagnosing COVID-19 pneumonia. In clinical settings, the segmentation of lung infections from computed tomography images enables rapid and accurate quantification and diagnosis of COVID-19. Segmentation of COVID-19 infections in the lungs poses a formidable challenge, primarily due to the indistinct boundaries and limited contrast presented by ground glass opacity manifestations. Moreover, the confounding similarity among infiltrates, lung tissues, and lung walls further complicates this segmentation task. To address these challenges, this paper introduces a novel deep network architecture, called CAD-Unet, for segmenting COVID-19 lung infections. In this architecture, capsule networks are incorporated into the existing Unet framework. Capsule networks represent a novel type of network architecture that differs from traditional convolutional neural networks. They utilize vectors for information transfer among capsules, facilitating the extraction of intricate lesion spatial information. Additionally, we design a capsule encoder path and establish a coupling path between the unet encoder and the capsule encoder. This design maximizes the complementary advantages of both network structures while achieving efficient information fusion. Finally, extensive experiments are conducted on four publicly available datasets, encompassing binary segmentation tasks and multi-class segmentation tasks. The experimental results demonstrate the superior segmentation performance of the proposed model. The code has been released at: https://github.com/AmanoTooko-jie/CAD-Unet.

Radiomics for lung cancer diagnosis, management, and future prospects.

Boubnovski Martell M, Linton-Reid K, Chen M, Aboagye EO

pubmed logopapersJul 1 2025
Lung cancer remains the leading cause of cancer-related mortality worldwide, with its early detection and effective treatment posing significant clinical challenges. Radiomics, the extraction of quantitative features from medical imaging, has emerged as a promising approach for enhancing diagnostic accuracy, predicting treatment responses, and personalising patient care. This review explores the role of radiomics in lung cancer diagnosis and management, with methods ranging from handcrafted radiomics to deep learning techniques that can capture biological intricacies. The key applications are highlighted across various stages of lung cancer care, including nodule detection, histology prediction, and disease staging, where artificial intelligence (AI) models demonstrate superior specificity and sensitivity. The article also examines future directions, emphasising the integration of large language models, explainable AI (XAI), and super-resolution imaging techniques as transformative developments. By merging diverse data sources and incorporating interpretability into AI models, radiomics stands poised to redefine clinical workflows, offering more robust and reliable tools for lung cancer diagnosis, treatment planning, and outcome prediction. These advancements underscore radiomics' potential in supporting precision oncology and improving patient outcomes through data-driven insights.

Gradual poisoning of a chest x-ray convolutional neural network with an adversarial attack and AI explainability methods.

Lee SB

pubmed logopapersJul 1 2025
Given artificial intelligence's transformative effects, studying safety is important to ensure it is implemented in a beneficial way. Convolutional neural networks are used in radiology research for prediction but can be corrupted through adversarial attacks. This study investigates the effect of an adversarial attack, through poisoned data. To improve generalizability, we create a generic ResNet pneumonia classification model and then use it as an example by subjecting it to BadNets adversarial attacks. The study uses various poisoned datasets of different compositions (2%, 16.7% and 100% ratios of poisoned data) and two different test sets (a normal set of test data and one that contained poisoned images) to study the effects of BadNets. To provide a visual effect of the progressing corruption of the models, SHapley Additive exPlanations (SHAP) were used. As corruption progressed, interval analysis revealed that performance on a valid test set decreased while the model learned to predict better on a poisoned test set. SHAP visualization showed focus on the trigger. In the 16.7% poisoned model, SHAP focus did not fixate on the trigger in the normal test set. Minimal effects were seen in the 2% model. SHAP visualization showed decreasing performance was correlated with increasing focus on the trigger. Corruption could potentially be masked in the 16.7% model unless subjected specifically to poisoned data. A minimum threshold for corruption may exist. The study demonstrates insights that can be further studied in future work and with future models. It also identifies areas of potential intervention for safeguarding models against adversarial attacks.

Preoperative Prediction of STAS Risk in Primary Lung Adenocarcinoma Using Machine Learning: An Interpretable Model with SHAP Analysis.

Wang P, Cui J, Du H, Qian Z, Zhan H, Zhang H, Ye W, Meng W, Bai R

pubmed logopapersJul 1 2025
Accurate preoperative prediction of spread through air spaces (STAS) in primary lung adenocarcinoma (LUAD) is critical for optimizing surgical strategies and improving patient outcomes. To develop a machine learning (ML) based model to predict STAS using preoperative CT imaging features and clinicopathological data, while enhancing interpretability through shapley additive explanations (SHAP) analysis. This multicenter retrospective study included 1237 patients with pathologically confirmed primary LUAD from three hospitals. Patients from Center 1 (n=932) were divided into a training set (n=652) and an internal test set (n=280). Patients from Centers 2 (n=165) and 3 (n=140) formed external validation sets. CT imaging features and clinical variables were selected using Boruta and least absolute shrinkage and selection operator regression. Seven ML models were developed and evaluated using five-fold cross-validation. Performance was assessed using F1 score, recall, precision, specificity, sensitivity, and area under the receiver operating characteristic curve (AUC). The Extreme Gradient Boosting (XGB) model achieved AUCs of 0.973 (training set), 0.862 (internal test set), and 0.842/0.810 (external validation sets). SHAP analysis identified nodule type, carcinoembryonic antigen, maximum nodule diameter, and lobulated sign as key features for predicting STAS. Logistic regression analysis confirmed these as independent risk factors. The XGB model demonstrated high predictive accuracy and interpretability for STAS. By integrating widely available clinical and imaging features, this model offers a practical and effective tool for preoperative risk stratification, supporting personalized surgical planning in primary LUAD management.

2.5D deep learning radiomics and clinical data for predicting occult lymph node metastasis in lung adenocarcinoma.

Huang X, Huang X, Wang K, Bai H, Lu X, Jin G

pubmed logopapersJul 1 2025
Occult lymph node metastasis (OLNM) refers to lymph node involvement that remains undetectable by conventional imaging techniques, posing a significant challenge in the accurate staging of lung adenocarcinoma. This study aims to investigate the potential of combining 2.5D deep learning radiomics with clinical data to predict OLNM in lung adenocarcinoma. Retrospective contrast-enhanced CT images were collected from 1,099 patients diagnosed with lung adenocarcinoma across two centers. Multivariable analysis was performed to identify independent clinical risk factors for constructing clinical signatures. Radiomics features were extracted from the enhanced CT images to develop radiomics signatures. A 2.5D deep learning approach was used to extract deep learning features from the images, which were then aggregated using multi-instance learning (MIL) to construct MIL signatures. Deep learning radiomics (DLRad) signatures were developed by integrating the deep learning features with radiomic features. These were subsequently combined with clinical features to form the combined signatures. The performance of the resulting signatures was evaluated using the area under the curve (AUC). The clinical model achieved AUCs of 0.903, 0.866, and 0.785 in the training, validation, and external test cohorts The radiomics model yielded AUCs of 0.865, 0.892, and 0.796 in the training, validation, and external test cohorts. The MIL model demonstrated AUCs of 0.903, 0.900, and 0.852 in the training, validation, and external test cohorts, respectively. The DLRad model showed AUCs of 0.910, 0.908, and 0.875 in the training, validation, and external test cohorts. Notably, the combined model consistently outperformed all other models, achieving AUCs of 0.940, 0.923, and 0.898 in the training, validation, and external test cohorts. The integration of 2.5D deep learning radiomics with clinical data demonstrates strong capability for OLNM in lung adenocarcinoma, potentially aiding clinicians in developing more personalized treatment strategies.

Lung cancer screening with low-dose CT: definition of positive, indeterminate, and negative screen results. A nodule management recommendation from the European Society of Thoracic Imaging.

Snoeckx A, Silva M, Prosch H, Biederer J, Frauenfelder T, Gleeson F, Jacobs C, Kauczor HU, Parkar AP, Schaefer-Prokop C, Prokop M, Revel MP

pubmed logopapersJul 1 2025
Early detection of lung cancer through low-dose CT lung cancer screening in a high-risk population has proven to reduce lung cancer-specific mortality. Nodule management plays a pivotal role in early detection and further diagnostic approaches. The European Society of Thoracic Imaging (ESTI) has established a nodule management recommendation to improve the handling of pulmonary nodules detected during screening. For solid nodules, the primary method for assessing the likelihood of malignancy is to monitor nodule growth using volumetry software. For subsolid nodules, the aggressiveness is determined by measuring the solid part. The ESTI-recommendation enhances existing protocols but puts a stronger focus on lesion aggressiveness. The main goals are to minimise the overall number of follow-up examinations while preventing the risk of a major stage shift and reducing the risk of overtreatment. KEY POINTS: Question Assessment of nodule growth and management according to guidelines is essential in lung cancer screening. Findings Assessment of nodule aggressiveness defines follow-up in lung cancer screening. Clinical relevance The ESTI nodule management recommendation aims to reduce follow-up examinations while preventing major stage shift and overtreatment.

MedGround-R1: Advancing Medical Image Grounding via Spatial-Semantic Rewarded Group Relative Policy Optimization

Huihui Xu, Yuanpeng Nie, Hualiang Wang, Ying Chen, Wei Li, Junzhi Ning, Lihao Liu, Hongqiu Wang, Lei Zhu, Jiyao Liu, Xiaomeng Li, Junjun He

arxiv logopreprintJul 1 2025
Medical Image Grounding (MIG), which involves localizing specific regions in medical images based on textual descriptions, requires models to not only perceive regions but also deduce spatial relationships of these regions. Existing Vision-Language Models (VLMs) for MIG often rely on Supervised Fine-Tuning (SFT) with large amounts of Chain-of-Thought (CoT) reasoning annotations, which are expensive and time-consuming to acquire. Recently, DeepSeek-R1 demonstrated that Large Language Models (LLMs) can acquire reasoning abilities through Group Relative Policy Optimization (GRPO) without requiring CoT annotations. In this paper, we adapt the GRPO reinforcement learning framework to VLMs for Medical Image Grounding. We propose the Spatial-Semantic Rewarded Group Relative Policy Optimization to train the model without CoT reasoning annotations. Specifically, we introduce Spatial-Semantic Rewards, which combine spatial accuracy reward and semantic consistency reward to provide nuanced feedback for both spatially positive and negative completions. Additionally, we propose to use the Chain-of-Box template, which integrates visual information of referring bounding boxes into the <think> reasoning process, enabling the model to explicitly reason about spatial regions during intermediate steps. Experiments on three datasets MS-CXR, ChestX-ray8, and M3D-RefSeg demonstrate that our method achieves state-of-the-art performance in Medical Image Grounding. Ablation studies further validate the effectiveness of each component in our approach. Code, checkpoints, and datasets are available at https://github.com/bio-mlhui/MedGround-R1

Enhanced pulmonary nodule detection with U-Net, YOLOv8, and swin transformer.

Wang X, Wu H, Wang L, Chen J, Li Y, He X, Chen T, Wang M, Guo L

pubmed logopapersJul 1 2025
Lung cancer remains the leading cause of cancer-related mortality worldwide, emphasizing the critical need for early pulmonary nodule detection to improve patient outcomes. Current methods encounter challenges in detecting small nodules and exhibit high false positive rates, placing an additional diagnostic burden on radiologists. This study aimed to develop a two-stage deep learning model integrating U-Net, Yolov8s, and the Swin transformer to enhance pulmonary nodule detection in computer tomography (CT) images, particularly for small nodules, with the goal of improving detection accuracy and reducing false positives. We utilized the LUNA16 dataset (888 CT scans) and an additional 308 CT scans from Tianjin Chest Hospital. Images were preprocessed for consistency. The proposed model first employs U-Net for precise lung segmentation, followed by Yolov8s augmented with the Swin transformer for nodule detection. The Shape-aware IoU (SIoU) loss function was implemented to improve bounding box predictions. For the LUNA16 dataset, the model achieved a precision of 0.898, a recall of 0.851, and a mean average precision at 50% IoU (mAP50) of 0.879, outperforming state-of-the-art models. The Tianjin Chest Hospital dataset has a precision of 0.855, a recall of 0.872, and an mAP50 of 0.862. This study presents a two-stage deep learning model that leverages U-Net, Yolov8s, and the Swin transformer for enhanced pulmonary nodule detection in CT images. The model demonstrates high accuracy and a reduced false positive rate, suggesting its potential as a useful tool for early lung cancer diagnosis and treatment.
Page 16 of 42416 results
Show
per page

Ready to Sharpen Your Edge?

Join hundreds of your peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.