Sort by:
Page 106 of 2402393 results

Transfer Learning for Automated Two-class Classification of Pulmonary Tuberculosis in Chest X-Ray Images.

Nayyar A, Shrivastava R, Jain S

pubmed logopapersJul 21 2025
Early and precise diagnosis is essential for effectively treating and managing pulmonary tuberculosis. The purpose of this research is to leverage artificial intelligence (AI), specifically convolutional neural networks (CNNs), to expedite the diagnosis of tuberculosis (TB) using chest X-ray (CXR) images. Mycobacterium tuberculosis, an aerobic bacterium, is the causative agent of TB. The disease remains a global health challenge, particularly in densely populated countries. Early detection via chest X-rays is crucial, but limited medical expertise hampers timely diagnosis. This study explores the application of CNNs, a highly efficient method, for automated TB detection, especially in areas with limited medical expertise. Previously trained models, specifically VGG-16, VGG-19, ResNet 50, and Inception v3, were used to validate the data. Effective feature extraction and classification in medical image analysis, especially in TB diagnosis, is facilitated by the distinct design and capabilities that each model offers. VGG-16 and VGG-19 are very good at identifying minute distinctions and hierarchical characteristics from CXR images; on the other hand, ResNet 50 avoids overfitting while retaining both low and high-level features. The inception v3 model is quite useful for examining various complex patterns in a CXR image with its capacity to extract multi-scale features. Inception v3 outperformed other models, attaining 97.60% accuracy without pre-processing and 98.78% with pre-processing. The proposed model shows promising results as a tool for improving TB diagnosis, and reducing the global impact of the disease, but further validation with larger and more diverse datasets is needed.

AI-Assisted Semiquantitative Measurement of Murine Bleomycin-Induced Lung Fibrosis Using In Vivo Micro-CT: An End-to-End Approach.

Cheng H, Gao T, Sun Y, Huang F, Gu X, Shan C, Wang B, Luo S

pubmed logopapersJul 21 2025
Small animal models are crucial for investigating idiopathic pulmonary fibrosis (IPF) and developing preclinical therapeutic strategies. However, there are several limitations to the quantitative measurements used in the longitudinal assessment of experimental lung fibrosis, e.g., histological or biochemical analyses introduce inter-individual variability, while image-derived biomarker has yet to directly and accurately quantify the severity of lung fibrosis. This study investigates artificial intelligence (AI)-assisted, end-to-end, semi-quantitative measurement of lung fibrosis using in vivo micro-CT. Based on the bleomycin (BLM)-induced lung fibrosis mouse model, the AI model predicts histopathological scores from in vivo micro-CT images, directly correlating these images with the severity of lung fibrosis in mice. Fibrosis severity was graded by the Ashcroft scale: none (0), mild (1-3), moderate (4-5), severe (≥6).The overall accuracy, precision, recall, and F1 scores of the lung fibrosis severity-stratified 3-fold cross validation on 225 micro-CT images for the proposed AI model were 92.9%, 90.9%, 91.6%, and 91.0%. The overall area under the receiver operating characteristic curve (AUROC) was 0.990 (95% CI: 0.977, 1.000), with AUROC values of 1.000 for none (100 images, 95% CI: 0.997, 1.000), 0.969 for mild (43 images, 95% CI: 0.918, 1.000), 0.992 for moderate (36 images, 95% CI: 0.962, 1.000), and 0.992 for severe (46 images, 95% CI: 0.967, 1.000). Preliminary results indicate that AI-assisted, in vivo micro-CT-based semi-quantitative measurements of murine are feasible and likely accurate. This novel method holds promise as a tool to improve the reproducibility of experimental studies in animal models of IPF.

Deep Learning-Driven Multimodal Fusion Model for Prediction of Middle Cerebral Artery Aneurysm Rupture Risk.

Jia X, Chen Y, Zheng K, Chen C, Liu J

pubmed logopapersJul 21 2025
The decision to treat unruptured intracranial aneurysms remains a clinical dilemma. Middle cerebral artery (MCA) aneurysms represent a prevalent subtype of intracranial aneurysms. This study aims to develop a multimodal fusion deep learning model for stratifying rupture risk in MCA aneurysms. We retrospectively enrolled internal cohort and two external validation datasets with 578 and 51 MCA aneurysms, respectively. Multivariate logistic regression analysis was performed to identify independent predictors of rupture. Aneurysm morphological parameters were quantified using reconstructed CT angiography (CTA) images. Radiomics features of aneurysms were extracted through computational analysis. We developed MCANet - a multimodal data-driven classification model integrating raw CTA images, radiomics features, clinical parameters, and morphological characteristics - to establish an aneurysm rupture risk assessment framework. External validation was conducted using datasets from two independent medical centers to evaluate model generalizability and small-sample robustness. Four key metrics, including accuracy, F1-score, precision, and recall, were employed to assess model performance. In the internal cohort, 369 aneurysms were ruptured. Independent predictors of rupture included: the presence of multiple aneurysms, aneurysm location, aneurysm angle, presence of daughter-sac aneurysm, and height-width ratio. MCANet demonstrated satisfactory predictive performance with 91.38% accuracy, 96.33% sensitivity, 90.52% precision, and 93.33% F1-score. External validation maintained good discriminative ability across both independent cohorts. The MCANet model effectively integrates multimodal heterogeneous data for MCA aneurysm rupture risk prediction, demonstrating clinical applicability even in data-constrained scenarios. This model shows potential to optimize therapeutic decision-making and mitigate patient anxiety through individualized risk assessment.

DREAM: A framework for discovering mechanisms underlying AI prediction of protected attributes

Gadgil, S. U., DeGrave, A. J., Janizek, J. D., Xu, S., Nwandu, L., Fonjungo, F., Lee, S.-I., Daneshjou, R.

medrxiv logopreprintJul 21 2025
Recent advances in Artificial Intelligence (AI) have started disrupting the healthcare industry, especially medical imaging, and AI devices are increasingly being deployed into clinical practice. Such classifiers have previously demonstrated the ability to discern a range of protected demographic attributes (like race, age, sex) from medical images with unexpectedly high performance, a sensitive task which is difficult even for trained physicians. In this study, we motivate and introduce a general explainable AI (XAI) framework called DREAM (DiscoveRing and Explaining AI Mechanisms) for interpreting how AI models trained on medical images predict protected attributes. Focusing on two modalities, radiology and dermatology, we are successfully able to train high-performing classifiers for predicting race from chest x-rays (ROC-AUC score of [~]0.96) and sex from dermoscopic lesions (ROC-AUC score of [~]0.78). We highlight how incorrect use of these demographic shortcuts can have a detrimental effect on the performance of a clinically relevant downstream task like disease diagnosis under a domain shift. Further, we employ various XAI techniques to identify specific signals which can be leveraged to predict sex. Finally, we propose a technique, which we callremoval via balancing, to quantify how much a signal contributes to the classification performance. Using this technique and the signals identified, we are able to explain [~]15% of the total performance for radiology and [~]42% of the total performance for dermatology. We envision DREAM to be broadly applicable to other modalities and demographic attributes. This analysis not only underscores the importance of cautious AI application in healthcare but also opens avenues for improving the transparency and reliability of AI-driven diagnostic tools.

Imaging-aided diagnosis and treatment based on artificial intelligence for pulmonary nodules: A review.

Gao H, Li J, Wu Y, Tang Z, He X, Zhao F, Chen Y, He X

pubmed logopapersJul 21 2025
Pulmonary nodules are critical indicators for the early detection of lung cancer; however, their diagnosis and management pose significant challenges due to the variability in nodule characteristics, reader fatigue, and limited clinical expertise, often leading to diagnostic errors. The rapid advancement of artificial intelligence (AI) presents promising solutions to address these issues. This review compares traditional rule-based methods, handcrafted feature-based machine learning, radiomics, deep learning, and hybrid models incorporating Transformers or attention mechanisms. It systematically compares their methodologies, clinical applications (diagnosis, treatment, prognosis), and dataset usage to evaluate performance, applicability, and limitations in pulmonary nodule management. AI advances have significantly improved pulmonary nodule management, with transformer-based models achieving leading accuracy in segmentation, classification, and subtyping. The fusion of multimodal imaging CT, PET, and MRI further enhances diagnostic precision. Additionally, AI aids treatment planning and prognosis prediction by integrating radiomics with clinical data. Despite these advances, challenges remain, including domain shift, high computational demands, limited interpretability, and variability across multi-center datasets. Artificial intelligence (AI) has transformative potential in improving the diagnosis and treatment of lung nodules, especially in improving the accuracy of lung cancer treatment and patient prognosis, where significant progress has been made.

An Improved Diagnostic Deep Learning Model for Cervical Lymphadenopathy Characterization.

Gong W, Li M, Wang S, Jiang Y, Wu J, Li X, Ma C, Luo H, Zhou H

pubmed logopapersJul 21 2025
To validate the diagnostic performance of a B-mode ultrasound-based deep learning (DL) model in distinguishing benign and malignant cervical lymphadenopathy (CLP). A total of 210 CLPs with conclusive pathological results were retrospectively included and separated as training (n = 169) or test cohort (n = 41) randomly at a ratio of 4:1. A DL model integrating convolutional neural network, deformable convolution network and attention mechanism was developed. Three diagnostic models were developed: (a) Model I, CLPs with at least one suspicious B-mode ultrasound feature (ratio of longitudinal to short diameter < 2, irregular margin, hyper-echogenicity, hilus absence, cystic necrosis and calcification) were deemed malignant; (b) Model II: total risk score of B-mode ultrasound features obtained by multivariate logistic regression and (c) Model III: CLPs with positive DL output are deemed malignant. The diagnostic utility of these models was assessed by the area under the receiver operating curve (AUC) and corresponding sensitivity and specificity. Multivariate analysis indicated that DL positive result was the most important factor associated with malignant CLPs [odds ratio (OR) = 39.05, p < 0.001], only followed by hilus absence (OR = 6.01, p = 0.001) in the training cohort. In the test cohort, the AUC of the DL model (0.871) was significantly higher than that in model I (AUC = 0.681, p = 0.04) and model II (AUC = 0.679, p = 0.03), respectively. In addition, model III obtained 93.3% specificity, which was significantly higher than that in model I (40.0%, p = 0.002) and model II (60.0%, p = 0.03), respectively. Although the sensitivity of model I was the highest, it did not show a significant difference compared to that of model III (96.2% vs.80.8%, p = 0.083). B-mode ultrasound-based DL is a potentially robust tool for the differential diagnosis of benign and malignant CLPs.

Mammo-SAE: Interpreting Breast Cancer Concept Learning with Sparse Autoencoders

Krishna Kanth Nakka

arxiv logopreprintJul 21 2025
Interpretability is critical in high-stakes domains such as medical imaging, where understanding model decisions is essential for clinical adoption. In this work, we introduce Sparse Autoencoder (SAE)-based interpretability to breast imaging by analyzing {Mammo-CLIP}, a vision--language foundation model pretrained on large-scale mammogram image--report pairs. We train a patch-level \texttt{Mammo-SAE} on Mammo-CLIP to identify and probe latent features associated with clinically relevant breast concepts such as \textit{mass} and \textit{suspicious calcification}. Our findings reveal that top activated class level latent neurons in the SAE latent space often tend to align with ground truth regions, and also uncover several confounding factors influencing the model's decision-making process. Additionally, we analyze which latent neurons the model relies on during downstream finetuning for improving the breast concept prediction. This study highlights the promise of interpretable SAE latent representations in providing deeper insight into the internal workings of foundation models at every layer for breast imaging.

The added value for MRI radiomics and deep-learning for glioblastoma prognostication compared to clinical and molecular information

D. Abler, O. Pusterla, A. Joye-Kühnis, N. Andratschke, M. Bach, A. Bink, S. M. Christ, P. Hagmann, B. Pouymayou, E. Pravatà, P. Radojewski, M. Reyes, L. Ruinelli, R. Schaer, B. Stieltjes, G. Treglia, W. Valenzuela, R. Wiest, S. Zoergiebel, M. Guckenberger, S. Tanadini-Lang, A. Depeursinge

arxiv logopreprintJul 21 2025
Background: Radiomics shows promise in characterizing glioblastoma, but its added value over clinical and molecular predictors has yet to be proven. This study assessed the added value of conventional radiomics (CR) and deep learning (DL) MRI radiomics for glioblastoma prognosis (<= 6 vs > 6 months survival) on a large multi-center dataset. Methods: After patient selection, our curated dataset gathers 1152 glioblastoma (WHO 2016) patients from five Swiss centers and one public source. It included clinical (age, gender), molecular (MGMT, IDH), and baseline MRI data (T1, T1 contrast, FLAIR, T2) with tumor regions. CR and DL models were developed using standard methods and evaluated on internal and external cohorts. Sub-analyses assessed models with different feature sets (imaging-only, clinical/molecular-only, combined-features) and patient subsets (S-1: all patients, S-2: with molecular data, S-3: IDH wildtype). Results: The best performance was observed in the full cohort (S-1). In external validation, the combined-feature CR model achieved an AUC of 0.75, slightly, but significantly outperforming clinical-only (0.74) and imaging-only (0.68) models. DL models showed similar trends, though without statistical significance. In S-2 and S-3, combined models did not outperform clinical-only models. Exploratory analysis of CR models for overall survival prediction suggested greater relevance of imaging data: across all subsets, combined-feature models significantly outperformed clinical-only models, though with a modest advantage of 2-4 C-index points. Conclusions: While confirming the predictive value of anatomical MRI sequences for glioblastoma prognosis, this multi-center study found standard CR and DL radiomics approaches offer minimal added value over demographic predictors such as age and gender.

An ensemble multimodal approach for predicting first episode psychosis using structural MRI and cognitive assessments

Zhang, S.

medrxiv logopreprintJul 21 2025
Classification between first episode psychosis (FEP) patients and healthy controls is of particular interest to the study of schizophrenia. However, predicting psychosis with cognitive assessments alone is prone to human errors and often lacks biological evidence to back up the findings. In this work, we combined a multimodal dataset of structural MRI and cognitive data to disentangle the detection of first-episode psychosis with a machine learning approach. For this purpose, we proposed a robust detection pipeline that explores the variables in high-order feature space. We applied the pipeline to Human Connectome Project for Early Psychosis (HCP-EP) dataset with 108 participants in EP and 47 controls. The pipeline demonstrated strong performance with 74.67% balanced accuracy on this task. Further feature analysis shows that the model is capable of identifying verified causative biological factors for the occurrence of psychosis based on volumetric MRI measurements, which suggests the potential of data-driven approaches for the search for neuroimaging biomarkers in future studies.

Artificial intelligence in radiology: diagnostic sensitivity of ChatGPT for detecting hemorrhages in cranial computed tomography scans.

Bayar-Kapıcı O, Altunışık E, Musabeyoğlu F, Dev Ş, Kaya Ö

pubmed logopapersJul 21 2025
Chat Generative Pre-trained Transformer (ChatGPT)-4V, a large language model developed by OpenAI, has been explored for its potential application in radiology. This study assesses ChatGPT-4V's diagnostic performance in identifying various types of intracranial hemorrhages in non-contrast cranial computed tomography (CT) images. Intracranial hemorrhages were presented to ChatGPT using the clearest 2D imaging slices. The first question, "Q1: Which imaging technique is used in this image?" was asked to determine the imaging modality. ChatGPT was then prompted with the second question, "Q2: What do you see in this image and what is the final diagnosis?" to assess whether the CT scan was normal or showed pathology. For CT scans containing hemorrhage that ChatGPT did not interpret correctly, a follow-up question-"Q3: There is bleeding in this image. Which type of bleeding do you see?"-was used to evaluate whether this guidance influenced its response. ChatGPT accurately identified the imaging technique (Q1) in all cases but demonstrated difficulty diagnosing epidural hematoma (EDH), subdural hematoma (SDH), and subarachnoid hemorrhage (SAH) when no clues were provided (Q2). When a hemorrhage clue was introduced (Q3), ChatGPT correctly identified EDH in 16.7% of cases, SDH in 60%, and SAH in 15.6%, and achieved 100% diagnostic accuracy for hemorrhagic cerebrovascular disease. Its sensitivity, specificity, and accuracy for Q2 were 23.6%, 92.5%, and 57.4%, respectively. These values improved substantially with the clue in Q3, with sensitivity rising to 50.9% and accuracy to 71.3%. ChatGPT also demonstrated higher diagnostic accuracy in larger hemorrhages in EDH and SDH images. Although the model performs well in recognizing imaging modalities, its diagnostic accuracy substantially improves when guided by additional contextual information. These findings suggest that ChatGPT's diagnostic performance improves with guided prompts, highlighting its potential as a supportive tool in clinical radiology.
Page 106 of 2402393 results
Show
per page

Ready to Sharpen Your Edge?

Join hundreds of your peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.