Sort by:
Page 21 of 2352341 results

An attention aided wavelet convolutional neural network for lung nodule characterization.

Halder A

pubmed logopapersSep 21 2025
Lung cancer is a leading cause of cancer-related mortality worldwide, necessitating the development of accurate and efficient diagnostic methods. Early detection and accurate characterization of pulmonary nodules significantly influence patient prognosis and treatment planning and can improve the five-year survival rate. However, distinguishing benign from malignant nodules using conventional imaging techniques remain a clinical challenge due to subtle structural similarities. Therefore, to address this issue, this study proposes a novel two-pathway wavelet-based deep learning computer-aided diagnosis (CADx) framework forimproved lung nodule classification using high-resolution computed tomography (HRCT) images. The proposed Wavelet-based Lung Cancer Detection Network (WaveLCDNet) is capable of characterizing lung nodules images through a hierarchical feature extraction pipeline consisting of convolutional neural network (CNN) blocks and trainable wavelet blocks for multi-resolution analysis. The introduced wavelet block can capture both spatial and frequency-domain information, preserving fine-grained texture details essential for nodule characterization. Additionally, in this work, convolutional block attention module (CBAM) based attention mechanism has been introduced to enhance discriminative feature learning. The extracted features from both pathways are adaptively fused and processed using global average pooling (GAP) operation. The introduced WaveLCDNet is trained and evaluated on the publicly accessible LIDC-IDRI dataset and achieved sensitivity, specificity, accuracy of 96.89%, 95.52%, and 96.70% for nodule characterization. In addition, the developed framework was externally validated on the Kaggle DSB2017 test dataset, achieving 95.90% accuracy with a Brier Score of 0.0215 for lung nodule characterization, reinforcing its reliability across independent imaging sources and its practical value for integration into real-world diagnostic workflows. By effectively combining multi-scale convolutional filtering with wavelet-based multi-resolution analysisand attention mechanisms, the introduced framework outperforms different recent most state-of-the-art deep learning models and offers a promising CADx solution forenhancing lung cancer screening early diagnosis in clinical settings.

Benign vs malignant tumors classification from tumor outlines in mammography scans using artificial intelligence techniques.

Beni HM, Asaei FY

pubmed logopapersSep 21 2025
Breast cancer is one of the most important causes of death among women due to cancer. With the early diagnosis of this condition, the probability of survival will increase. For this purpose, medical imaging methods, especially mammography, are used for screening and early diagnosis of breast abnormalities. The main goal of this study is to distinguish benign or malignant tumors based on tumor morphology features extracted from tumor outlines extracted from mammography images. Unlike previous studies, this study does not use the mammographic image itself but only extracts the exact outline of the tumor. These outlines were extracted from a new and publicly available mammography database published in 2024. The features outlines were calculated using known pre-trained Convolutional Neural Networks (CNN), including VGG16, ResNet50, Xception65, AlexNet, DenseNet, GoogLeNet, Inception-v3, and a combination of them to improve performance. These pre-trained networks have been used in many studies in various fields. In the classification part, known Machine Learning (ML) algorithms, such as Support Vector Machine (SVM), K-Nearest Neighbor (KNN), Neural Network (NN), Naïve Bayes (NB), Decision Tree (DT), and a combination of them have been compared in outcome measures, namely accuracy, specificity, sensitivity, and precision. Also, with the use of data augmentation, the dataset size was increased about 6-8 times, and the K-fold cross-validation technique (K = 5) was used in this study. Based on the performed simulations, a combination of the features from all pre-trained deep networks and the NB classifier resulted in the best possible outcomes with 88.13 % accuracy, 92.52 % specificity, 83.73 % sensitivity, and 92.04 % precision. Furthermore, validation on DMID dataset using ResNet50 features along with NB classifier, led to 92.03 % accuracy, 95.57 % specificity, 88.49 % sensitivity, and 95.23 % precision. This study sheds light on using AI algorithms to prevent biopsy tests and speed up breast cancer tumor classification using tumor outlines in mammographic images.

Causal Representation Learning from Multimodal Clinical Records under Non-Random Modality Missingness

Zihan Liang, Ziwen Pan, Ruoxuan Xiong

arxiv logopreprintSep 21 2025
Clinical notes contain rich patient information, such as diagnoses or medications, making them valuable for patient representation learning. Recent advances in large language models have further improved the ability to extract meaningful representations from clinical texts. However, clinical notes are often missing. For example, in our analysis of the MIMIC-IV dataset, 24.5% of patients have no available discharge summaries. In such cases, representations can be learned from other modalities such as structured data, chest X-rays, or radiology reports. Yet the availability of these modalities is influenced by clinical decision-making and varies across patients, resulting in modality missing-not-at-random (MMNAR) patterns. We propose a causal representation learning framework that leverages observed data and informative missingness in multimodal clinical records. It consists of: (1) an MMNAR-aware modality fusion component that integrates structured data, imaging, and text while conditioning on missingness patterns to capture patient health and clinician-driven assignment; (2) a modality reconstruction component with contrastive learning to ensure semantic sufficiency in representation learning; and (3) a multitask outcome prediction model with a rectifier that corrects for residual bias from specific modality observation patterns. Comprehensive evaluations across MIMIC-IV and eICU show consistent gains over the strongest baselines, achieving up to 13.8% AUC improvement for hospital readmission and 13.1% for ICU admission.

Advanced Ultrasound Quantitative Analysis in Hepatology: A Systematic Review of Methodologies for Characterizing Focal Liver Lesions.

Li S, Liu H, Li W, Gao X

pubmed logopapersSep 21 2025
This systematic review evaluates advanced ultrasound quantitative techniques including contrast-enhanced ultrasound, elastography, quantitative ultrasound (QUS), multiparametric ultrasound, and artificial intelligence for characterizing focal liver lesions (FLLs). It critically appraises their technical principles, parameter extraction methodologies, and clinical validation frameworks. It further integrates and comparatively analyzes their diagnostic performance across major FLL subtypes, including hepatocellular carcinoma, metastases, hemangioma, and focal nodular hyperplasia. This work provides a foundation for improving noninvasive FLL diagnosis and highlights the imperative for standardization and clinical translation of advanced QUS in hepatology.

A Comparative Evaluation of Meta-Learning Models for Few-Shot Chest X-Ray Disease Classification.

Quiñonez-Baca LC, Ramirez-Alonso G, Gaxiola F, Manzo-Martinez A, Cornejo R, Lopez-Flores DR

pubmed logopapersSep 21 2025
<b>Background/Objectives</b>: The limited availability of labeled data, particularly in the medical domain, poses a significant challenge for training accurate diagnostic models. While deep learning techniques have demonstrated notable efficacy in image-based tasks, they require large annotated datasets. In data-scarce scenarios-especially involving rare diseases-their performance deteriorates significantly. Meta-learning offers a promising alternative by enabling models to adapt quickly to new tasks using prior knowledge and only a few labeled examples. This study aims to evaluate the effectiveness of representative meta-learning models for thoracic disease classification in chest X-rays. <b>Methods</b>: We conduct a comparative evaluation of four meta-learning models: Prototypical Networks, Relation Networks, MAML, and FoMAML. First, we assess five backbone architectures (ConvNeXt, DenseNet-121, ResNet-50, MobileNetV2, and ViT) using a Prototypical Network. The best-performing backbone is then used across all meta-learning models for fair comparison. Experiments are performed on the ChestX-ray14 dataset under a 2-way setting with multiple <i>k</i>-shot configurations. <b>Results</b>: Prototypical Networks combined with DenseNet-121 achieved the best performance, with a recall of 68.1%, an F1-score of 67.4%, and a precision of 0.693 in the 2-way, 10-shot configuration. In a disease-specific analysis, Hernia obtains the best classification results. Furthermore, Prototypical and Relation Networks demonstrate significantly higher computational efficiency, requiring fewer FLOPs and shorter execution times than MAML and FoMAML. <b>Conclusions</b>: Prototype-based meta-learning, particularly with DenseNet-121, proves to be a robust and computationally efficient approach for few-shot chest X-ray disease classification. These findings highlight its potential for real-world clinical applications, especially in scenarios with limited annotated medical data.

Chest computed tomography-based artificial intelligence-aided latent class analysis for diagnosis of severe pneumonia.

Chu C, Guo Y, Lu Z, Gui T, Zhao S, Cui X, Lu S, Jiang M, Li W, Gao C

pubmed logopapersSep 20 2025
There is little literature describing the artificial intelligence (AI)-aided diagnosis of severe pneumonia (SP) subphenotypes and the association of the subphenotypes with the ventilatory treatment efficacy. The aim of our study is to illustrate whether clinical and biological heterogeneity, such as ventilation and gas-exchange, exists among patients with SP using chest computed tomography (CT)-based AI-aided latent class analysis (LCA). This retrospective study included 413 patients hospitalized at Xinhua Hospital diagnosed with SP from June 1, 2015 to May 30, 2020. AI quantification results of chest CT and their combination with additional clinical variables were used to develop LCA models in an SP population. The optimal subphenotypes were determined though evaluating statistical indicators of all the LCA models, and clinical implications of them such as guiding ventilation strategies were further explored by statistical methods. The two-class LCA model based on AI quantification results of chest CT can describe the biological characteristics of the SP population well and hence yielded the two clinical subphenotypes. Patients with subphenotype-1 had milder infections ( P <0.001) than patients with subphenotype-2 and had lower 30-day ( P <0.001) and 90-day ( P <0.001) mortality, and lower in-hospital ( P = 0.001) and 2-year ( P <0.001) mortality. Patients with subphenotype-1 showed a better match between the percentage of non-infected lung volume (used to quantify ventilation) and oxygen saturation (used to reflect gas exchange), compared with patients with subphenotype-2. There were significant differences in the matching degree of lung ventilation and gas exchange between the two subphenotypes ( P <0.001). Compared with patients with subphenotype-2, those with subphenotype-1 showed a relatively better match between CT-based AI metrics of the non-infected region and oxygenation, and their clinical outcomes were effectively improved after receiving invasive ventilation treatment. A two-class LCA model based on AI quantification results of chest CT in the SP population particularly revealed clinical heterogeneity of lung function. Identifying the degree of match between ventilation and gas-exchange may help guide decisions about assisted ventilation.

Predictive Analysis of Neoadjuvant Chemotherapy Efficacy in Breast Cancer Using Multi-Region Ultrasound Imaging Features Combined With Pathological Parameters.

Wei C, Jia Y, Gu Y, He Z, Nie F

pubmed logopapersSep 20 2025
This study aimed to analyze the correlation between the ultrasonographic radiomic features of multiple regions within and surrounding the primary tumor in breast cancer patients prior to receiving neoadjuvant chemotherapy (NAC) and the efficacy of NAC. By integrating clinical and pathological parameters, a predictive model was constructed to provide an accurate basis for personalized treatment and precise prognosis in breast cancer patients. This retrospective study included 321 breast cancer patients who underwent NAC treatment at the Second Hospital of Lanzhou University from January 2019 to December 2024. According to post-operative pathological results, the patients were divided into pathological complete response (PCR) and non-pathological complete response (non-PCR) groups. Regions of interest were outlined on 2-D ultrasound images using Itk-snap software. The intra-tumor (Intra) region and 5 mm (Peri-5 mm), 10 mm (Peri-10 mm) and 15 mm (Peri-15 mm) the peri-tumoralregions were demarcated, with radiomics features extracted from each region. Patients were randomly divided into a training set (n = 224) and a validation set (n = 97) in a 7:3 ratio. All features underwent Z-score normalization followed by dimensionality reduction using t-tests, Pearson correlation coefficients and least absolute shrinkage and selection operator. Radiomics models for Intra, Peri-5 mm, Peri-10 mm, Peri-15 mm and the combined intra-tumoral and peri-tumoral regions (Intra-tumoral, Peri-tumoral, IntraPeri) were constructed using a random forest machine-learning classifier. The predictive performance of the models was assessed by plotting receiver operating characteristic curves and calculating the area under the curve (AUC). Additionally, calibration curves and decision curve analysis were plotted to evaluate the model's goodness of fit and clinical net benefit RESULTS: A total of 214 radiomics features were extracted from the intra-tumoral and multi-region peri-tumoral areas. Using the least absolute shrinkage and selection operator regression model, eight intra-tumoral radiomics features, eight peri-10 mm radiomics features and nine IntraPeri-10 mm radiomics features were selected as being closely associated with PCR. The AUC of the intra-tumoral model was 0.860 and 0.823 in the training and validation sets, respectively. The AUCs of the peri-5 mm, Peri-10 mm and Peri-15 mm models were 0.836, 0.854 and 0.822 in the training set, and 0.793, 0.799 and 0.792 in the validation set. Among them, the AUC of the IntraPeri-10 mm model in the validation set was 0.842 (95% confidence interval [CI]: 0.764-0.921), which was superior to the AUC of the IntraPeri-5 mm model (0.831; 95% CI: 0.758-0.914) and the IntraPeri-15 mm model (0.838; 95% CI: 0.761-0.917). The combined model based on IntraPeri-10 mm and clinical pathological parameters (HER-2, Ki-67) achieved an AUC of 0.869 (95% CI: 0.800-0.937). The Delong test showed that the AUC of the combined model was significantly superior to that of the other models. The calibration curve indicated that the combined model had a good fit, and decision curve analysis demonstrated that the combined model provided a better clinical net benefit. The peri-10 mm region is the optimal predictive area for the tumor's surrounding tissue after NAC in breast cancer. The IntraPeri-10 mm model, incorporating clinical pathological parameters, performs better at predicting the efficacy of NAC in breast cancer and can accurately assess treatment response, offering valuable guidance for subsequent treatment decisions.

Comparison of Prostate-Specific Membrane Antigen Positron Emission Tomography and Conventional Imaging Modalities in the Detection of Biochemical Recurrence of Prostate Cancer and Assessment of the Role of Artificial Intelligence: A Systematic Review and Meta-analysis.

Zhang H, Xie C, Huang C, Jiang Z, Tang Q

pubmed logopapersSep 20 2025
We conducted a systematic review and meta-analysis to assess and compare the diagnostic performance of prostate-specific membrane antigen positron-emission tomography (PSMA PET) with conventional imaging modalities in detecting biochemical recurrence of prostate cancer, and to assess the role of artificial intelligence in this context. A comprehensive search of PubMed, Embase, Web of Science, the Cochrane Library, and Scopus was conducted for studies, initially on May 7, 2025, and updated on July 28, 2025. Studies that compared PSMA PET with conventional imaging and assessed artificial intelligence for detecting biochemical recurrence of prostate cancer were considered. The QUADAS-2 technique was employed to evaluate study quality. Diagnosis accuracy and detection rates were aggregated utilizing a bivariate random-effects model. A total of 7637 patients from 67 studies were included. PSMA PET demonstrated significantly higher overall diagnostic accuracy for biochemical recurrence of prostate cancer compared to mpMRI, CT, and AI test sets, with accuracy values of (0.89 vs. 0.71, 0.45, and 0.76, P<0.01). For local recurrence, mpMRI outperformed PSMA PET and CT (0.93 vs. 0.84 and 0.77, P<0.01). PSMA PET was superior in detecting lymph node metastasis than mpMRI and CT (0.89 vs. 0.79 and 0.72, P<0.05). For bone metastasis, PSMA PET outperformed mpMRI, CT, and Bone scan (0.95 vs. 0.85, 0.81, and 0.80, P<0.05). For visceral metastasis, PSMA PET outperformed mpMRI (0.96 vs. 0.89, P=0.23), and CT (0.96 vs. 0.78, P<0.05). 21 studies involving 3113 samples were included to evaluate the performance of artificial intelligence in detecting biochemical recurrence of prostate cancer. The pooled sensitivity, specificity, DOR, and AUC of AI test sets in detecting biochemical recurrence of prostate cancer were 0.77, 0.76, 10.39, and 0.79. Heterogeneity limits the generalizability of our findings. PSMA PET outperformed mpMRI and CT in detecting overall, local recurrence, bone, and visceral metastasis, while mpMRI was more effective for local recurrence. While AI exhibits potential diagnostic efficacy. Despite promising results, heterogeneity and limited validation highlight the need for further research to support routine clinical use.

Enhancing the reliability of Alzheimer's disease prediction in MRI images.

Islam J, Furqon EN, Farady I, Alex JSR, Shih CT, Kuo CC, Lin CY

pubmed logopapersSep 19 2025
Alzheimer's Disease (AD) diagnostic procedures employing Magnetic Resonance Imaging (MRI) analysis encounter considerable obstacles pertaining to reliability and accuracy, especially when deep learning models are utilized within clinical environments. Present deep learning methodologies for MRI-based AD detection frequently demonstrate spatial dependencies and exhibit deficiencies in robust validation mechanisms. Extant validation techniques inadequately integrate anatomical knowledge and exhibit challenges in feature interpretability across a range of imaging conditions. To address this fundamental gap, we introduce a reverse validation paradigm that systematically repositions anatomical structures to test whether models recognize features based on anatomical characteristics rather than spatial memorization. Our research endeavors to rectify these shortcomings by proposing three innovative methodologies: Feature Position Invariance (FPI) for the validation of anatomical features, biomarker location augmentation aimed at enhancing spatial learning, and High-Confidence Cohort (HCC) selection for the reliable identification of training samples. The FPI methodology leverages reverse validation approach to substantiate model predictions through the reconstruction of anatomical features, bolstered by our extensive data augmentation strategy and a confidence-based sample selection technique. The application of this framework utilizing YOLO and MobileNet architecture has yielded significant advancements in both binary and three-class AD classification tasks, achieving state-of-the-art accuracy with enhancements of 2-4 % relative to baseline models. Additionally, our methodology generates interpretable insights through anatomy-aligned validation, establishing direct links between model decisions and neuropathological features. Our experimental findings reveal consistent performance across various anatomical presentations, signifying that the framework effectively enhances both the reliability and interpretability of AD diagnosis through MRI analysis, thereby equipping medical professionals with a more robust diagnostic support system.

Multimodal AI-driven Biomarker for Early Detection of Cancer Cachexia

Ahmed, S., Parker, N., Park, M., Davis, E. W., Jeong, D., Permuth, J. B., Schabath, M. B., Yilmaz, Y., Rasool, G.

medrxiv logopreprintSep 19 2025
Cancer cachexia, a multifactorial metabolic syndrome characterized by severe muscle wasting and weight loss, contributes to poor outcomes across various cancer types but lacks a standardized, generalizable biomarker for early detection. We present a multimodal AI-based biomarker trained on real-world clinical, radiologic, laboratory, and unstructured clinical note data, leveraging foundation models and large language models (LLMs) to identify cachexia at the time of cancer diagnosis. Prediction accuracy improved with each added modality: 77% using clinical variables alone, 81% with added laboratory data, and 85% with structured symptom features extracted from clinical notes. Incorporating embeddings from clinical text and CT images further improved accuracy to 92%. The framework also demonstrated prognostic utility, improving survival prediction as data modalities were integrated. Designed for real-world clinical deployment, the framework accommodates missing modalities without requiring imputation or case exclusion, supporting scalability across diverse oncology settings. Unlike prior models trained on curated datasets, our approach utilizes standard-of-care clinical data, facilitating integration into oncology workflows. In contrast to fixed-threshold composite indices such as the cachexia index (CXI), the model generates patient-specific predictions, enabling adaptable, cancer-agnostic performance. To enhance clinical reliability and safety, the framework incorporates uncertainty estimation to flag low-confidence cases for expert review. This work advances a clinically applicable, scalable, and trustworthy AI-driven decision support tool for early cachexia detection and personalized oncology care.
Page 21 of 2352341 results
Show
per page

Ready to Sharpen Your Edge?

Join hundreds of your peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.