Sort by:
Page 106 of 3513502 results

Multivariate Fields of Experts

Stanislas Ducotterd, Michael Unser

arxiv logopreprintAug 8 2025
We introduce the multivariate fields of experts, a new framework for the learning of image priors. Our model generalizes existing fields of experts methods by incorporating multivariate potential functions constructed via Moreau envelopes of the $\ell_\infty$-norm. We demonstrate the effectiveness of our proposal across a range of inverse problems that include image denoising, deblurring, compressed-sensing magnetic-resonance imaging, and computed tomography. The proposed approach outperforms comparable univariate models and achieves performance close to that of deep-learning-based regularizers while being significantly faster, requiring fewer parameters, and being trained on substantially fewer data. In addition, our model retains a relatively high level of interpretability due to its structured design.

Artificial intelligence in radiology, nuclear medicine and radiotherapy: Perceptions, experiences and expectations from the medical radiation technologists in Central and South America.

Mendez-Avila C, Torre S, Arce YV, Contreras PR, Rios J, Raza NO, Gonzalez H, Hernandez YC, Cabezas A, Lucero M, Ezquerra V, Malamateniou C, Solis-Barquero SM

pubmed logopapersAug 8 2025
Artificial intelligence (AI) has been growing in the field of medical imaging and clinical practice. It is essential to comprehend the perceptions, experiences, and expectations regarding AI implementation among medical radiation technologists (MRTs) working in radiology, nuclear medicine, and radiotherapy. Some global studies tend to inform about AI implementation, but there is almost no information from Central and South American professionals. This study aimed to understand the perceptions of the impact of AI on the MRTs, as well as the varying experiences and expectations these professionals have regarding its implementation. An online survey was conducted among Central and South American MRTs for the collection of qualitative data concerning the primary perceptions regarding the implementation of AI in radiology, nuclear medicine, and radiotherapy. The analysis considered descriptive statistics in closed-ended questions and dimension codification for open-ended responses. A total of 398 valid responses were obtained, and it was determined that 98.5 % (n = 392) of the respondents agreed with the implementation of AI in clinical practice. The primary contributions of AI that were identified were the optimization of processes, greater diagnostic accuracy, and the possibility of job expansion. On the other hand, concerns were raised regarding the delay in providing training opportunities and limited avenues for learning in this domain, the displacement of roles, and dehumanization in clinical practice. This sample of participants likely represents mostly professionals who have more AI knowledge than others. It is therefore important to interpret these results with caution. Our findings indicate strong professional confidence in AI's capacity to improve imaging quality while maintaining patient safety standards. However, user resistance may disturb implementation efforts. Our results highlight the dual need for (a) comprehensive professional training programs and (b) user education initiatives that demonstrate AI's clinical value in radiology. We therefore recommend a carefully structured, phased AI implementation approach, guided by evidence-based guidelines and validated training protocols from existing research. AI is already present in medical imaging, but its effective implementations depend on building acceptance and trust through education and training, enabling MRTs to use it safely for patient benefit.

Ensemble deep learning model for early diagnosis and classification of Alzheimer's disease using MRI scans.

Robinson Jeyapaul S, Kombaiya S, Jeya Kumar AK, Stanley VJ

pubmed logopapersAug 8 2025
BackgroundAlzheimer's disease (AD) is an irreversible neurodegenerative disorder characterized by progressive cognitive and memory decline. Accurate prediction of high-risk individuals enables early detection and better patient care.ObjectiveThis study aims to enhance MRI-based AD classification through advanced image preprocessing, optimal feature selection, and ensemble deep learning techniques.MethodsThe study employs advanced image preprocessing techniques such as normalization, affine transformation, and denoising to improve MRI quality. Brain structure segmentation is performed using the adaptive DeepLabV3 + approach for precise AD diagnosis. A novel optimal feature selection framework, H-IBMFO, integrates the Improved Beluga Whale Optimizer and Manta Foraging Optimization. An ensemble deep learning model combining MobileNet V2, DarkNet, and ResNet is used for classification. MATLAB is utilized for implementation.ResultsThe proposed system achieves 98.7% accuracy, with 98% precision, 98% sensitivity, 99% specificity, and 98% F-measure, demonstrating superior classification performance with minimal false positives and negatives.ConclusionsThe study establishes an efficient framework for AD classification, significantly improving early detection through optimized feature selection and deep learning. The high accuracy and reliability of the system validate its effectiveness in diagnosing AD stages.

Thyroid Volume Measurement With AI-Assisted Freehand 3D Ultrasound Compared to 2D Ultrasound-A Clinical Trial.

Rask KB, Makouei F, Wessman MHJ, Kristensen TT, Todsen T

pubmed logopapersAug 8 2025
Accurate thyroid volume assessment is critical in thyroid disease diagnostics, yet conventional high-resolution 2D ultrasound has limitations. Freehand 3D ultrasound with AI-assisted segmentation presents a potential advancement, but its clinical accuracy requires validation. This prospective clinical trial included 14 patients scheduled for total thyroidectomy. Preoperative thyroid volume was measured using both 2D ultrasound (ellipsoid method) and freehand 3D ultrasound with AI segmentation. Postoperative thyroid volume, determined via the water displacement method, served as the reference standard. The median postoperative thyroid volume was 14.8 mL (IQR 8.8-20.2). The median volume difference was 1.7 mL (IQR 1.2-3.3) for 3D ultrasound and 3.6 mL (IQR 2.3-6.6) for 2D ultrasound (p = 0.02). The inter-operator reliability coefficient for 3D ultrasound was 0.986 (p < 0.001). These findings suggest that freehand 3D ultrasound with AI-assisted segmentation provides superior accuracy and reproducibility compared to 2D ultrasound and may enhance clinical thyroid volume assessment. ClinicalTrials.gov identifier: NCT05510609.

A Cohort Study of Pediatric Severe Community-Acquired Pneumonia Involving AI-Based CT Image Parameters and Electronic Health Record Data.

He M, Yuan J, Liu A, Pu R, Yu W, Wang Y, Wang L, Nie X, Yi J, Xue H, Xie J

pubmed logopapersAug 8 2025
Community-acquired pneumonia (CAP) is a significant concern for children worldwide and is associated with a high morbidity and mortality. To improve patient outcomes, early intervention and accurate diagnosis are essential. Artificial intelligence (AI) can mine and label imaging data and thus may contribute to precision research and personalized clinical management. The baseline characteristics of 230 children with severe CAP hospitalized from January 2023 to October 2024 were retrospectively analyzed. The patients were divided into two groups according to the presence of respiratory failure. The predictive ability of AI-derived chest CT (computed tomography) indices alone for respiratory failure was assessed via logistic regression analysis. ROC (receiver operating characteristic) curves were plotted for these regression models. After adjusting for age, white blood cell count, neutrophils, lymphocytes, creatinine, wheezing, and fever > 5 days, a greater number of involved lung lobes [odds ratio 1.347, 95% confidence interval (95% CI) 1.036-1.750, P = 0.026] and bilateral lung involvement (odds ratio 2.734, 95% CI 1.084-6.893, P = 0.033) were significantly associated with respiratory failure. The discriminatory power (as measured by the area under curve) of Model 2 and Model 3, which included electronic health record data and the accuracy of CT imaging features, was better than that of Model 0 and Model 1, which contained only the chest CT parameters. The sensitivity and specificity of Model 2 at the optimal critical value (0.441) were 84.3% and 59.8%, respectively. The sensitivity and specificity of Model 3 at the optimal critical value (0.446) were 68.6% and 76.0%, respectively. The use of AI-derived chest CT indices may achieve high diagnostic accuracy and guide precise interventions for patients with severe CAP. However, clinical, laboratory, and AI-derived chest CT indices should be included to accurately predict and treat severe CAP.

Development and validation of a transformer-based deep learning model for predicting distant metastasis in non-small cell lung cancer using <sup>18</sup>FDG PET/CT images.

Hu N, Luo Y, Tang M, Yan G, Yuan S, Li F, Lei P

pubmed logopapersAug 8 2025
This study aimed to develop and validate a hybrid deep learning (DL) model that integrates convolutional neural network (CNN) and vision transformer (ViT) architectures to predict distant metastasis (DM) in patients with non-small cell lung cancer (NSCLC) using <sup>18</sup>F-FDG PET/CT images. A retrospective analysis was conducted on a cohort of consecutively registered patients who were newly diagnosed and untreated for NSCLC. A total of 167 patients with available PET/CT images were included in the analysis. DL features were extracted using a combination of CNN and ViT architectures, followed by feature selection, model construction, and evaluation of model performance using the receiver operating characteristic (ROC) and the area under the curve (AUC). The ViT-based DL model exhibited strong predictive capabilities in both the training and validation cohorts, achieving AUCs of 0.824 and 0.830 for CT features, and 0.602 and 0.694 for PET features, respectively. Notably, the model that integrated both PET and CT features demonstrated a notable AUC of 0.882 in the validation cohort, outperforming models that utilized either PET or CT features alone. Furthermore, this model outperformed the CNN model (ResNet 50), which achieved an AUC of 0.752 [95% CI 0.613, 0.890], p < 0.05. Decision curve analysis further supported the efficacy of the ViT-based DL model. The ViT-based DL developed in this study demonstrates considerable potential in predicting DM in patients with NSCLC, potentially informing the creation of personalized treatment strategies. Future validation through prospective studies with larger cohorts is necessary.

Machine learning diagnostic model for amyotrophic lateral sclerosis analysis using MRI-derived features.

Gil Chong P, Mazon M, Cerdá-Alberich L, Beser Robles M, Carot JM, Vázquez-Costa JF, Martí-Bonmatí L

pubmed logopapersAug 8 2025
Amyotrophic Lateral Sclerosis is a devastating motor neuron disease characterized by its diagnostic difficulty. Currently, no reliable biomarkers exist in the diagnosis process. In this scenario, our purpose is the application of machine learning algorithms to imaging MRI-derived variables for the development of diagnostic models that facilitate and shorten the process. A dataset of 211 patients (114 ALS, 45 mimic, 22 genetic carriers and 30 control) with MRI-derived features of volumetry, cortical thickness and local iron (via T2* mapping, and visual assessment of susceptibility imaging). A binary classification task approach has been taken to classify patients with and without ALS. A sequential modeling methodology, understood from an iterative improvement perspective, has been followed, analyzing each group's performance separately to adequately improve modelling. Feature filtering techniques, dimensionality reduction techniques (PCA, kernel PCA), oversampling techniques (SMOTE, ADASYN) and classification techniques (logistic regression, LASSO, Ridge, ElasticNet, Support Vector Classifier, K-neighbors, random forest) were included. Three subsets of available data have been used for each proposed architecture: a subset containing automatic retrieval MRI-derived data, a subset containing the variables from the visual analysis of the susceptibility imaging and a subset containing all features. The best results have been attained with all the available data through a voting classifier composed of five different classifiers: accuracy = 0.896, AUC = 0.929, sensitivity = 0.886, specificity = 0.929. These results confirm the potential of ML techniques applied to imaging variables of volumetry, cortical thickness, and local iron for the development of diagnostic model as a clinical tool for decision-making support.

GAN-MRI enhanced multi-organ MRI segmentation: a deep learning perspective.

Channarayapatna Srinivasa A, Bhat SS, Baduwal D, Sim ZTJ, Patil SS, Amarapur A, Prakash KNB

pubmed logopapersAug 8 2025
Clinical magnetic resonance imaging (MRI) is a high-resolution tool widely used for detailed anatomical imaging. However, prolonged scan times often lead to motion artefacts and patient discomfort. Fast acquisition techniques can reduce scan times but often produce noisy, low-contrast images, compromising segmentation accuracy essential for diagnosis and treatment planning. To address these limitations, we developed an end-to-end framework that incorporates BIDS-based data organiser and anonymizer, a GAN-based MR image enhancement model (GAN-MRI), AssemblyNet for brain region segmentation, and an attention-residual U-Net with Guided loss for abdominal and thigh segmentation. Thirty brain scans (5,400 slices) and 32 abdominal (1,920 slices) and 55 thigh scans (2,200 slices) acquired from multiple MRI scanners (GE, Siemens, Toshiba) underwent evaluation. Image quality improved significantly, with SNR and CNR for brain scans increasing from 28.44 to 42.92 (p < 0.001) and 11.88 to 18.03 (p < 0.001), respectively. Abdominal scans exhibited SNR increases from 35.30 to 50.24 (p < 0.001) and CNR from 10,290.93 to 93,767.22 (p < 0.001). Double-blind evaluations highlighted improved visualisations of anatomical structures and bias field correction. Segmentation performance improved substantially in the thigh (muscle: + 21%, IMAT: + 9%) and abdominal regions (SSAT: + 1%, DSAT: + 2%, VAT: + 12%), while brain segmentation metrics remained largely stable, reflecting the robustness of the baseline model. Proposed framework is designed to handle data from multiple anatomies with variations from different MRI scanners and centres by enhancing MRI scan and improving segmentation accuracy, diagnostic precision and treatment planning while reducing scan times and maintaining patient comfort.

GPT-4 for automated sequence-level determination of MRI protocols based on radiology request forms from clinical routine.

Terzis R, Kaya K, Schömig T, Janssen JP, Iuga AI, Kottlors J, Lennartz S, Gietzen C, Gözdas C, Müller L, Hahnfeldt R, Maintz D, Dratsch T, Pennig L

pubmed logopapersAug 8 2025
This study evaluated GPT-4's accuracy in MRI sequence selection based on radiology request forms (RRFs), comparing its performance to radiology residents. This retrospective study included 100 RRFs across four subspecialties (cardiac imaging, neuroradiology, musculoskeletal, and oncology). GPT-4 and two radiology residents (R1: 2 years, R2: 5 years MRI experience) selected sequences based on each patient's medical history and clinical questions. Considering imaging society guidelines, five board-certified specialized radiologists assessed protocols based on completeness, quality, and utility in consensus, using 5-point Likert scales. Clinical applicability was rated binarily by the institution's lead radiographer. GPT-4 achieved median scores of 3 (1-5) for completeness, 4 (1-5) for quality, and 4 (1-5) for utility, comparable to R1 (3 (1-5), 4 (1-5), 4 (1-5); each p > 0.05) but inferior to R2 (4 (1-5), 5 (1-5); p < 0.01, respectively, and 5 (1-5); p < 0.001). Subspecialty protocol quality varied: GPT-4 matched R1 (4 (2-4) vs. 4 (2-5), p = 0.20) and R2 (4 (2-5); p = 0.47) in cardiac imaging; showed no differences in neuroradiology (all 5 (1-5), p > 0.05); scored lower than R1 and R2 in musculoskeletal imaging (3 (2-5) vs. 4 (3-5); p < 0.01, and 5 (3-5); p < 0.001); and matched R1 (4 (1-5) vs. 2 (1-4), p = 0.12) as well as R2 (5 (2-5); p = 0.20) in oncology. GPT-4-based protocols were clinically applicable in 95% of cases, comparable to R1 (95%) and R2 (96%). GPT-4 generated MRI protocols with notable completeness, quality, utility, and clinical applicability, excelling in standardized subspecialties like cardiac and neuroradiology imaging while yielding lower accuracy in musculoskeletal examinations. Question Long MRI acquisition times limit patient access, making accurate protocol selection crucial for efficient diagnostics, though it's time-consuming and error-prone, especially for inexperienced residents. Findings GPT-4 generated MRI protocols of remarkable yet inconsistent quality, performing on par with an experienced resident in standardized fields, but moderately in musculoskeletal examinations. Clinical relevance The large language model can assist less experienced radiologists in determining detailed MRI protocols and counteract increasing workloads. The model could function as a semi-automatic tool, generating MRI protocols for radiologists' confirmation, optimizing resource allocation, and improving diagnostics and cost-effectiveness.

An Anisotropic Cross-View Texture Transfer with Multi-Reference Non-Local Attention for CT Slice Interpolation.

Uhm KH, Cho H, Hong SH, Jung SW

pubmed logopapersAug 8 2025
Computed tomography (CT) is one of the most widely used non-invasive imaging modalities for medical diagnosis. In clinical practice, CT images are usually acquired with large slice thicknesses due to the high cost of memory storage and operation time, resulting in an anisotropic CT volume with much lower inter-slice resolution than in-plane resolution. Since such inconsistent resolution may lead to difficulties in disease diagnosis, deep learning-based volumetric super-resolution methods have been developed to improve inter-slice resolution. Most existing methods conduct single-image super-resolution on the through-plane or synthesize intermediate slices from adjacent slices; however, the anisotropic characteristic of 3D CT volume has not been well explored. In this paper, we propose a novel cross-view texture transfer approach for CT slice interpolation by fully utilizing the anisotropic nature of 3D CT volume. Specifically, we design a unique framework that takes high-resolution in-plane texture details as a reference and transfers them to low-resolution through-plane images. To this end, we introduce a multi-reference non-local attention module that extracts meaningful features for reconstructing through-plane high-frequency details from multiple in-plane images. Through extensive experiments, we demonstrate that our method performs significantly better in CT slice interpolation than existing competing methods on public CT datasets including a real-paired benchmark, verifying the effectiveness of the proposed framework. The source code of this work is available at https://github.com/khuhm/ACVTT.
Page 106 of 3513502 results
Show
per page

Ready to Sharpen Your Edge?

Join hundreds of your peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.