Sort by:
Page 207 of 3933923 results

Spondyloarthritis Research and Treatment Network (SPARTAN) Clinical and Imaging Year in Review 2024.

Ferrandiz-Espadin R, Liew JW

pubmed logopapersJul 1 2025
Diagnostic delay remains a critical challenge in axial spondyloarthritis (axSpA). This review highlights key clinical and imaging research from 2024 that addresses this persistent issue, with a focus on the evolving roles of MRI, artificial intelligence (AI), and updated Canadian management recommendations. Multiple studies published in 2024 emphasized the continued problem of diagnostic delay in axSpA. Studies support the continued use of sacroiliac joint MRI as a central diagnostic tool for axSpA, particularly in patients with chronic back pain and associated conditions like uveitis, psoriasis (PsO), or inflammatory bowel disease. AI-based tools for interpreting sacroiliac joint MRIs demonstrated moderate agreement with expert assessments, offering a potential solution to variability and limited access to expert musculoskeletal radiology. These innovations may support earlier diagnosis and reduce misclassification. Innovative models of care, including patient-initiated telemedicine visits, reduced in-person visit frequency without compromising clinical outcomes in patients with stable axSpA. Updated Canadian treatment guidelines introduced more robust data on Janus kinase (JAK) inhibitors and offered stronger support for tapering biologics in patients with sustained low disease activity or remission, while advising against abrupt discontinuation. This clinical and imaging year in review covers challenges and innovations in axSpA, emphasizing the need for early access to care and the development of tools to support prompt diagnosis and sustained continuity of care.

How I Do It: Three-Dimensional MR Neurography and Zero Echo Time MRI for Rendering of Peripheral Nerve and Bone.

Lin Y, Tan ET, Campbell G, Breighner RE, Fung M, Wolfe SW, Carrino JA, Sneag DB

pubmed logopapersJul 1 2025
MR neurography sequences provide excellent nerve-to-background soft tissue contrast, whereas a zero echo time (ZTE) MRI sequence provides cortical bone contrast. By demonstrating the spatial relationship between nerves and bones, a combination of rendered three-dimensional (3D) MR neurography and ZTE sequences provides a roadmap for clinical decision-making, particularly for surgical intervention. In this article, the authors describe the method for fused rendering of peripheral nerve and bone by combining nerve and bone structures from 3D MR neurography and 3D ZTE MRI, respectively. The described method includes scanning acquisition, postprocessing that entails deep learning-based reconstruction techniques, and rendering techniques. Representative case examples demonstrate the steps and clinical use of these techniques. Challenges in nerve and bone rendering are also discussed.

An efficient attention Densenet with LSTM for lung disease detection and classification using X-ray images supported by adaptive R2-Unet-based image segmentation.

Betha SK, Dev DR, Sunkara K, Kodavanti PV, Putta A

pubmed logopapersJul 1 2025
Lung diseases represent one of the most prevalent health challenges globally, necessitating accurate diagnosis to improve patient outcomes. This work presents a novel deep learning-aided lung disease classification framework comprising three key phases: image acquisition, segmentation, and classification. Initially, chest X-ray images are taken from standard datasets. The lung regions are segmented using an Adaptive Recurrent Residual U-Net (AR2-UNet), whose parameters are optimised using Enhanced Pufferfish Optimisation Algorithm (EPOA) to enhance segmentation accuracy. The segmented images are processed using "Attention-based Densenet with Long Short Term Memory(ADNet-LSTM)" for robust categorisation. Investigational results demonstrate that the proposed model achieves the highest classification accuracy of 93.92%, significantly outperforming several baseline models including ResNet with 90.77%, Inception with 89.55%, DenseNet with 89.66%, and "Long Short Term Memory (LSTM)" with 91.79%. Thus, the proposed framework offers a dependable and efficient solution for lung disease detection, supporting clinicians in early and accurate diagnosis.

GAN-based Denoising for Scan Time Reduction and Motion Correction of 18F FP-CIT PET/CT: A Multicenter External Validation Study.

Han H, Choo K, Jeon TJ, Lee S, Seo S, Kim D, Kim SJ, Lee SH, Yun M

pubmed logopapersJul 1 2025
AI-driven scan time reduction is rapidly transforming medical imaging with benefits such as improved patient comfort and enhanced efficiency. A Dual Contrastive Learning Generative Adversarial Network (DCLGAN) was developed to predict full-time PET scans from shorter, noisier scans, improving challenges in imaging patients with movement disorders. 18F FP-CIT PET/CT data from 391 patients with suspected Parkinsonism were used [250 training/validation, 141 testing (hospital A)]. Ground truth (GT) images were reconstructed from 15-minute scans, while denoised images (DIs) were generated from 1-, 3-, 5-, and 10-minute scans. Image quality was assessed using normalized root mean square error (NRMSE), peak signal-to-noise ratio (PSNR), structural similarity index measure (SSIM), visual analysis, and clinical metrics like BPND and ISR for diagnosis of non-neurodegenerative Parkinson disease (NPD), idiopathic PD (IPD), and atypical PD (APD). External validation used data from 2 hospitals with different scanners (hospital B: 1-, 3-, 5-, and 10-min; hospital C: 1-, 3-, and 5-min). In addition, motion artifact reduction was evaluated using the Dice similarity coefficient (DSC). In hospital A, NRMSE, PSNR, and SSIM values improved with scan duration, with the 5-minute DIs achieving optimal quality (NRMSE 0.008, PSNR 42.13, SSIM 0.98). Visual analysis rated DIs from scans ≥3 minutes as adequate or higher. The mean BPND differences (95% CI) for each DIs were 0.19 (-0.01, 0.40), 0.11 (-0.02, 0.24), 0.08 (-0.03, 0.18), and 0.01 (-0.06, 0.07), with the CIs significantly decreasing. ISRs with the highest effect sizes for differentiating NPD, IPD, and APD (PP/AP, PP/VS, PC/VP) remained stable post-denoising. External validation showed 10-minute DIs (hospital B) and 1-minute DIs (hospital C) reached benchmarks of hospital A's image quality metrics, with similar trends in visual analysis and BPND CIs. Furthermore, motion artifact correction in 9 patients yielded DSC improvements from 0.89 to 0.95 in striatal regions. The DL-model is capable of generating high-quality 18F FP-CIT PET images from shorter scans to enhance patient comfort, minimize motion artifacts, and maintain diagnostic precision. Furthermore, our study plays an important role in providing insights into how imaging quality assessment metrics can be used to determine the appropriate scan duration for different scanners with varying sensitivities.

Computed Tomography Advancements in Plaque Analysis: From Histology to Comprehensive Plaque Burden Assessment.

Catapano F, Lisi C, Figliozzi S, Scialò V, Politi LS, Francone M

pubmed logopapersJul 1 2025
Advancements in coronary computed tomography angiography (CCTA) facilitated the transition from traditional histological approaches to comprehensive plaque burden assessment. Recent updates in the European Society of Cardiology (ESC) guidelines emphasize CCTA's role in managing chronic coronary syndrome by enabling detailed monitoring of atherosclerotic plaque progression. Limitations of conventional CCTA, such as spatial resolution challenges in accurately characterizing plaque components like thin-cap fibroatheromas and necrotic lipid-rich cores, are addressed with photon-counting detector CT (PCD-CT) technology. PCD-CT offers enhanced spatial resolution and spectral imaging, improving the detection and characterization of high-risk plaque features while reducing artifacts. The integration of artificial intelligence (AI) in plaque analysis enhances diagnostic accuracy through automated plaque characterization and radiomics. These technological advancements support a comprehensive approach to plaque assessment, incorporating hemodynamic evaluations, morphological metrics, and AI-driven analysis, thereby enabling personalized patient care and improved prediction of acute clinical events.

Patient-specific deep learning tracking for real-time 2D pancreas localisation in kV-guided radiotherapy.

Ahmed AM, Madden L, Stewart M, Chow BVY, Mylonas A, Brown R, Metz G, Shepherd M, Coronel C, Ambrose L, Turk A, Crispin M, Kneebone A, Hruby G, Keall P, Booth JT

pubmed logopapersJul 1 2025
In pancreatic stereotactic body radiotherapy (SBRT), accurate motion management is crucial for the safe delivery of high doses per fraction. Intra-fraction tracking with magnetic resonance imaging-guidance for gated SBRT has shown potential for improved local control. Visualisation of pancreas (and surrounding organs) remains challenging in intra-fraction kilo-voltage (kV) imaging, requiring implanted fiducials. In this study, we investigate patient-specific deep-learning approaches to track the gross-tumour-volume (GTV), pancreas-head and the whole-pancreas in intra-fraction kV images. Conditional-generative-adversarial-networks were trained and tested on data from 25 patients enrolled in an ethics-approved pancreatic SBRT trial for contour prediction on intra-fraction 2D kV images. Labelled digitally-reconstructed-radiographs (DRRs) were generated from contoured planning-computed-tomography (CTs) (CT-DRRs) and cone-beam-CTs (CBCT-DRRs). A population model was trained using CT-DRRs of 19 patients. Two patient-specific model types were created for six additional patients by fine-tuning the population model using CBCT-DRRs (CBCT-models) or CT-DRRs (CT-models) acquired in exhale-breath-hold. Model predictions on unseen triggered-kV images from the corresponding six patients were evaluated against projected-contours using Dice-Similarity-Coefficient (DSC), centroid-error (CE), average Hausdorff-distance (AHD), and Hausdorff-distance at 95th-percentile (HD95). The mean ± 1SD (standard-deviation) DSCs were 0.86 ± 0.09 (CBCT-models) and 0.78 ± 0.12 (CT-models). For AHD and CE, the CBCT-model predicted contours within 2.0 mm ≥90.3 % of the time, while HD95 was within 5.0 mm ≥90.0 % of the time, and had a prediction time of 29.2 ± 3.7 ms per contour. The patient-specific CBCT-models outperformed the CT-models and predicted the three contours with 90th-percentile error ≤2.0 mm, indicating the potential for clinical real-time application.

A Contrast-Enhanced Ultrasound Cine-Based Deep Learning Model for Predicting the Response of Advanced Hepatocellular Carcinoma to Hepatic Arterial Infusion Chemotherapy Combined With Systemic Therapies.

Han X, Peng C, Ruan SM, Li L, He M, Shi M, Huang B, Luo Y, Liu J, Wen H, Wang W, Zhou J, Lu M, Chen X, Zou R, Liu Z

pubmed logopapersJul 1 2025
Recently, a hepatic arterial infusion chemotherapy (HAIC)-associated combination therapeutic regimen, comprising HAIC and systemic therapies (molecular targeted therapy plus immunotherapy), referred to as HAIC combination therapy, has demonstrated promising anticancer effects. Identifying individuals who may potentially benefit from HAIC combination therapy could contribute to improved treatment decision-making for patients with advanced hepatocellular carcinoma (HCC). This dual-center study was a retrospective analysis of prospectively collected data with advanced HCC patients who underwent HAIC combination therapy and pretreatment contrast-enhanced ultrasound (CEUS) evaluations from March 2019 to March 2023. Two deep learning models, AE-3DNet and 3DNet, along with a time-intensity curve-based model, were developed for predicting therapeutic responses from pretreatment CEUS cine images. Diagnostic metrics, including the area under the receiver-operating-characteristic curve (AUC), were calculated to compare the performance of the models. Survival analysis was used to assess the relationship between predicted responses and prognostic outcomes. The model of AE-3DNet was constructed on the top of 3DNet, with innovative incorporation of spatiotemporal attention modules to enhance the capacity for dynamic feature extraction. 326 patients were included, 243 of whom formed the internal validation cohort, which was utilized for model development and fivefold cross-validation, while the rest formed the external validation cohort. Objective response (OR) or non-objective response (non-OR) were observed in 63% (206/326) and 37% (120/326) of the participants, respectively. Among the three efficacy prediction models assessed, AE-3DNet performed superiorly with AUC values of 0.84 and 0.85 in the internal and external validation cohorts, respectively. AE-3DNet's predicted response survival curves closely resembled actual clinical outcomes. The deep learning model of AE-3DNet developed based on pretreatment CEUS cine performed satisfactorily in predicting the responses of advanced HCC to HAIC combination therapy, which may serve as a promising tool for guiding combined therapy and individualized treatment strategies. Trial Registration: NCT02973685.

Deep Learning Models for CT Segmentation of Invasive Pulmonary Aspergillosis, Mucormycosis, Bacterial Pneumonia and Tuberculosis: A Multicentre Study.

Li Y, Huang F, Chen D, Zhang Y, Zhang X, Liang L, Pan J, Tan L, Liu S, Lin J, Li Z, Hu G, Chen H, Peng C, Ye F, Zheng J

pubmed logopapersJul 1 2025
The differential diagnosis of invasive pulmonary aspergillosis (IPA), pulmonary mucormycosis (PM), bacterial pneumonia (BP) and pulmonary tuberculosis (PTB) are challenging due to overlapping clinical and imaging features. Manual CT lesion segmentation is time-consuming, deep-learning (DL)-based segmentation models offer a promising solution, yet disease-specific models for these infections remain underexplored. We aimed to develop and validate dedicated CT segmentation models for IPA, PM, BP and PTB to enhance diagnostic accuracy. Methods:Retrospective multi-centre data (115 IPA, 53 PM, 130 BP, 125 PTB) were used for training/internal validation, with 21 IPA, 8PM, 30 BP and 31 PTB cases for external validation. Expert-annotated lesions served as ground truth. An improved 3D U-Net architecture was employed for segmentation, with preprocessing steps including normalisations, cropping and data augmentation. Performance was evaluated using Dice coefficients. Results:Internal validation achieved Dice scores of 78.83% (IPA), 93.38% (PM), 80.12% (BP) and 90.47% (PTB). External validation showed slightly reduced but robust performance: 75.09% (IPA), 77.53% (PM), 67.40% (BP) and 80.07% (PTB). The PM model demonstrated exceptional generalisability, scoring 83.41% on IPA data. Cross-validation revealed mutual applicability, with IPA/PTB models achieving > 75% Dice for each other's lesions. BP segmentation showed lower but clinically acceptable performance ( >72%), likely due to complex radiological patterns. Disease-specific DL segmentation models exhibited high accuracy, particularly for PM and PTB. While IPA and BP models require refinement, all demonstrated cross-disease utility, suggesting immediate clinical value for preliminary lesion annotation. Future efforts should enhance datasets and optimise models for intricate cases.

Synthetic Versus Classic Data Augmentation: Impacts on Breast Ultrasound Image Classification.

Medghalchi Y, Zakariaei N, Rahmim A, Hacihaliloglu I

pubmed logopapersJul 1 2025
The effectiveness of deep neural networks (DNNs) for the ultrasound image analysis depends on the availability and accuracy of the training data. However, the large-scale data collection and annotation, particularly in medical fields, is often costly and time consuming, especially when healthcare professionals are already burdened with their clinical responsibilities. Ensuring that a model remains robust across different imaging conditions-such as variations in ultrasound devices and manual transducer operation-is crucial in the ultrasound image analysis. The data augmentation is a widely used solution, as it increases both the size and diversity of datasets, thereby enhancing the generalization performance of DNNs. With the advent of generative networks such as generative adversarial networks (GANs) and diffusion-based models, the synthetic data generation has emerged as a promising augmentation technique. However, comprehensive studies comparing classic and generative method-based augmentation methods are lacking, particularly in ultrasound-based breast cancer imaging, where variability in breast density, tumor morphology, and operator skill poses significant challenges. This study aims to compare the effectiveness of classic and generative network-based data augmentation techniques in improving the performance and robustness of breast ultrasound image classification models. Specifically, we seek to determine whether the computational intensity of generative networks is justified in data augmentation. This analysis will provide valuable insights into the role and benefits of each technique in enhancing the diagnostic accuracy of DNN for breast cancer diagnosis. The code for this work will be available at: ht.tps://github.com/yasamin-med/SCDA.git.

A Deep Learning Approach for Nerve Injury Classification in Brachial Plexopathies Using Magnetic Resonance Neurography with Modified Hiking Optimization Algorithm.

Dahou A, Elaziz MA, Khattap MG, Hassan HGEMA

pubmed logopapersJul 1 2025
Brachial plexopathies (BPs) encompass a complex spectrum of nerve injuries affecting motor and sensory function in the upper extremities. Diagnosis is challenging due to the intricate anatomy and symptom overlap with other neuropathies. Magnetic Resonance Neurography (MRN) provides advanced imaging but requires specialized interpretation. This study proposes an AI-based framework that combines deep learning (DL) with the modified Hiking Optimization Algorithm (MHOA) enhanced by a Comprehensive Learning (CL) technique to improve the classification of nerve injuries (neuropraxia, axonotmesis, neurotmesis) using MRN data. The framework utilizes MobileNetV4 for feature extraction and MHOA for optimized feature selection across different MRI sequences (STIR, T2, T1, and DWI). A dataset of 39 patients diagnosed with BP was used. The framework classifies injuries based on Seddon's criteria, distinguishing between normal and abnormal conditions as well as injury severity. The model achieved excellent performance, with 1.0000 accuracy in distinguishing normal from abnormal conditions using STIR and T2 sequences. For injury severity classification, accuracy was 0.9820 in STIR, outperforming the original HOA and other metaheuristic algorithms. Additionally, high classification accuracy (0.9667) was observed in DWI. The proposed framework outperformed traditional methods and demonstrated high sensitivity and specificity. The proposed AI-based framework significantly improves the diagnosis of BP by accurately classifying nerve injury types. By integrating DL and optimization techniques, it reduces diagnostic variability, making it a valuable tool for clinical settings with limited specialized neuroimaging expertise. This framework has the potential to enhance clinical decision-making and optimize patient outcomes through precise and timely diagnoses.
Page 207 of 3933923 results
Show
per page

Ready to Sharpen Your Edge?

Join hundreds of your peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.