Sort by:
Page 72 of 91907 results

Structural and metabolic topological alterations associated with butylphthalide treatment in mild cognitive impairment: Data from a randomized, double-blind, placebo-controlled trial.

Han X, Gong S, Gong J, Wang P, Li R, Chen R, Xu C, Sun W, Li S, Chen Y, Yang Y, Luan H, Wen B, Guo J, Lv S, Wei C

pubmed logopapersJun 1 2025
Effective intervention for mild cognitive impairment (MCI) is key for preventing dementia. As a neuroprotective agent, butylphthalide has the potential to treat MCI due to Alzheimer disease (AD). However, the pharmacological mechanism of butylphthalide from the brain network perspective is not clear. Therefore, we aimed to investigate the multimodal brain network changes associated with butylphthalide treatment in MCI due to AD. A total of 270 patients with MCI due to AD received either butylphthalide or placebo at a ratio of 1:1 for 1 year. Effective treatment was defined as a decrease in the Alzheimer's Disease Assessment Scale-Cognitive Subscale (ADAS-cog) > 2.5. Brain networks were constructed using T1-magnetic resonance imaging and fluorodeoxyglucose positron emission tomography. A support vector machine was applied to develop predictive models. Both treatment (drug vs. placebo)-time interactions and efficacy (effective vs. ineffective)-time interactions were detected on some overlapping structural network metrics. Simple effects analyses revealed a significantly increased global efficiency in the structural network under both treatment and effective treatment of butylphthalide. Among the overlapping metrics, an increased degree centrality of left paracentral lobule was significantly related to poorer cognitive improvement. The predictive model based on baseline multimodal network metrics exhibited high accuracy (88.93%) of predicting butylphthalide's efficacy. Butylphthalide may restore abnormal organization in structural networks of patients with MCI due to AD, and baseline network metrics could be predictive markers for therapeutic efficacy of butylphthalide. This study was registered in the Chinese Clinical Trial Registry (Registration Number: ChiCTR1800018362, Registration Date: 2018-09-13).

Integration of Deep Learning and Sub-regional Radiomics Improves the Prediction of Pathological Complete Response to Neoadjuvant Chemoradiotherapy in Locally Advanced Rectal Cancer Patients.

Wu X, Wang J, Chen C, Cai W, Guo Y, Guo K, Chen Y, Shi Y, Chen J, Lin X, Jiang X

pubmed logopapersJun 1 2025
The precise prediction of response to neoadjuvant chemoradiotherapy is crucial for tailoring perioperative treatment in patients diagnosed with locally advanced rectal cancer (LARC). This retrospective study aims to develop and validate a model that integrates deep learning and sub-regional radiomics from MRI imaging to predict pathological complete response (pCR) in patients with LARC. We retrospectively enrolled 768 eligible participants from three independent hospitals who had received neoadjuvant chemoradiotherapy followed by radical surgery. Pretreatment pelvic MRI scans (T2-weighted), were collected for annotation and feature extraction. The K-means approach was used to segment the tumor into sub-regions. Radiomics and deep learning features were extracted by the Pyradiomics and 3D ResNet50, respectively. The predictive models were developed using the radiomics, sub-regional radiomics, and deep learning features with the machine learning algorithm in training cohort, and then validated in the external tests. The models' performance was assessed using various metrics, including the area under the curve (AUC), decision curve analysis, Kaplan-Meier survival analysis. We constructed a combined model, named SRADL, which includes deep learning with sub-regional radiomics signatures, enabling precise prediction of pCR in LARC patients. SRADL had satisfactory performance for the prediction of pCR in the training cohort (AUC 0.925 [95% CI 0.894 to 0.948]), and in test 1 (AUC 0.915 [95% CI 0.869 to 0.949]) and in test 2 (AUC 0.902 [95% CI 0.846 to 0.945]). By employing optimal threshold of 0.486, the predicted pCR group had longer survival compared to predicted non-pCR group across three cohorts. SRADL also outperformed other single-modality prediction models. The novel SRADL, which integrates deep learning with sub-regional signatures, showed high accuracy and robustness in predicting pCR to neoadjuvant chemoradiotherapy using pretreatment MRI images, making it a promising tool for the personalized management of LARC.

Evaluation of a Deep Learning Denoising Algorithm for Dose Reduction in Whole-Body Photon-Counting CT Imaging: A Cadaveric Study.

Dehdab R, Brendel JM, Streich S, Ladurner R, Stenzl B, Mueck J, Gassenmaier S, Krumm P, Werner S, Herrmann J, Nikolaou K, Afat S, Brendlin A

pubmed logopapersJun 1 2025
Photon Counting CT (PCCT) offers advanced imaging capabilities with potential for substantial radiation dose reduction; however, achieving this without compromising image quality remains a challenge due to increased noise at lower doses. This study aims to evaluate the effectiveness of a deep learning (DL)-based denoising algorithm in maintaining diagnostic image quality in whole-body PCCT imaging at reduced radiation levels, using real intraindividual cadaveric scans. Twenty-four cadaveric human bodies underwent whole-body CT scans on a PCCT scanner (NAEOTOM Alpha, Siemens Healthineers) at four different dose levels (100%, 50%, 25%, and 10% mAs). Each scan was reconstructed using both QIR level 2 and a DL algorithm (ClariCT.AI, ClariPi Inc.), resulting in 192 datasets. Objective image quality was assessed by measuring CT value stability, image noise, and contrast-to-noise ratio (CNR) across consistent regions of interest (ROIs) in the liver parenchyma. Two radiologists independently evaluated subjective image quality based on overall image clarity, sharpness, and contrast. Inter-rater agreement was determined using Spearman's correlation coefficient, and statistical analysis included mixed-effects modeling to assess objective and subjective image quality. Objective analysis showed that the DL denoising algorithm did not significantly alter CT values (p ≥ 0.9975). Noise levels were consistently lower in denoised datasets compared to the Original (p < 0.0001). No significant differences were observed between the 25% mAs denoised and the 100% mAs original datasets in terms of noise and CNR (p ≥ 0.7870). Subjective analysis revealed strong inter-rater agreement (r ≥ 0.78), with the 50% mAs denoised datasets rated superior to the 100% mAs original datasets (p < 0.0001) and no significant differences detected between the 25% mAs denoised and 100% mAs original datasets (p ≥ 0.9436). The DL denoising algorithm maintains image quality in PCCT imaging while enabling up to a 75% reduction in radiation dose. This approach offers a promising method for reducing radiation exposure in clinical PCCT without compromising diagnostic quality.

Deep Learning-Enhanced Ultra-high-resolution CT Imaging for Superior Temporal Bone Visualization.

Brockstedt L, Grauhan NF, Kronfeld A, Mercado MAA, Döge J, Sanner A, Brockmann MA, Othman AE

pubmed logopapersJun 1 2025
This study assesses the image quality of temporal bone ultra-high-resolution (UHR) Computed tomography (CT) scans in adults and children using hybrid iterative reconstruction (HIR) and a novel, vendor-specific deep learning-based reconstruction (DLR) algorithm called AiCE Inner Ear. In a retrospective, single-center study (February 1-July 30, 2023), UHR-CT scans of 57 temporal bones of 35 patients (5 children, 23 male) with at least one anatomical unremarkable temporal bone were included. There is an adult computed tomography dose index volume (CTDIvol 25.6 mGy) and a pediatric protocol (15.3 mGy). Images were reconstructed using HIR at normal resolution (0.5-mm slice thickness, 512² matrix) and UHR (0.25-mm, 1024² and 2048² matrix) as well as with a vendor-specific DLR advanced intelligent clear-IQ engine inner ear (AiCE Inner Ear) at UHR (0.25-mm, 1024² matrix). Three radiologists evaluated 18 anatomic structures using a 5-point Likert scale. Signal-to-noise (SNR) and contrast-to-noise ratio (CNR) were measured automatically. In the adult protocol subgroup (n=30; median age: 51 [11-89]; 19 men) and the pediatric protocol subgroup (n=5; median age: 2 [1-3]; 4 men), UHR-CT with DLR significantly improved subjective image quality (p<0.024), reduced noise (p<0.001), and increased CNR and SNR (p<0.001). DLR also enhanced visualization of key structures, including the tendon of the stapedius muscle (p<0.001), tympanic membrane (p<0.009), and basal aspect of the osseous spiral lamina (p<0.018). Vendor-specific DLR-enhanced UHR-CT significantly improves temporal bone image quality and diagnostic performance.

A systematic review on deep learning-enabled coronary CT angiography for plaque and stenosis quantification and cardiac risk prediction.

Shrivastava P, Kashikar S, Parihar PH, Kasat P, Bhangale P, Shrivastava P

pubmed logopapersJun 1 2025
Coronary artery disease (CAD) is a major worldwide health concern, contributing significantly to the global burden of cardiovascular diseases (CVDs). According to the 2023 World Health Organization (WHO) report, CVDs account for approximately 17.9 million deaths annually. This emphasizies the need for advanced diagnostic tools such as coronary computed tomography angiography (CCTA). The incorporation of deep learning (DL) technologies could significantly improve CCTA analysis by automating the quantification of plaque and stenosis, thus enhancing the precision of cardiac risk assessments. A recent meta-analysis highlights the evolving role of CCTA in patient management, showing that CCTA-guided diagnosis and management reduced adverse cardiac events and improved event-free survival in patients with stable and acute coronary syndromes. An extensive literature search was carried out across various electronic databases, such as MEDLINE, Embase, and the Cochrane Library. This search utilized a specific strategy that included both Medical Subject Headings (MeSH) terms and pertinent keywords. The review adhered to PRISMA guidelines and focused on studies published between 2019 and 2024 that employed deep learning (DL) for coronary computed tomography angiography (CCTA) in patients aged 18 years or older. After implementing specific inclusion and exclusion criteria, a total of 10 articles were selected for systematic evaluation regarding quality and bias. This systematic review included a total of 10 studies, demonstrating the high diagnostic performance and predictive capabilities of various deep learning models compared to different imaging modalities. This analysis highlights the effectiveness of these models in enhancing diagnostic accuracy in imaging techniques. Notably, strong correlations were observed between DL-derived measurements and intravascular ultrasound findings, enhancing clinical decision-making and risk stratification for CAD. Deep learning-enabled CCTA represents a promising advancement in the quantification of coronary plaques and stenosis, facilitating improved cardiac risk prediction and enhancing clinical workflow efficiency. Despite variability in study designs and potential biases, the findings support the integration of DL technologies into routine clinical practice for better patient outcomes in CAD management.

Deep learning-based acceleration of high-resolution compressed sense MR imaging of the hip.

Marka AW, Meurer F, Twardy V, Graf M, Ebrahimi Ardjomand S, Weiss K, Makowski MR, Gersing AS, Karampinos DC, Neumann J, Woertler K, Banke IJ, Foreman SC

pubmed logopapersJun 1 2025
To evaluate a Compressed Sense Artificial Intelligence framework (CSAI) incorporating parallel imaging, compressed sense (CS), and deep learning for high-resolution MRI of the hip, comparing it with standard-resolution CS imaging. Thirty-two patients with femoroacetabular impingement syndrome underwent 3 T MRI scans. Coronal and sagittal intermediate-weighted TSE sequences with fat saturation were acquired using CS (0.6 ×0.8 mm resolution) and CSAI (0.3 ×0.4 mm resolution) protocols in comparable acquisition times (7:49 vs. 8:07 minutes for both planes). Two readers systematically assessed the depiction of the acetabular and femoral cartilage (in five cartilage zones), labrum, ligamentum capitis femoris, and bone using a five-point Likert scale. Diagnostic confidence and abnormality detection were recorded and analyzed using the Wilcoxon signed-rank test. CSAI significantly improved the cartilage depiction across most cartilage zones compared to CS. Overall Likert scores were 4.0 ± 0.2 (CS) vs 4.2 ± 0.6 (CSAI) for reader 1 and 4.0 ± 0.2 (CS) vs 4.3 ± 0.6 (CSAI) for reader 2 (p ≤ 0.001). Diagnostic confidence increased from 3.5 ± 0.7 and 3.9 ± 0.6 (CS) to 4.0 ± 0.6 and 4.1 ± 0.7 (CSAI) for readers 1 and 2, respectively (p ≤ 0.001). More cartilage lesions were detected with CSAI, with significant improvements in diagnostic confidence in certain cartilage zones such as femoral zone C and D for both readers. Labrum and ligamentum capitis femoris depiction remained similar, while bone depiction was rated lower. No abnormalities detected in CS were missed in CSAI. CSAI provides high-resolution hip MR images with enhanced cartilage depiction without extending acquisition times, potentially enabling more precise hip cartilage assessment.

Combating Medical Label Noise through more precise partition-correction and progressive hard-enhanced learning.

Zhang S, Chu S, Qiang Y, Zhao J, Wang Y, Wei X

pubmed logopapersJun 1 2025
Computer-aided diagnosis systems based on deep neural networks heavily rely on datasets with high-quality labels. However, manual annotation for lesion diagnosis relies on image features, often requiring professional experience and complex image analysis process. This inevitably introduces noisy labels, which can misguide the training of classification models. Our goal is to design an effective method to address the challenges posed by label noise in medical images. we propose a novel noise-tolerant medical image classification framework consisting of two phases: fore-training correction and progressive hard-sample enhanced learning. In the first phase, we design a dual-branch sample partition detection scheme that effectively classifies each instance into one of three subsets: clean, hard, or noisy. Simultaneously, we propose a hard-sample label refinement strategy based on class prototypes with confidence-perception weighting and an effective joint correction method for noisy samples, enabling the acquisition of higher-quality training data. In the second phase, we design a progressive hard-sample reinforcement learning method to enhance the model's ability to learn discriminative feature representations. This approach accounts for sample difficulty and mitigates the effects of label noise in medical datasets. Our framework achieves an accuracy of 82.39% on the pneumoconiosis dataset collected by our laboratory. On a five-class skin disease dataset with six different levels of label noise (0, 0.05, 0.1, 0.2, 0.3, and 0.4), the average accuracy over the last ten epochs reaches 88.51%, 86.64%, 85.02%, 83.01%, 81.95%, 77.89%, respectively; For binary polyp classification under noise rates of 0.2, 0.3, and 0.4, the average accuracy over the last ten epochs is 97.90%, 93.77%, 89.33%, respectively. The effectiveness of our proposed framework is demonstrated through its performance on three challenging datasets with both real and synthetic noise. Experimental results further demonstrate the robustness of our method across varying noise rates.

Explicit Abnormality Extraction for Unsupervised Motion Artifact Reduction in Magnetic Resonance Imaging.

Zhou Y, Li H, Liu J, Kong Z, Huang T, Ahn E, Lv Z, Kim J, Feng DD

pubmed logopapersJun 1 2025
Motion artifacts compromise the quality of magnetic resonance imaging (MRI) and pose challenges to achieving diagnostic outcomes and image-guided therapies. In recent years, supervised deep learning approaches have emerged as successful solutions for motion artifact reduction (MAR). One disadvantage of these methods is their dependency on acquiring paired sets of motion artifact-corrupted (MA-corrupted) and motion artifact-free (MA-free) MR images for training purposes. Obtaining such image pairs is difficult and therefore limits the application of supervised training. In this paper, we propose a novel UNsupervised Abnormality Extraction Network (UNAEN) to alleviate this problem. Our network is capable of working with unpaired MA-corrupted and MA-free images. It converts the MA-corrupted images to MA-reduced images by extracting abnormalities from the MA-corrupted images using a proposed artifact extractor, which intercepts the residual artifact maps from the MA-corrupted MR images explicitly, and a reconstructor to restore the original input from the MA-reduced images. The performance of UNAEN was assessed by experimenting with various publicly available MRI datasets and comparing them with state-of-the-art methods. The quantitative evaluation demonstrates the superiority of UNAEN over alternative MAR methods and visually exhibits fewer residual artifacts. Our results substantiate the potential of UNAEN as a promising solution applicable in real-world clinical environments, with the capability to enhance diagnostic accuracy and facilitate image-guided therapies.

P2TC: A Lightweight Pyramid Pooling Transformer-CNN Network for Accurate 3D Whole Heart Segmentation.

Cui H, Wang Y, Zheng F, Li Y, Zhang Y, Xia Y

pubmed logopapersJun 1 2025
Cardiovascular disease is a leading global cause of death, requiring accurate heart segmentation for diagnosis and surgical planning. Deep learning methods have been demonstrated to achieve superior performances in cardiac structures segmentation. However, there are still limitations in 3D whole heart segmentation, such as inadequate spatial context modeling, difficulty in capturing long-distance dependencies, high computational complexity, and limited representation of local high-level semantic information. To tackle the above problems, we propose a lightweight Pyramid Pooling Transformer-CNN (P2TC) network for accurate 3D whole heart segmentation. The proposed architecture comprises a dual encoder-decoder structure with a 3D pyramid pooling Transformer for multi-scale information fusion and a lightweight large-kernel Convolutional Neural Network (CNN) for local feature extraction. The decoder has two branches for precise segmentation and contextual residual handling. The first branch is used to generate segmentation masks for pixel-level classification based on the features extracted by the encoder to achieve accurate segmentation of cardiac structures. The second branch highlights contextual residuals across slices, enabling the network to better handle variations and boundaries. Extensive experimental results on the Multi-Modality Whole Heart Segmentation (MM-WHS) 2017 challenge dataset demonstrate that P2TC outperforms the most advanced methods, achieving the Dice scores of 92.6% and 88.1% in Computed Tomography (CT) and Magnetic Resonance Imaging (MRI) modalities respectively, which surpasses the baseline model by 1.5% and 1.7%, and achieves state-of-the-art segmentation results.

MedKAFormer: When Kolmogorov-Arnold Theorem Meets Vision Transformer for Medical Image Representation.

Wang G, Zhu Q, Song C, Wei B, Li S

pubmed logopapersJun 1 2025
Vision Transformers (ViTs) suffer from high parameter complexity because they rely on Multi-layer Perceptrons (MLPs) for nonlinear representation. This issue is particularly challenging in medical image analysis, where labeled data is limited, leading to inadequate feature representation. Existing methods have attempted to optimize either the patch embedding stage or the non-embedding stage of ViTs. Still, they have struggled to balance effective modeling, parameter complexity, and data availability. Recently, the Kolmogorov-Arnold Network (KAN) was introduced as an alternative to MLPs, offering a potential solution to the large parameter issue in ViTs. However, KAN cannot be directly integrated into ViT due to challenges such as handling 2D structured data and dimensionality catastrophe. To solve this problem, we propose MedKAFormer, the first ViT model to incorporate the Kolmogorov-Arnold (KA) theorem for medical image representation. It includes a Dynamic Kolmogorov-Arnold Convolution (DKAC) layer for flexible nonlinear modeling in the patch embedding stage. Additionally, it introduces a Nonlinear Sparse Token Mixer (NSTM) and a Nonlinear Dynamic Filter (NDF) in the non-embedding stage. These components provide comprehensive nonlinear representation while reducing model overfitting. MedKAFormer reduces parameter complexity by 85.61% compared to ViT-Base and achieves competitive results on 14 medical datasets across various imaging modalities and structures.
Page 72 of 91907 results
Show
per page

Ready to Sharpen Your Edge?

Join hundreds of your peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.