Sort by:
Page 157 of 3563559 results

MLRU++: Multiscale Lightweight Residual UNETR++ with Attention for Efficient 3D Medical Image Segmentation

Nand Kumar Yadav, Rodrigue Rizk, William CW Chen, KC Santosh

arxiv logopreprintJul 22 2025
Accurate and efficient medical image segmentation is crucial but challenging due to anatomical variability and high computational demands on volumetric data. Recent hybrid CNN-Transformer architectures achieve state-of-the-art results but add significant complexity. In this paper, we propose MLRU++, a Multiscale Lightweight Residual UNETR++ architecture designed to balance segmentation accuracy and computational efficiency. It introduces two key innovations: a Lightweight Channel and Bottleneck Attention Module (LCBAM) that enhances contextual feature encoding with minimal overhead, and a Multiscale Bottleneck Block (M2B) in the decoder that captures fine-grained details via multi-resolution feature aggregation. Experiments on four publicly available benchmark datasets (Synapse, BTCV, ACDC, and Decathlon Lung) demonstrate that MLRU++ achieves state-of-the-art performance, with average Dice scores of 87.57% (Synapse), 93.00% (ACDC), and 81.12% (Lung). Compared to existing leading models, MLRU++ improves Dice scores by 5.38% and 2.12% on Synapse and ACDC, respectively, while significantly reducing parameter count and computational cost. Ablation studies evaluating LCBAM and M2B further confirm the effectiveness of the proposed architectural components. Results suggest that MLRU++ offers a practical and high-performing solution for 3D medical image segmentation tasks. Source code is available at: https://github.com/1027865/MLRUPP

CLIF-Net: Intersection-guided Cross-view Fusion Network for Infection Detection from Cranial Ultrasound

Yu, M., Peterson, M. R., Burgoine, K., Harbaugh, T., Olupot-Olupot, P., Gladstone, M., Hagmann, C., Cowan, F. M., Weeks, A., Morton, S. U., Mulondo, R., Mbabazi-Kabachelor, E., Schiff, S. J., Monga, V.

medrxiv logopreprintJul 22 2025
This paper addresses the problem of detecting possible serious bacterial infection (pSBI) of infancy, i.e. a clinical presentation consistent with bacterial sepsis in newborn infants using cranial ultrasound (cUS) images. The captured image set for each patient enables multiview imagery: coronal and sagittal, with geometric overlap. To exploit this geometric relation, we develop a new learning framework, called the intersection-guided Crossview Local-and Image-level Fusion Network (CLIF-Net). Our technique employs two distinct convolutional neural network branches to extract features from coronal and sagittal images with newly developed multi-level fusion blocks. Specifically, we leverage the spatial position of these images to locate the intersecting region. We then identify and enhance the semantic features from this region across multiple levels using cross-attention modules, facilitating the acquisition of mutually beneficial and more representative features from both views. The final enhanced features from the two views are then integrated and projected through the image-level fusion layer, outputting pSBI and non-pSBI class probabilities. We contend that our method of exploiting multi-view cUS images enables a first of its kind, robust 3D representation tailored for pSBI detection. When evaluated on a dataset of 302 cUS scans from Mbale Regional Referral Hospital in Uganda, CLIF-Net demonstrates substantially enhanced performance, surpassing the prevailing state-of-the-art infection detection techniques.

LA-Seg: Disentangled sinogram pattern-guided transformer for lesion segmentation in limited-angle computed tomography.

Yoon JH, Lee YJ, Yoo SB

pubmed logopapersJul 21 2025
Limited-angle computed tomography (LACT) offers patient-friendly benefits, such as rapid scanning and reduced radiation exposure. However, the incompleteness of data in LACT often causes notable artifacts, posing challenges for precise medical interpretation. Although numerous approaches have been introduced to reconstruct LACT images into complete computed tomography (CT) scans, they focus on improving image quality and operate separately from lesion segmentation models, often overlooking essential lesion-specific information. This is because reconstruction models are primarily optimized to satisfy overall image quality rather than local lesion-specific regions, in a non-end-to-end setup where each component is optimized independently and may not contribute to reaching the global minimum of the overall objective function. To address this problem, we propose LA-Seg, a transformer-based segmentation model using the sinogram domain of LACT data. The LA-Seg method uses an auxiliary reconstruction task to estimates incomplete sinogram regions to enhance segmentation robustness. Applying transformers adapted from video prediction models captures the spatial structure and sequential patterns in sinograms and reconstructs features in incomplete regions using a disentangled representation guided by distinctive patterns. We propose contrastive abnormal feature loss to distinguish between normal and abnormal regions better. The experimental results demonstrate that LA-Seg consistently surpasses existing medical segmentation approaches in diverse LACT conditions. The source code is provided at https://github.com/jhyoon964/LA-Seg.

Artificial intelligence in radiology: diagnostic sensitivity of ChatGPT for detecting hemorrhages in cranial computed tomography scans.

Bayar-Kapıcı O, Altunışık E, Musabeyoğlu F, Dev Ş, Kaya Ö

pubmed logopapersJul 21 2025
Chat Generative Pre-trained Transformer (ChatGPT)-4V, a large language model developed by OpenAI, has been explored for its potential application in radiology. This study assesses ChatGPT-4V's diagnostic performance in identifying various types of intracranial hemorrhages in non-contrast cranial computed tomography (CT) images. Intracranial hemorrhages were presented to ChatGPT using the clearest 2D imaging slices. The first question, "Q1: Which imaging technique is used in this image?" was asked to determine the imaging modality. ChatGPT was then prompted with the second question, "Q2: What do you see in this image and what is the final diagnosis?" to assess whether the CT scan was normal or showed pathology. For CT scans containing hemorrhage that ChatGPT did not interpret correctly, a follow-up question-"Q3: There is bleeding in this image. Which type of bleeding do you see?"-was used to evaluate whether this guidance influenced its response. ChatGPT accurately identified the imaging technique (Q1) in all cases but demonstrated difficulty diagnosing epidural hematoma (EDH), subdural hematoma (SDH), and subarachnoid hemorrhage (SAH) when no clues were provided (Q2). When a hemorrhage clue was introduced (Q3), ChatGPT correctly identified EDH in 16.7% of cases, SDH in 60%, and SAH in 15.6%, and achieved 100% diagnostic accuracy for hemorrhagic cerebrovascular disease. Its sensitivity, specificity, and accuracy for Q2 were 23.6%, 92.5%, and 57.4%, respectively. These values improved substantially with the clue in Q3, with sensitivity rising to 50.9% and accuracy to 71.3%. ChatGPT also demonstrated higher diagnostic accuracy in larger hemorrhages in EDH and SDH images. Although the model performs well in recognizing imaging modalities, its diagnostic accuracy substantially improves when guided by additional contextual information. These findings suggest that ChatGPT's diagnostic performance improves with guided prompts, highlighting its potential as a supportive tool in clinical radiology.

Facilitators and Barriers to Implementing AI in Routine Medical Imaging: Systematic Review and Qualitative Analysis.

Wenderott K, Krups J, Weigl M, Wooldridge AR

pubmed logopapersJul 21 2025
Artificial intelligence (AI) is rapidly advancing in health care, particularly in medical imaging, offering potential for improved efficiency and reduced workload. However, there is little systematic evidence on process factors for successful AI technology implementation into clinical workflows. This study aimed to systematically assess and synthesize the facilitators and barriers to AI implementation reported in studies evaluating AI solutions in routine medical imaging. We conducted a systematic review of 6 medical databases. Using a qualitative content analysis, we extracted the reported facilitators and barriers, outcomes, and moderators in the implementation process of AI. Two reviewers analyzed and categorized the data separately. We then used epistemic network analysis to explore their relationships across different stages of AI implementation. Our search yielded 13,756 records. After screening, we included 38 original studies in our final review. We identified 12 key dimensions and 37 subthemes that influence the implementation of AI in health care workflows. Key dimensions included evaluation of AI use and fit into workflow, with frequency depending considerably on the stage of the implementation process. In total, 20 themes were mentioned as both facilitators and barriers to AI implementation. Studies often focused predominantly on performance metrics over the experiences or outcomes of clinicians. This systematic review provides a thorough synthesis of facilitators and barriers to successful AI implementation in medical imaging. Our study highlights the usefulness of AI technologies in clinical care and the fit of their integration into routine clinical workflows. Most studies did not directly report facilitators and barriers to AI implementation, underscoring the importance of comprehensive reporting to foster knowledge sharing. Our findings reveal a predominant focus on technological aspects of AI adoption in clinical work, highlighting the need for holistic, human-centric consideration to fully leverage the potential of AI in health care. PROSPERO CRD42022303439; https://www.crd.york.ac.uk/PROSPERO/view/CRD42022303439. RR2-10.2196/40485.

Lightweight Network Enhancing High-Resolution Feature Representation for Efficient Low Dose CT Denoising.

Li J, Li Y, Qi F, Wang S, Zhang Z, Huang Z, Yu Z

pubmed logopapersJul 21 2025
Low-dose computed tomography plays a crucial role in reducing radiation exposure in clinical imaging, however, the resultant noise significantly impacts image quality and diagnostic precision. Recent transformer-based models have demonstrated strong denoising capabilities but are often constrained by high computational complexity. To overcome these limitations, we propose AMFA-Net, an adaptive multi-order feature aggregation network that provides a lightweight architecture for enhancing highresolution feature representation in low-dose CT imaging. AMFA-Net effectively integrates local and global contexts within high-resolution feature maps while learning discriminative representations through multi-order context aggregation. We introduce an agent-based self-attention crossshaped window transformer block that efficiently captures global context in high-resolution feature maps, which is subsequently fused with backbone features to preserve critical structural information. Our approach employs multiorder gated aggregation to adaptively guide the network in capturing expressive interactions that may be overlooked in fused features, thereby producing robust representations for denoised image reconstruction. Experiments on two challenging public datasets with 25% and 10% full-dose CT image quality demonstrate that our method surpasses state-of-the-art approaches in denoising performance with low computational cost, highlighting its potential for realtime medical applications.

Fully automated pedicle screw manufacturer identification in plain radiograph with deep learning methods.

Waranusast R, Riyamongkol P, Weerakul S, Chaibhuddanugul N, Laoruengthana A, Mahatthanatrakul A

pubmed logopapersJul 21 2025
Pedicle screw manufacturer identification is crucial for revision surgery planning; however, this information is occasionally unavailable. We developed a deep learning-based algorithm to identify the pedicle screw manufacturer from plain radiographs. We collected anteroposterior (AP) and lateral radiographs from 276 patients who had thoracolumbar spine surgery with pedicle screws from three international manufacturers. The samples were randomly assigned to training sets (178), validation sets (40), and test sets (58). The algorithm incorporated a convolutional neural network (CNN) model to classify the radiograph as AP and lateral, followed by YOLO object detection to locate the pedicle screw. Another CNN classifier model then identified the manufacturer of each pedicle screw in AP and lateral views. The voting scheme determined the final classification. For comparison, two spine surgeons independently evaluated the same test set, and the accuracy was compared. The mean age of the patients was 59.5 years, with 1,887 pedicle screws included. The algorithm achieved a perfect accuracy of 100% for the AP radiograph, 98.9% for the lateral radiograph, and 100% when both views were considered. By comparison, the spine surgeons achieved 97.1% accuracy. Statistical analysis revealed near-perfect agreement between the algorithm and the surgeons. We have successfully developed an algorithm for pedicle screw manufacturer identification, which demonstrated excellent accuracy and was comparable to experienced spine surgeons.

LLM-driven Medical Report Generation via Communication-efficient Heterogeneous Federated Learning.

Che H, Jin H, Gu Z, Lin Y, Jin C, Chen H

pubmed logopapersJul 21 2025
Large Language Models (LLMs) have demonstrated significant potential in Medical Report Generation (MRG), yet their development requires large amounts of medical image-report pairs, which are commonly scattered across multiple centers. Centralizing these data is exceptionally challenging due to privacy regulations, thereby impeding model development and broader adoption of LLM-driven MRG models. To address this challenge, we present FedMRG, the first framework that leverages Federated Learning (FL) to enable privacy-preserving, multi-center development of LLM-driven MRG models, specifically designed to overcome the critical challenge of communication-efficient LLM training under multi-modal data heterogeneity. To start with, our framework tackles the fundamental challenge of communication overhead in federated LLM tuning by employing low-rank factorization to efficiently decompose parameter updates, significantly reducing gradient transmission costs and making LLM-driven MRG feasible in bandwidth-constrained FL settings. Furthermore, we observed the dual heterogeneity in MRG under the FL scenario: varying image characteristics across medical centers, as well as diverse reporting styles and terminology preferences. To address the data heterogeneity, we further enhance FedMRG with (1) client-aware contrastive learning in the MRG encoder, coupled with diagnosis-driven prompts, which capture both globally generalizable and locally distinctive features while maintaining diagnostic accuracy; and (2) a dual-adapter mutual boosting mechanism in the MRG decoder that harmonizes generic and specialized adapters to address variations in reporting styles and terminology. Through extensive evaluation of our established FL-MRG benchmark, we demonstrate the generalizability and adaptability of FedMRG, underscoring its potential in harnessing multi-center data and generating clinically accurate reports while maintaining communication efficiency.

Noninvasive Deep Learning System for Preoperative Diagnosis of Follicular-Like Thyroid Neoplasms Using Ultrasound Images: A Multicenter, Retrospective Study.

Shen H, Huang Y, Yan W, Zhang C, Liang T, Yang D, Feng X, Liu S, Wang Y, Cao W, Cheng Y, Chen H, Ni Q, Wang F, You J, Jin Z, He W, Sun J, Yang D, Liu L, Cao B, Zhang X, Li Y, Pei S, Zhang S, Zhang B

pubmed logopapersJul 21 2025
To propose a deep learning (DL) system for the preoperative diagnosis of follicular-like thyroid neoplasms (FNs) using routine ultrasound images. Preoperative diagnosis of malignancy in nodules suspicious for an FN remains challenging. Ultrasound, fine-needle aspiration cytology, and intraoperative frozen section pathology cannot unambiguously distinguish between benign and malignant FNs, leading to unnecessary biopsies and operations in benign nodules. This multicenter, retrospective study included 3634 patients who underwent ultrasound and received a definite diagnosis of FN from 11 centers, comprising thyroid follicular adenoma (n=1748), follicular carcinoma (n=299), and follicular variant of papillary thyroid carcinoma (n=1587). Four DL models including Inception-v3, ResNet50, Inception-ResNet-v2, and DenseNet161 were constructed on a training set (n=2587, 6178 images) and were verified on an internal validation set (n=648, 1633 images) and an external validation set (n=399, 847 images). The diagnostic efficacy of the DL models was evaluated against the ACR TI-RADS regarding the area under the curve (AUC), sensitivity, specificity, and unnecessary biopsy rate. When externally validated, the four DL models yielded robust and comparable performance, with AUCs of 82.2%-85.2%, sensitivities of 69.6%-76.0%, and specificities of 84.1%-89.2%, which outperformed the ACR TI-RADS. Compared to ACR TI-RADS, the DL models showed a higher biopsy rate of malignancy (71.6% -79.9% vs 37.7%, P<0.001) and a significantly lower unnecessary FNAB rate (8.5% -12.8% vs 40.7%, P<0.001). This study provides a noninvasive DL tool for accurate preoperative diagnosis of FNs, showing better performance than ACR TI-RADS and reducing unnecessary invasive interventions.

Ultra-low dose imaging in a standard axial field-of-view PET.

Lima T, Gomes CV, Fargier P, Strobel K, Leimgruber A

pubmed logopapersJul 21 2025
Though ultra-low dose (ULD) imaging offers notable benefits, its widespread clinical adoption faces challenges. Long-axial field-of-view (LAFOV) PET/CT systems are expensive and scarce, while artificial intelligence (AI) shows great potential but remains largely limited to specific systems and is not yet widely used in clinical practice. However, integrating AI techniques and technological advancements into ULD imaging is helping bridge the gap between standard axial field-of-view (SAFOV) and LAFOV PET/CT systems. This paper offers an initial evaluation of ULD capabilities using one of the latest SAFOV PET/CT device. A patient injected with 16.4 MBq <sup>18</sup>F-FDG underwent a local protocol consisting of a dynamic acquisition (first 30 min) of the abdominal section and a static whole body 74 min post-injection on a GE Omni PET/CT. From the acquired images we computed the dosimetry and compared clinical output from kidney function and brain uptake to kidney model and normal databases, respectively. The effective PET dose for this patient was 0.27 ± 0.01 mSv and the absorbed doses were 0.56 mGy, 0.89 mGy and 0.20 mGy, respectively to the brain, heart, and kidneys. The recorded kidney concentration closely followed the kidney model, matching the increase and decrease in activity concentration over time. Normal values for the z-score were observed for the brain uptake, indicating typical brain function and activity patterns consistent with healthy individuals. The signal to noise ration obtained in this study (13.1) was comparable to the LAFOV reported values. This study shows promising capabilities of ultra-low-dose imaging in SAFOV PET devices, previously deemed unattainable with SAFOV PET imaging.
Page 157 of 3563559 results
Show
per page

Ready to Sharpen Your Edge?

Join hundreds of your peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.