Sort by:
Page 48 of 61604 results

Enhancing Boundary Accuracy in Semantic Segmentation of Chest X-Ray Images Using Gaussian Process Regression.

Aljaddouh B, D Malathi D

pubmed logopapersMay 22 2025
This research aims to enhance X-ray lung segmentation by addressing boundary distortions in anatomical structures, with the objective of refining segmentation boundaries and improving the morphological shape of segmented objects. The proposed approach combines the K-segment principal curve with Gaussian Process Regression (GPR) to refine segmentation boundaries, evaluated using lung X-ray datasets at varying resolutions. Several state-of-the-art models, including U-Net, SegNet, and TransUnet, were also assessed for comparison. The model employed a custom kernel for GPR, combining Radial Basis Function (RBF) with a cosine similarity term. The effectiveness of the model was evaluated using metrics such as the Dice Coefficient (DC) and Jaccard Index (JC) for segmentation accuracy, along with Average Symmetric Surface Distance (ASSD) and Hausdorff Distance (HD) for boundary alignment. The proposed method achieved superior segmentation performance, particularly at the highest resolution (1024x1024 pixels), with a DC of 95.7% for the left lung and 94.1% for the right lung. Among the different models, TransUnet outperformed others across both the semantic segmentation and boundary refinement stages, showing significant improvements in DC, JC, ASSD, and HD. The results indicate that the proposed boundary refinement approach effectively improves the segmentation quality of lung X-rays, excelling in refining well-defined structures and achieving superior boundary alignment, showcasing its potential for clinical applications. However, limitations exist when dealing with irregular or unpredictable shapes, suggesting areas for future enhancement.

FasNet: a hybrid deep learning model with attention mechanisms and uncertainty estimation for liver tumor segmentation on LiTS17.

Singh R, Gupta S, Almogren A, Rehman AU, Bharany S, Altameem A, Choi J

pubmed logopapersMay 21 2025
Liver cancer, especially hepatocellular carcinoma (HCC), remains one of the most fatal cancers globally, emphasizing the critical need for accurate tumor segmentation to enable timely diagnosis and effective treatment planning. Traditional imaging techniques, such as CT and MRI, rely on manual interpretation, which can be both time-intensive and subject to variability. This study introduces FasNet, an innovative hybrid deep learning model that combines ResNet-50 and VGG-16 architectures, incorporating Channel and Spatial Attention mechanisms alongside Monte Carlo Dropout to improve segmentation precision and reliability. FasNet leverages ResNet-50's robust feature extraction and VGG-16's detailed spatial feature capture to deliver superior liver tumor segmentation accuracy. Channel and spatial attention mechanisms could selectively focus on the most relevant features and spatial regions for suitable segmentation with good accuracy and reliability. Monte Carlo Dropout estimates uncertainty and adds robustness, which is critical for high-stakes medical applications. Tested on the LiTS17 dataset, FasNet achieved a Dice Coefficient of 0.8766 and a Jaccard Index of 0.8487, surpassing several state-of-the-art methods. The Channel and Spatial Attention mechanisms in FasNet enhance feature selection, focusing on the most relevant spatial and channel information, while Monte Carlo Dropout improves model robustness and uncertainty estimation. These results position FasNet as a powerful diagnostic tool, offering precise and automated liver tumor segmentation that aids in early detection and precise treatment, ultimately enhancing patient outcomes.

The Desmoid Dilemma: Challenges and Opportunities in Assessing Tumor Burden and Therapeutic Response.

Chang YC, Nixon B, Souza F, Cardoso FN, Dayan E, Geiger EJ, Rosenberg A, D'Amato G, Subhawong T

pubmed logopapersMay 21 2025
Desmoid tumors are rare, locally invasive soft-tissue tumors with unpredictable clinical behavior. Imaging plays a crucial role in their diagnosis, measurement of disease burden, and assessment of treatment response. However, desmoid tumors' unique imaging features present challenges to conventional imaging metrics. The heterogeneous nature of these tumors, with a variable composition (fibrous, myxoid, or cellular), complicates accurate delineation of tumor boundaries and volumetric assessment. Furthermore, desmoid tumors can demonstrate prolonged stability or spontaneous regression, and biologic quiescence is often manifested by collagenization rather than bulk size reduction, making traditional size-based response criteria, such as Response Evaluation Criteria in Solid Tumors (RECIST), suboptimal. To overcome these limitations, advanced imaging techniques offer promising opportunities. Functional and parametric imaging methods, such as diffusion-weighted MRI, dynamic contrast-enhanced MRI, and T2 relaxometry, can provide insights into tumor cellularity and maturation. Radiomics and artificial intelligence approaches may enhance quantitative analysis by extracting and correlating complex imaging features with biological behavior. Moreover, imaging biomarkers could facilitate earlier detection of treatment efficacy or resistance, enabling tailored therapy. By integrating advanced imaging into clinical practice, it may be possible to refine the evaluation of disease burden and treatment response, ultimately improving the management and outcomes of patients with desmoid tumors.

Three-Blind Validation Strategy of Deep Learning Models for Image Segmentation.

Larroza A, Pérez-Benito FJ, Tendero R, Perez-Cortes JC, Román M, Llobet R

pubmed logopapersMay 21 2025
Image segmentation plays a central role in computer vision applications such as medical imaging, industrial inspection, and environmental monitoring. However, evaluating segmentation performance can be particularly challenging when ground truth is not clearly defined, as is often the case in tasks involving subjective interpretation. These challenges are amplified by inter- and intra-observer variability, which complicates the use of human annotations as a reliable reference. To address this, we propose a novel validation framework-referred to as the three-blind validation strategy-that enables rigorous assessment of segmentation models in contexts where subjectivity and label variability are significant. The core idea is to have a third independent expert, blind to the labeler identities, assess a shuffled set of segmentations produced by multiple human annotators and/or automated models. This allows for the unbiased evaluation of model performance and helps uncover patterns of disagreement that may indicate systematic issues with either human or machine annotations. The primary objective of this study is to introduce and demonstrate this validation strategy as a generalizable framework for robust model evaluation in subjective segmentation tasks. We illustrate its practical implementation in a mammography use case involving dense tissue segmentation while emphasizing its potential applicability to a broad range of segmentation scenarios.

Cardiac Magnetic Resonance Imaging in the German National Cohort: Automated Segmentation of Short-Axis Cine Images and Post-Processing Quality Control

Full, P. M., Schirrmeister, R. T., Hein, M., Russe, M. F., Reisert, M., Ammann, C., Greiser, K. H., Niendorf, T., Pischon, T., Schulz-Menger, J., Maier-Hein, K. H., Bamberg, F., Rospleszcz, S., Schlett, C. L., Schuppert, C.

medrxiv logopreprintMay 21 2025
PurposeTo develop a segmentation and quality control pipeline for short-axis cardiac magnetic resonance (CMR) cine images from the prospective, multi-center German National Cohort (NAKO). Materials and MethodsA deep learning model for semantic segmentation, based on the nnU-Net architecture, was applied to full-cycle short-axis cine images from 29,908 baseline participants. The primary objective was to determine data on structure and function for both ventricles (LV, RV), including end diastolic volumes (EDV), end systolic volumes (ESV), and LV myocardial mass. Quality control measures included a visual assessment of outliers in morphofunctional parameters, inter- and intra-ventricular phase differences, and LV time-volume curves (TVC). These were adjudicated using a five-point rating scale, ranging from five (excellent) to one (non-diagnostic), with ratings of three or lower subject to exclusion. The predictive value of outlier criteria for inclusion and exclusion was analyzed using receiver operating characteristics. ResultsThe segmentation model generated complete data for 29,609 participants (incomplete in 1.0%) and 5,082 cases (17.0 %) were visually assessed. Quality assurance yielded a sample of 26,899 participants with excellent or good quality (89.9%; exclusion of 1,875 participants due to image quality issues and 835 cases due to segmentation quality issues). TVC was the strongest single discriminator between included and excluded participants (AUC: 0.684). Of the two-category combinations, the pairing of TVC and phases provided the greatest improvement over TVC alone (AUC difference: 0.044; p<0.001). The best performance was observed when all three categories were combined (AUC: 0.748). Extending the quality-controlled sample to include acceptable quality ratings, a total of 28,413 (95.0%) participants were available. ConclusionThe implemented pipeline facilitated the automated segmentation of an extensive CMR dataset, integrating quality control measures. This methodology ensures that ensuing quantitative analyses are conducted with a diminished risk of bias.

TAGS: 3D Tumor-Adaptive Guidance for SAM

Sirui Li, Linkai Peng, Zheyuan Zhang, Gorkem Durak, Ulas Bagci

arxiv logopreprintMay 21 2025
Foundation models (FMs) such as CLIP and SAM have recently shown great promise in image segmentation tasks, yet their adaptation to 3D medical imaging-particularly for pathology detection and segmentation-remains underexplored. A critical challenge arises from the domain gap between natural images and medical volumes: existing FMs, pre-trained on 2D data, struggle to capture 3D anatomical context, limiting their utility in clinical applications like tumor segmentation. To address this, we propose an adaptation framework called TAGS: Tumor Adaptive Guidance for SAM, which unlocks 2D FMs for 3D medical tasks through multi-prompt fusion. By preserving most of the pre-trained weights, our approach enhances SAM's spatial feature extraction using CLIP's semantic insights and anatomy-specific prompts. Extensive experiments on three open-source tumor segmentation datasets prove that our model surpasses the state-of-the-art medical image segmentation models (+46.88% over nnUNet), interactive segmentation frameworks, and other established medical FMs, including SAM-Med2D, SAM-Med3D, SegVol, Universal, 3D-Adapter, and SAM-B (at least +13% over them). This highlights the robustness and adaptability of our proposed framework across diverse medical segmentation tasks.

BrainView: A Cloud-based Deep Learning System for Brain Image Segmentation, Tumor Detection and Visualization.

Ghose P, Jamil HM

pubmed logopapersMay 21 2025
A brain tumor is an abnormal growth in the brain that disrupts its functionality and poses a significant threat to human life by damaging neurons. Early detection and classification of brain tumors are crucial to prevent complications and maintain good health. Recent advancements in deep learning techniques have shown immense potential in image classification and segmentation for tumor identification and classification. In this study, we present a platform, BrainView, for detection, and segmentation of brain tumors from Magnetic Resonance Images (MRI) using deep learning. We utilized EfficientNetB7 pre-trained model to design our proposed DeepBrainNet classification model for analyzing brain MRI images to classify its type. We also proposed a EfficinetNetB7 based image segmentation model, called the EffB7-UNet, for tumor localization. Experimental results show significantly high classification (99.96%) and segmentation (92.734%) accuracies for our proposed models. Finally, we discuss the contours of a cloud application for BrainView using Flask and Flutter to help researchers and clinicians use our machine learning models online for research purposes.

Enhancing nuclei segmentation in breast histopathology images using U-Net with backbone architectures.

C V LP, V G B, Bhooshan RS

pubmed logopapersMay 21 2025
Breast cancer remains a leading cause of mortality among women worldwide, underscoring the need for accurate and timely diagnostic methods. Precise segmentation of nuclei in breast histopathology images is crucial for effective diagnosis and prognosis, offering critical insights into tumor characteristics and informing treatment strategies. This paper presents an enhanced U-Net architecture utilizing ResNet-34 as an advanced backbone, aimed at improving nuclei segmentation performance. The proposed model is evaluated and compared with standard U-Net and its other variants, including U-Net with VGG-16 and Inception-v3 backbones, using the BreCaHad dataset with nuclei masks generated through ImageJ software. The U-Net model with ResNet-34 backbone achieved superior performance, recording an Intersection over Union (IoU) score of 0.795, significantly outperforming the basic U-Net's IoU score of 0.725. The integration of advanced backbones and data augmentation techniques substantially improved segmentation accuracy, especially on limited medical imaging datasets. Comparative analysis demonstrated that ResNet-34 consistently surpassed other configurations across multiple metrics, including IoU, accuracy, precision, and F1 score. Further validation on the BNS and MoNuSeg-2018 datasets confirmed the robustness of the proposed model. This study highlights the potential of advanced deep learning architectures combined with augmentation methods to address challenges in nuclei segmentation, contributing to the development of more effective clinical diagnostic tools and improved patient care outcomes.

SAMA-UNet: Enhancing Medical Image Segmentation with Self-Adaptive Mamba-Like Attention and Causal-Resonance Learning

Saqib Qamar, Mohd Fazil, Parvez Ahmad, Ghulam Muhammad

arxiv logopreprintMay 21 2025
Medical image segmentation plays an important role in various clinical applications, but existing models often struggle with the computational inefficiencies and challenges posed by complex medical data. State Space Sequence Models (SSMs) have demonstrated promise in modeling long-range dependencies with linear computational complexity, yet their application in medical image segmentation remains hindered by incompatibilities with image tokens and autoregressive assumptions. Moreover, it is difficult to achieve a balance in capturing both local fine-grained information and global semantic dependencies. To address these challenges, we introduce SAMA-UNet, a novel architecture for medical image segmentation. A key innovation is the Self-Adaptive Mamba-like Aggregated Attention (SAMA) block, which integrates contextual self-attention with dynamic weight modulation to prioritise the most relevant features based on local and global contexts. This approach reduces computational complexity and improves the representation of complex image features across multiple scales. We also suggest the Causal-Resonance Multi-Scale Module (CR-MSM), which enhances the flow of information between the encoder and decoder by using causal resonance learning. This mechanism allows the model to automatically adjust feature resolution and causal dependencies across scales, leading to better semantic alignment between the low-level and high-level features in U-shaped architectures. Experiments on MRI, CT, and endoscopy images show that SAMA-UNet performs better in segmentation accuracy than current methods using CNN, Transformer, and Mamba. The implementation is publicly available at GitHub.

Seeing the Trees for the Forest: Rethinking Weakly-Supervised Medical Visual Grounding

Ta Duc Huy, Duy Anh Huynh, Yutong Xie, Yuankai Qi, Qi Chen, Phi Le Nguyen, Sen Kim Tran, Son Lam Phung, Anton van den Hengel, Zhibin Liao, Minh-Son To, Johan W. Verjans, Vu Minh Hieu Phan

arxiv logopreprintMay 21 2025
Visual grounding (VG) is the capability to identify the specific regions in an image associated with a particular text description. In medical imaging, VG enhances interpretability by highlighting relevant pathological features corresponding to textual descriptions, improving model transparency and trustworthiness for wider adoption of deep learning models in clinical practice. Current models struggle to associate textual descriptions with disease regions due to inefficient attention mechanisms and a lack of fine-grained token representations. In this paper, we empirically demonstrate two key observations. First, current VLMs assign high norms to background tokens, diverting the model's attention from regions of disease. Second, the global tokens used for cross-modal learning are not representative of local disease tokens. This hampers identifying correlations between the text and disease tokens. To address this, we introduce simple, yet effective Disease-Aware Prompting (DAP) process, which uses the explainability map of a VLM to identify the appropriate image features. This simple strategy amplifies disease-relevant regions while suppressing background interference. Without any additional pixel-level annotations, DAP improves visual grounding accuracy by 20.74% compared to state-of-the-art methods across three major chest X-ray datasets.
Page 48 of 61604 results
Show
per page

Ready to Sharpen Your Edge?

Join hundreds of your peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.