Sort by:
Page 88 of 99990 results

Incorporating organ deformation in biological modeling and patient outcome study for permanent prostate brachytherapy.

To S, Mavroidis P, Chen RC, Wang A, Royce T, Tan X, Zhu T, Lian J

pubmed logopapersMay 28 2025
Permanent prostate brachytherapy has inherent intraoperative organ deformation due to the inflatable trans-rectal ultrasound probe cover. Since the majority of the dose is delivered postoperatively with no deformation, the dosimetry approved at the time of implant may not accurately represent the dose delivered to the target and organs at risk. We aimed to evaluate the biological effect of the prostate deformation and its correlation with patient-reported outcomes. We prospectively acquired ultrasound images of the prostate pre- and postprobe cover inflation for 27 patients undergoing I-125 seed implant. The coordinates of implanted seeds from approved clinical plan were transferred to deformation-corrected prostate to simulate the actual dosimetry using a machine learning-based deformable image registration. The DVHs of both sets of plans were reduced to biologically effective dose (BED) distribution and subsequently to Tumor Control Probability (TCP) and Normal Tissue Complication Probability (NTCP) metrics. The change in fourteen patient-reported rectal and urinary symptoms between pretreatment to 6 months post-op time points were correlated with the TCP and NTCP metrics using the area under the curve (AUC) and odds ratio (OR). Between the clinical and the deformation corrected research plans, the mean TCP decreased by 9.4% (p < 0.01), whereas mean NTCP of rectum decreased by 10.3% and that of urethra increased by 16.3%, respectively (p < 0.01). For the diarrhea symptom, the deformation corrected research plans showed AUC=0.75 and OR = 8.9 (1.3-58.8) for the threshold NTCP>20%, while the clinical plan showed AUC=0.56 and OR = 1.4 (0.2 to 9.0). For the symptom of urinary control, the deformation corrected research plans showed AUC = 0.70, OR = 6.9 (0.6 to 78.0) for the threshold of NTCP>15%, while the clinical plan showed AUC = 0.51 and no positive OR. Taking organ deformation into consideration, clinical brachytherapy plans showed worse tumor coverage, worse urethra sparing but better rectal sparing. The deformation corrected research plans showed a stronger correlation with the patient-reported outcome than the clinical plans for the symptoms of diarrhea and urinary control.

Integrating SEResNet101 and SE-VGG19 for advanced cervical lesion detection: a step forward in precision oncology.

Ye Y, Chen Y, Pan J, Li P, Ni F, He H

pubmed logopapersMay 28 2025
Cervical cancer remains a significant global health issue, with accurate differentiation between low-grade (LSIL) and high-grade squamous intraepithelial lesions (HSIL) crucial for effective screening and management. Current methods, such as Pap smears and HPV testing, often fall short in sensitivity and specificity. Deep learning models hold the potential to enhance the accuracy of cervical cancer screening but require thorough evaluation to ascertain their practical utility. This study compares the performance of two advanced deep learning models, SEResNet101 and SE-VGG19, in classifying cervical lesions using a dataset of 3,305 high-quality colposcopy images. We assessed the models based on their accuracy, sensitivity, specificity, and area under the receiver operating characteristic curve (AUC). The SEResNet101 model demonstrated superior performance over SE-VGG19 across all evaluated metrics. Specifically, SEResNet101 achieved a sensitivity of 95%, a specificity of 97%, and an AUC of 0.98, compared to 89% sensitivity, 93% specificity, and an AUC of 0.94 for SE-VGG19. These findings suggest that SEResNet101 could significantly reduce both over- and under-treatment rates by enhancing diagnostic precision. Our results indicate that SEResNet101 offers a promising enhancement over existing screening methods, integrating advanced deep learning algorithms to significantly improve the precision of cervical lesion classification. This study advocates for the inclusion of SEResNet101 in clinical workflows to enhance cervical cancer screening protocols, thereby improving patient outcomes. Future work should focus on multicentric trials to validate these findings and facilitate widespread clinical adoption.

Efficient feature extraction using light-weight CNN attention-based deep learning architectures for ultrasound fetal plane classification.

Sivasubramanian A, Sasidharan D, Sowmya V, Ravi V

pubmed logopapersMay 28 2025
Ultrasound fetal imaging is beneficial to support prenatal development because it is affordable and non-intrusive. Nevertheless, fetal plane classification (FPC) remains challenging and time-consuming for obstetricians since it depends on nuanced clinical aspects, which increases the difficulty in identifying relevant features of the fetal anatomy. Thus, to assist with its accurate feature extraction, a lightweight artificial intelligence architecture leveraging convolutional neural networks and attention mechanisms is proposed to classify the largest benchmark ultrasound dataset. The approach fine-tunes from lightweight EfficientNet feature extraction backbones pre-trained on the ImageNet1k. to classify key fetal planes such as the brain, femur, thorax, cervix, and abdomen. Our methodology incorporates the attention mechanism to refine features and 3-layer perceptrons for classification, achieving superior performance with the highest Top-1 accuracy of 96.25%, Top-2 accuracy of 99.80% and F1-Score of 0.9576. Importantly, the model has 40x fewer trainable parameters than existing benchmark ensemble or transformer pipelines, facilitating easy deployment on edge devices to help clinical practitioners with real-time FPC. The findings are also interpreted using GradCAM to carry out clinical correlation to aid doctors with diagnostics and improve treatment plans for expectant mothers.

A vessel bifurcation landmark pair dataset for abdominal CT deformable image registration (DIR) validation.

Criscuolo ER, Zhang Z, Hao Y, Yang D

pubmed logopapersMay 28 2025
Deformable image registration (DIR) is an enabling technology in many diagnostic and therapeutic tasks. Despite this, DIR algorithms have limited clinical use, largely due to a lack of benchmark datasets for quality assurance during development. DIRs of intra-patient abdominal CTs are among the most challenging registration scenarios due to significant organ deformations and inconsistent image content. To support future algorithm development, here we introduce our first-of-its-kind abdominal CT DIR benchmark dataset, comprising large numbers of highly accurate landmark pairs on matching blood vessel bifurcations. Abdominal CT image pairs of 30 patients were acquired from several publicly available repositories as well as the authors' institution with IRB approval. The two CTs of each pair were originally acquired for the same patient but on different days. An image processing workflow was developed and applied to each CT image pair: (1) Abdominal organs were segmented with a deep learning model, and image intensity within organ masks was overwritten. (2) Matching image patches were manually identified between two CTs of each image pair. (3) Vessel bifurcation landmarks were labeled on one image of each image patch pair. (4) Image patches were deformably registered, and landmarks were projected onto the second image. (5) Landmark pair locations were refined manually or with an automated process. This workflow resulted in 1895 total landmark pairs, or 63 per case on average. Estimates of the landmark pair accuracy using digital phantoms were 0.7 mm ± 1.2 mm. The data are published in Zenodo at https://doi.org/10.5281/zenodo.14362785. Instructions for use can be found at https://github.com/deshanyang/Abdominal-DIR-QA. This dataset is a first-of-its-kind for abdominal DIR validation. The number, accuracy, and distribution of landmark pairs will allow for robust validation of DIR algorithms with precision beyond what is currently available.

Contrast-Enhanced Ultrasound for Hepatocellular Carcinoma Diagnosis-<i>AJR</i> Expert Panel Narrative Review.

Li L, Burgio MD, Fetzer DT, Ferraioli G, Lyshchik A, Meloni MF, Rafailidis V, Sidhu PS, Vilgrain V, Wilson SR, Zhou J

pubmed logopapersMay 28 2025
Despite growing clinical use of contrast-enhanced ultrasound (CEUS), inconsistency remains in the modality's role in clinical pathways for hepatocellular carcinoma (HCC) diagnosis and management. This AJR Expert Panel Narrative Review provides practical insights on the use of CEUS for the diagnosis of HCC across populations, including individuals at high risk for HCC, individuals with metabolic dysfunction-associated steatotic liver disease, and remaining individuals not at high risk for HCC. Considerations addressed with respect to high-risk patients include CEUS diagnostic criteria for HCC, use of CEUS for differentiating HCC from non-HCC malignancy, use of CEUS for small (≤2 cm) lesions, use of CEUS for characterizing occult lesions on B-mode ultrasound, and use of CEUS for indeterminate lesions on CT or MRI. Representative literature addressing the use of CEUS for HCC diagnosis as well as gaps in knowledge requiring further investigation are highlighted. Throughout these discussions, the article distinguishes two broad types of ultrasound contrast agents used for liver imaging: pure blood-pool agents and a combined blood-pool and Kupffer-cell agent. Additional topics include the use of CEUS for treatment response assessment after nonradiation therapies and implications of artificial intelligence technologies. The article concludes with a series of consensus statements from the author panel.

Toward diffusion MRI in the diagnosis and treatment of pancreatic cancer.

Lee J, Lin T, He Y, Wu Y, Qin J

pubmed logopapersMay 28 2025
Pancreatic cancer is a highly aggressive malignancy with rising incidence and mortality rates, often diagnosed at advanced stages. Conventional imaging methods, such as computed tomography (CT) and magnetic resonance imaging (MRI), struggle to assess tumor characteristics and vascular involvement, which are crucial for treatment planning. This paper explores the potential of diffusion magnetic resonance imaging (dMRI) in enhancing pancreatic cancer diagnosis and treatment. Diffusion-based techniques, such as diffusion-weighted imaging (DWI), diffusion tensor imaging (DTI), intravoxel incoherent motion (IVIM), and diffusion kurtosis imaging (DKI), combined with emerging AI‑powered analysis, provide insights into tissue microstructure, allowing for earlier detection and improved evaluation of tumor cellularity. These methods may help assess prognosis and monitor therapy response by tracking diffusion and perfusion metrics. However, challenges remain, such as standardized protocols and robust data analysis pipelines. Ongoing research, including deep learning applications, aims to improve reliability, and dMRI shows promise in providing functional insights and improving patient outcomes. Further clinical validation is necessary to maximize its benefits.

Dose calculation in nuclear medicine with magnetic resonance imaging images using Monte Carlo method.

Vu LH, Thao NTP, Trung NT, Hau PVT, Hong Loan TT

pubmed logopapersMay 27 2025
In recent years, scientists have been trying to convert magnetic resonance imaging (MRI) images into computed tomography (CT) images for dose calculations while taking advantage of the benefits of MRI images. The main approaches for image conversion are bulk density, Atlas registration, and machine learning. These methods have limitations in accuracy and time consumption and require large datasets to convert images. In this study, the novel 'voxels spawn voxels' technique combined with the 'orthonormalize' feature in Carimas software was developed to build a conversion dataset from MRI intensity to Hounsfield unit value for some structural regions including gluteus maximus, liver, kidneys, spleen, pancreas, and colon. The original CT images and the converted MRI images were imported into the Geant4/Gamos software for dose calculation. It gives good results (<5%) in most organs except the intestine (18%).

Deep learning network enhances imaging quality of low-b-value diffusion-weighted imaging and improves lesion detection in prostate cancer.

Liu Z, Gu WJ, Wan FN, Chen ZZ, Kong YY, Liu XH, Ye DW, Dai B

pubmed logopapersMay 27 2025
Diffusion-weighted imaging with higher b-value improves detection rate for prostate cancer lesions. However, obtaining high b-value DWI requires more advanced hardware and software configuration. Here we use a novel deep learning network, NAFNet, to generate a deep learning reconstructed (DLR<sub>1500</sub>) images from 800 b-value to mimic 1500 b-value images, and to evaluate its performance and lesion detection improvements based on whole-slide images (WSI). We enrolled 303 prostate cancer patients with both 800 and 1500 b-values from Fudan University Shanghai Cancer Centre between 2017 and 2020. We assigned these patients to the training and validation set in a 2:1 ratio. The testing set included 36 prostate cancer patients from an independent institute who had only preoperative DWI at 800 b-value. Two senior radiology doctors and two junior radiology doctors read and delineated cancer lesions on DLR<sub>1500</sub>, original 800 and 1500 b-values DWI images. WSI were used as the ground truth to assess the lesion detection improvement of DLR<sub>1500</sub> images in the testing set. After training and generating, within junior radiology doctors, the diagnostic AUC based on DLR<sub>1500</sub> images is not inferior to that based on 1500 b-value images (0.832 (0.788-0.876) vs. 0.821 (0.747-0.899), P = 0.824). The same phenomenon is also observed in senior radiology doctors. Furthermore, in the testing set, DLR<sub>1500</sub> images could significantly enhance junior radiology doctors' diagnostic performance than 800 b-value images (0.848 (0.758-0.938) vs. 0.752 (0.661-0.843), P = 0.043). DLR<sub>1500</sub> DWIs were comparable in quality to original 1500 b-value images within both junior and senior radiology doctors. NAFNet based DWI enhancement can significantly improve the image quality of 800 b-value DWI, and therefore promote the accuracy of prostate cancer lesion detection for junior radiology doctors.

Prostate Cancer Screening with Artificial Intelligence-Enhanced Micro-Ultrasound: A Comparative Study with Traditional Methods

Muhammad Imran, Wayne G. Brisbane, Li-Ming Su, Jason P. Joseph, Wei Shao

arxiv logopreprintMay 27 2025
Background and objective: Micro-ultrasound (micro-US) is a novel imaging modality with diagnostic accuracy comparable to MRI for detecting clinically significant prostate cancer (csPCa). We investigated whether artificial intelligence (AI) interpretation of micro-US can outperform clinical screening methods using PSA and digital rectal examination (DRE). Methods: We retrospectively studied 145 men who underwent micro-US guided biopsy (79 with csPCa, 66 without). A self-supervised convolutional autoencoder was used to extract deep image features from 2D micro-US slices. Random forest classifiers were trained using five-fold cross-validation to predict csPCa at the slice level. Patients were classified as csPCa-positive if 88 or more consecutive slices were predicted positive. Model performance was compared with a classifier using PSA, DRE, prostate volume, and age. Key findings and limitations: The AI-based micro-US model and clinical screening model achieved AUROCs of 0.871 and 0.753, respectively. At a fixed threshold, the micro-US model achieved 92.5% sensitivity and 68.1% specificity, while the clinical model showed 96.2% sensitivity but only 27.3% specificity. Limitations include a retrospective single-center design and lack of external validation. Conclusions and clinical implications: AI-interpreted micro-US improves specificity while maintaining high sensitivity for csPCa detection. This method may reduce unnecessary biopsies and serve as a low-cost alternative to PSA-based screening. Patient summary: We developed an AI system to analyze prostate micro-ultrasound images. It outperformed PSA and DRE in detecting aggressive cancer and may help avoid unnecessary biopsies.

Dual-energy CT combined with histogram parameters in the assessment of perineural invasion in colorectal cancer.

Wang Y, Tan H, Li S, Long C, Zhou B, Wang Z, Cao Y

pubmed logopapersMay 27 2025
The purpose is to evaluate the predictive value of dual-energy CT (DECT) combined with histogram parameters and a clinical prediction model for perineural invasion (PNI) in colorectal cancer (CRC). We retrospectively analyzed clinical and imaging data from 173 CRC patients who underwent preoperative DECT-enhanced scanning at two centers. Data from Qinghai University Affiliated Hospital (n = 120) were randomly divided into training and validation sets, while data from Lanzhou University Second Hospital (n = 53) served as the external validation set. Regions of interest (ROIs) were delineated to extract spectral and histogram parameters, and multivariate logistic regression identified optimal predictors. Six machine learning models-support vector machine (SVM), decision tree (DT), random forest (RF), logistic regression (LR), k-nearest neighbors (KNN), and extreme gradient boosting (XGBoost)-were constructed. Model performance and clinical utility were assessed using receiver operating characteristic (ROC) curves, calibration curves, and decision curve analysis (DCA). Four independent predictive factors were identified through multivariate analysis: entropy, CT40<sub>KeV</sub>, CEA, and skewness. Among the six classifier models, RF model demonstrated the best performance in the training set (AUC = 0.918, 95% CI: 0.862-0.969). In the validation set, RF outperformed other models (AUC = 0.885, 95% CI: 0.772-0.972). Notably, in the external validation set, the XGBoost model achieved the highest performance (AUC = 0.823, 95% CI: 0.672-0.945). Dual-energy CT-based combined with histogram parameters and clinical prediction modeling can be effectively used for preoperative noninvasive assessment of perineural invasion in colorectal cancer.
Page 88 of 99990 results
Show
per page

Ready to Sharpen Your Edge?

Join hundreds of your peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.