A deep learning model leveraging SAM improved segmentation speed and matched radiologist performance in classifying ovarian lesions on MRI.
Key Details
- 1Researchers from Johns Hopkins developed an MRI-based, end-to-end DL pipeline incorporating Meta’s Segment Anything Model (SAM) and DenseNet-121.
- 2Integrated model reduced segmentation time by 4 minutes per lesion compared to manual segmentation.
- 3SAM achieved a Dice coefficient of 0.86 to 0.88 for lesion segmentation.
- 4DL model achieved an AUC of 0.85 internally and 0.79 externally, on par with radiologists’ AUC of 0.84 (p > 0.05).
- 5Training data included 534 lesions (internal) and 87 lesions (external) from the US and Taiwan.
Why It Matters
This study demonstrates that foundation AI models can efficiently streamlines lesion segmentation and classification with accuracy rivaling radiologists, suggesting a path toward more automated and collaborative clinical workflows in radiology.

Source
AuntMinnie
Related News

•Health Imaging
AI as Second Reader Surpasses Radiologists in Breast Cancer Screening
AI used as a second reader on mammograms improves cancer detection rates compared to radiologists alone.

•Health Imaging
AI-Powered Ultrasound Tool Predicts Delivery Timing for Pregnant Patients
Researchers have created an AI model using ultrasound to accurately forecast expectant mothers’ delivery timelines.

•AuntMinnie
ChatGPT-4 Turbo Powers Postdeployment Monitoring of ICH Detection AI
Researchers found ChatGPT-4 Turbo could efficiently monitor the performance of Aidoc's ICH detection AI across real-world radiology practices.