Back to all papers

Foundation versus domain-specific models for left ventricular segmentation on cardiac ultrasound.

June 6, 2025pubmed logopapers

Authors

Chao CJ,Gu YR,Kumar W,Xiang T,Appari L,Wu J,Farina JM,Wraith R,Jeong J,Arsanjani R,Kane GC,Oh JK,Langlotz CP,Banerjee I,Fei-Fei L,Adeli E

Affiliations (12)

  • Department of Cardiovascular Medicine, Mayo Clinic Rochester, Rochester, MN, USA.
  • Stanford Institute for Human-Centered Artificial Intelligence, Stanford University, Stanford, CA, USA.
  • Department of Radiology, Mayo Clinic Arizona, Scottsdale, AZ, USA.
  • School of Computing and Augmented Intelligence, Arizona State University, Phoenix, AZ, USA.
  • Department of Cardiovascular Medicine, Mayo Clinic Arizona, Scottsdale, AZ, USA.
  • Center of Artificial Intelligence in Medical Imaging, Stanford University, Stanford, CA, USA.
  • Department of Radiology, Stanford University, Stanford, CA, USA.
  • Stanford Institute for Human-Centered Artificial Intelligence, Stanford University, Stanford, CA, USA. [email protected].
  • Department of Computer Science, Stanford University, Stanford, CA, USA. [email protected].
  • Stanford Institute for Human-Centered Artificial Intelligence, Stanford University, Stanford, CA, USA. [email protected].
  • Department of Computer Science, Stanford University, Stanford, CA, USA. [email protected].
  • Department of Psychiatry and Behavioral Sciences, Stanford University, Stanford, CA, USA. [email protected].

Abstract

The Segment Anything Model (SAM) was fine-tuned on the EchoNet-Dynamic dataset and evaluated on external transthoracic echocardiography (TTE) and Point-of-Care Ultrasound (POCUS) datasets from CAMUS (University Hospital of St Etienne) and Mayo Clinic (99 patients: 58 TTE, 41 POCUS). Fine-tuned SAM was superior or comparable to MedSAM. The fine-tuned SAM also outperformed EchoNet and U-Net models, demonstrating strong generalization, especially on apical 2-chamber (A2C) images (fine-tuned SAM vs. EchoNet: CAMUS-A2C: DSC 0.891 ± 0.040 vs. 0.752 ± 0.196, p < 0.0001) and POCUS (DSC 0.857 ± 0.047 vs. 0.667 ± 0.279, p < 0.0001). Additionally, SAM-enhanced workflow reduced annotation time by 50% (11.6 ± 4.5 sec vs. 5.7 ± 1.7 sec, p < 0.0001) while maintaining segmentation quality. We demonstrated an effective strategy for fine-tuning a vision foundation model for enhancing clinical workflow efficiency and supporting human-AI collaboration.

Topics

Journal Article

Ready to Sharpen Your Edge?

Subscribe to join 7,600+ peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.