Back to all papers

User-centric eXplainable AI criteria for implementing AI-based denoising in PET/CT.

Authors

Champendal M,Ribeiro RT,Müller H,Prior JO,Sá Dos Reis C

Affiliations (5)

  • HESAV School of Health Sciences - Vaud, HES-SO University of Applied Sciences and Arts Western Switzerland, Avenue de Beaumont 21, 1011 Lausanne, Switzerland; Faculty of Biology and Medicine, University of Lausanne, Unicentre, CH, 1015 Lausanne, Switzerland. Electronic address: [email protected].
  • HESAV School of Health Sciences - Vaud, HES-SO University of Applied Sciences and Arts Western Switzerland, Avenue de Beaumont 21, 1011 Lausanne, Switzerland. Electronic address: [email protected].
  • Informatics Institute, University of Applied Sciences Western Switzerland (HES-SO Valais) Sierre, CH, Rue de l'Industrie 23, 1950 Sion, Switzerland; Medical Faculty, University of Geneva, CH, Rue Michel Servet, Genève 4 11211, Switzerland. Electronic address: [email protected].
  • Faculty of Biology and Medicine, University of Lausanne, Unicentre, CH, 1015 Lausanne, Switzerland; Nuclear Medicine and Molecular Imaging Department, Lausanne University Hospital (CHUV): Lausanne, CH, Rue du Bugnon 46, 1011 Lausanne, Switzerland. Electronic address: [email protected].
  • HESAV School of Health Sciences - Vaud, HES-SO University of Applied Sciences and Arts Western Switzerland, Avenue de Beaumont 21, 1011 Lausanne, Switzerland. Electronic address: [email protected].

Abstract

The clinical adoption of AI-based denoising in PET/CT relies on the development of transparent and trustworthy tools that align with the radiographers' needs and support integration into routine practice. This study aims to determine the key characteristics of an eXplainable Artificial Intelligence (XAI)/tool aligning the radiographers' needs to facilitate the clinical adoption of AI-based denoising algorithm in PET/CT. Two focus groups were organised, involving ten voluntary participants recruited from nuclear medicine departments from Western-Switzerland, forming a convenience sample of radiographers. Two different scenarios, matching or mismatching the ground truth were used to identify their needs and the questions they would like to ask to understand the AI-denoising algorithm. Additionally, the characteristics that an XAI tool should possess to best meet their needs were investigated. Content analysis was performed following the three steps outlined by Wanlin. Ethics cleared the study. Ten radiographers (aged 31-60y) identified two levels of explanation: (1) simple, global explanations with numerical confidence levels for rapid understanding in routine settings; (2) detailed, case-specific explanations using mixed formats where necessary, depending on the clinical situation and users to build confidence and support decision-making. Key questions include the functions of the algorithm ('what'), the clinical context ('when') and the dependency of the results ('how'). An effective XAI tool should be easy, adaptable, user-friendly and not disruptive to workflows. Radiographers need two levels of explanation from XAI tools: global summaries that preserve workflow efficiency and detailed, case-specific insights when needed. Meeting these needs is key to fostering trust, understanding, and integration of AI-based denoising in PET/CT. Implementing adaptive XAI tools tailored to radiographers' needs can support clinical workflows and accelerate the adoption of AI in PET/CT imaging.

Topics

Journal Article

Ready to Sharpen Your Edge?

Join hundreds of your peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.