Evaluating artificial intelligence chatbots for patient education in oral and maxillofacial radiology.

Authors

Helvacioglu-Yigit D,Demirturk H,Ali K,Tamimi D,Koenig L,Almashraqi A

Affiliations (5)

  • College of Dental Medicine, QU Health, Qatar University, Doha, Qatar. Electronic address: [email protected].
  • University of Pittsburgh School of Dental Medicine, Pittsburgh, PA, USA; Oral and Maxillofacial Radiology Consultant, Private Practice, Wexford, PA, USA.
  • College of Dental Medicine, QU Health, Qatar University, Doha, Qatar.
  • Oral and Maxillofacial Radiology Consultant, Private Practice, Orlando, FL, USA.
  • Marquette University School of Dentistry, Milwaukee, WI, USA.

Abstract

This study aimed to compare the quality and readability of the responses generated by 3 publicly available artificial intelligence (AI) chatbots in answering frequently asked questions (FAQs) related to Oral and Maxillofacial Radiology (OMR) to assess their suitability for patient education. Fifteen OMR-related questions were selected from professional patient information websites. These questions were posed to ChatGPT-3.5 by OpenAI, Gemini 1.5 Pro by Google, and Copilot by Microsoft to generate responses. Three board-certified OMR specialists evaluated the responses regarding scientific adequacy, ease of understanding, and overall reader satisfaction. Readability was assessed using the Flesch-Kincaid Grade Level (FKGL) and Flesch Reading Ease (FRE) scores. The Wilcoxon signed-rank test was conducted to compare the scores assigned by the evaluators to the responses from the chatbots and professional websites. Interevaluator agreement was examined by calculating the Fleiss kappa coefficient. There were no significant differences between groups in terms of scientific adequacy. In terms of readability, chatbots had overall mean FKGL and FRE scores of 12.97 and 34.11, respectively. Interevaluator agreement level was generally high. Although chatbots are relatively good at responding to FAQs, validating AI-generated information using input from healthcare professionals can enhance patient care and safety. Readability of the text content in the chatbots and websites requires high reading levels.

Topics

Artificial IntelligencePatient Education as TopicRadiologyRadiography, DentalJournal Article

Ready to Sharpen Your Edge?

Join hundreds of your peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.