Isfahan University of Medical Sciences

Science Communicator Platform

Stay connected! Follow us on X network (Twitter):
Share this content! On (X network) By
Efficacy and Empathy of Ai Chatbots in Answering Frequently Asked Questions on Oral Oncology Publisher



Rokhshad R1 ; Khoury ZH2 ; Mohammadrahimi H3 ; Motie P4 ; Price JB5 ; Tavares T6 ; Jessri M7, 8 ; Bavarian R9, 10 ; Sciubba JJ11 ; Sultan AS5, 12
Authors
Show Affiliations
Authors Affiliations
  1. 1. Department of Pediatric Dentistry, Loma Linda School of Dentistry, CA, United States
  2. 2. Department of Oral Diagnostic Sciences & Research, School of Dentistry, Meharry Medical College, TN, United States
  3. 3. Department of Dentistry and Oral Health, Aarhus University, Aarhus, Denmark
  4. 4. Medical Image and Signal Processing Research Center, Isfahan University of Medical Sciences, Isfahan, Iran
  5. 5. Division of Artificial Intelligence Research, Department of Oncology and Diagnostic Sciences, University of Maryland School of Dentistry, Baltimore, MD, United States
  6. 6. Department of Comprehensive Dentistry, UT Health San Antonio, School of Dentistry, San Antonio, TX, United States
  7. 7. Oral Medicine and Pathology Department, School of Dentistry, University of Queensland, Herston, QLD, Australia
  8. 8. Oral Medicine Department, MetroNorth Hospital and Health Services, Queensland Health, QLD, Australia
  9. 9. Department of Oral and Maxillofacial Surgery, Massachusetts General Hospital, Boston, MA, United States
  10. 10. Department of Oral and Maxillofacial Surgery, Harvard School of Dental Medicine, Boston, MA, United States
  11. 11. Department of Otolaryngology, Head & Neck Surgery, The Johns Hopkins University, Baltimore, MD, United States
  12. 12. University of Maryland Marlene and Stewart Greenebaum Comprehensive Cancer Center, Baltimore, MD, United States

Source: Oral Surgery, Oral Medicine, Oral Pathology and Oral Radiology Published:2025


Abstract

Objectives: Artificial intelligence chatbots have demonstrated feasibility and efficacy in improving health outcomes. In this study, responses from 5 different publicly available AI chatbots—Bing, GPT-3.5, GPT-4, Google Bard, and Claude—to frequently asked questions related to oral cancer were evaluated. Study design: Relevant patient-related frequently asked questions about oral cancer were obtained from two main sources: public health websites and social media platforms. From these sources, 20 oral cancer-related questions were selected. Four board-certified specialists in oral medicine/oral and maxillofacial pathology assessed the answers using modified version of the global quality score on a 5-point Likert scale. Additionally, readability was measured using the Flesch–Kincaid Grade Level and Flesch Reading Ease scores. Responses were also assessed for empathy using a validated 5-point scale. Results: Specialists ranked GPT-4 with highest total score of 17.3 ± 1.5, while Bing received the lowest at 14.9 ± 2.2. Bard had the highest Flesch Reading Ease score of 62 ± 7; and ChatGPT-3.5 and Claude received the lowest scores (more challenging readability). GPT-4 and Bard emerged as the most superior chatbots in terms of empathy and accurate citations on patient-related frequently asked questions pertaining to oral cancer. GPT-4 had highest overall quality, whereas Bing showed the lowest level of quality, empathy, and accuracy for citations. Conclusion: GPT-4 demonstrated the highest quality responses to frequently asked questions pertaining to oral cancer. Although impressive in their ability to guide patients on common oral cancer topics, most chatbots did not perform well when assessed for empathy or citation accuracy. © 2025