Chatbots Vs. Human Experts: Evaluating Diagnostic Performance of Chatbots in Uveitis and the Perspectives on AI Adoption in Ophthalmology

William Rojas-Carabali, Alok Sen, Aniruddha Agarwal, Gavin Tan, Carol Y Cheung, Andres Rousselot, Rajdeep Agrawal, Renee Liu, Carlos Cifuentes-González, Tobias Elze, John H Kempen, Lucia Sobrin, Quan Dong Nguyen, Alejandra de-la-Torre, Bernett Lee, Vishali Gupta, Rupesh Agrawal*

*Corresponding author for this work

Research output: Contribution to journalArticleAcademicpeer-review

Abstract

PURPOSE: To assess the diagnostic performance of two chatbots, ChatGPT and Glass, in uveitis diagnosis compared to renowned uveitis specialists, and evaluate clinicians' perception about utilizing artificial intelligence (AI) in ophthalmology practice. METHODS: Six cases were presented to uveitis experts, ChatGPT (version 3.5 and 4.0) and Glass 1.0, and diagnostic accuracy was analyzed. Additionally, a survey about the emotions, confidence in utilizing AI-based tools, and the likelihood of incorporating such tools in clinical practice was done. RESULTS: Uveitis experts accurately diagnosed all cases (100%), while ChatGPT achieved a diagnostic success rate of 66% and Glass 1.0 achieved 33%. Most attendees felt excited or optimistic about utilizing AI in ophthalmology practice. Older age and high level of education were positively correlated with increased inclination to adopt AI-based tools. CONCLUSIONS: ChatGPT demonstrated promising diagnostic capabilities in uveitis cases and ophthalmologist showed enthusiasm for the integration of AI into clinical practice.
Original languageEnglish
Pages (from-to)1-8
Number of pages8
JournalOcular Immunology and Inflammation
DOIs
Publication statusE-pub ahead of print - Oct 2023

Keywords

  • Artificial intelligence
  • ChatGPT
  • diagnosis
  • large language model
  • ophthalmology

Cite this