Abstract

Introduction The introduction of Artificial Intelligence (AI) tools like ChatGPT and Google Bard promises transformative advances in clinical diagnostics. The aim of this study is to examine the ability of these two AI tools to diagnose various medical scenarios. Methods Experts from varied medical domains curated 20 case scenarios, each paired with its ideal diagnostic answer. Both AI systems, ChatGPT (updated in September 2021) and Google Bard (updated in January 2023), were tasked with diagnosing these cases. Their outcomes were recorded and subsequently assessed by human medical professionals. Results In the diagnostic evaluations, ChatGPT achieved an accuracy of 90%, correctly diagnosing 18 out of 20 cases, while Google Bard displayed an 80% accuracy rate, correctly answering 16 questions. Notably, both AIs faltered in specific complex scenarios. For instance, both systems misdiagnosed a labor situation, and while ChatGPT incorrectly identified a case of hypertrophic pyloric stenosis, Google Bard suggested a less suitable diagnostic procedure (pelvic ultrasound) for a 56-year-old patient. Conclusion This study showcases the promising capabilities of ChatGPT and Google Bard in the realm of clinical diagnostics, with both AI tools achieving commendable accuracy rates.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call