Abstract
Artificial intelligence (AI)-driven technologies offer transformative potential in plastic surgery, spanning pre-operative planning, surgical procedures, and post-operative care, with the promise of improved patient outcomes. To compare the web-based ChatGPT-4o (omni; OpenAI, San Francisco, CA) and Gemini Advanced (Alphabet Inc., Mountain View, CA), focusing on their data upload feature and examining outcomes before and after exposure to CME articles, particularly regarding their efficacy relative to human participants. Participants and LLMs completed 22 multiple-choice questions to assess baseline knowledge of CME topics. Initially, both LLMs and participants answered without article access. In incognito mode, the LLMs repeated the tests over 6 days. After accessing the articles, responses from both LLMs and participants were extracted and analyzed. There was a significant increase in mean scores after the article was read in the resident group, indicating a significant rise. In the LLM groups, the ChatGPT-4.o (omni) group showed no significant difference between pre- and post-article scores, but the Gemini Advanced group demonstrated a significant increase. It can be stated that the ChatGPT-4.o and Gemini Advanced groups have higher accuracy means compared to the resident group in both pre and post-article periods. The analysis between human participants and LLMs indicates promising implications for the incorporation of LLMs in medical education. As these models increase in sophistication, they offer the potential to serve as supplementary tools within traditional learning environments. This could aid in bridging the gap between theoretical knowledge and practical implementation.
Published Version
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have