Comparative performance of ChatGPT, Gemini, and final-year emergency medicine clerkship students in answering multiple-choice questions: implications for the use of AI in medical education.

Journal: International journal of emergency medicine
Published Date:

Abstract

BACKGROUND: The integration of artificial intelligence (AI) into medical education has gained significant attention, particularly with the emergence of advanced language models, such as ChatGPT and Gemini. While these tools show promise for answering multiple-choice questions (MCQs), their efficacy in specialized domains, such as Emergency Medicine (EM) clerkship, remains underexplored. This study aimed to evaluate and compare the accuracy of ChatGPT, Gemini, and final-year EM students when it comes to answering text-only and image-based MCQs, in order to assess AI's potential for use as a supplementary tool in the field of medical education.

Authors

  • Shaikha Nasser Al-Thani
    Department of Emergency Medicine, Hamad Medical Corporation, P O Box 3050, Doha, Qatar.
  • Shahzad Anjum
    Department of Emergency Medicine, Hamad Medical Corporation, P O Box 3050, Doha, Qatar.
  • Zain Ali Bhutta
    Department of Emergency Medicine, Hamad Medical Corporation, P O Box 3050, Doha, Qatar.
  • Sarah Bashir
    University of Aberdeen, Aberdeenshire, UK.
  • Muhammad Azhar Majeed
    Department of Emergency Medicine, Hamad Medical Corporation, P O Box 3050, Doha, Qatar.
  • Anfal Sher Khan
    Weill Cornell Medicine, Doha, Qatar.
  • Khalid Bashir
    Department of Emergency Medicine, Hamad Medical Corporation, P O Box 3050, Doha, Qatar. kbashir@hamad.qa.

Keywords

No keywords available for this article.