AI versus human-generated multiple-choice questions for medical education: a cohort study in a high-stakes examination.
Journal:
BMC medical education
PMID:
39923067
Abstract
BACKGROUND: The creation of high-quality multiple-choice questions (MCQs) is essential for medical education assessments but is resource-intensive and time-consuming when done by human experts. Large language models (LLMs) like ChatGPT-4o offer a promising alternative, but their efficacy remains unclear, particularly in high-stakes exams.