Evaluating accuracy and reproducibility of large language model performance on critical care assessments in pharmacy education.
Journal:
Frontiers in artificial intelligence
Published Date:
Jan 9, 2025
Abstract
BACKGROUND: Large language models (LLMs) have demonstrated impressive performance on medical licensing and diagnosis-related exams. However, comparative evaluations to optimize LLM performance and ability in the domain of comprehensive medication management (CMM) are lacking. The purpose of this evaluation was to test various LLMs performance optimization strategies and performance on critical care pharmacotherapy questions used in the assessment of Doctor of Pharmacy students.
Authors
Keywords
No keywords available for this article.