Comparing a Large Language Model with Previous Deep Learning Models on Named Entity Recognition of Adverse Drug Events.

Journal: Studies in health technology and informatics
PMID:

Abstract

The ability to fine-tune pre-trained deep learning models to learn how to process a downstream task using a large training set allow to significantly improve performances of named entity recognition. Large language models are recent models based on the Transformers architecture that may be conditioned on a new task with in-context learning, by providing a series of instructions or prompt. These models only require few examples and such approach is defined as few shot learning. Our objective was to compare performances of named entity recognition of adverse drug events between state of the art deep learning models fine-tuned on Pubmed abstracts and a large language model using few-shot learning. Hussain et al's state of the art model (PMID: 34422092) significantly outperformed the ChatGPT-3.5 model (F1-Score: 97.6% vs 86.0%). Few-shot learning is a convenient way to perform named entity recognition when training examples are rare, but performances are still inferior to those of a deep learning model fine-tuned with several training examples. Perspectives are to evaluate few-shot prompting with GPT-4 and perform fine-tuning on GPT-3.5.

Authors

  • Théophile Tiffet
    Unit of Public health, University hospital of Saint-Etienne, France.
  • Alexis Pikaar
    Sorbonne Université, INSERM, Univ Paris 13, Laboratoire d'Informatique Médicale et d'Ingénierie des Connaissances pour la e-Santé, LIMICS, 75006, Paris, France.
  • Béatrice Trombert-Paviot
    Public health and medical information unit, Saint Etienne University Hospital, France.
  • Marie-Christine Jaulent
    INSERM UMRS 1142, Medicine Faculty, Pierre and Marie Curie University, Sorbonne Universities, Paris, France.
  • Cédric Bousquet
    Sorbonne Université, INSERM, Université Paris 13, LIMICS, Paris, France.