Oversampling effect in pretraining for bidirectional encoder representations from transformers (BERT) to localize medical BERT and enhance biomedical BERT.

Journal: Artificial intelligence in medicine
Published Date:

Abstract

BACKGROUND: Pretraining large-scale neural language models on raw texts has made a significant contribution to improving transfer learning in natural language processing. With the introduction of transformer-based language models, such as bidirectional encoder representations from transformers (BERT), the performance of information extraction from free text has improved significantly in both the general and medical domains. However, it is difficult to train specific BERT models to perform well in domains for which few databases of a high quality and large size are publicly available.

Authors

  • Shoya Wada
    Department of Medical Informatics, Osaka University Graduate School of Medicine, Suita, Osaka, Japan.
  • Toshihiro Takeda
    Department of Medical Informatics, Osaka University Graduate School of Medicine, Suita, Osaka, Japan.
  • Katsuki Okada
    Department of Medical Informatics, Osaka University Graduate School of Medicine, Osaka, Japan.
  • Shirou Manabe
    Department of Medical Informatics, Osaka University Graduate School of Medicine, Osaka, Japan.
  • Shozo Konishi
    Department of Medical Informatics, Osaka University Graduate School of Medicine.
  • Jun Kamohara
    Faculty of Medicine, Osaka University, Japan.
  • Yasushi Matsumura
    Department of Medical Informatics, Osaka University Graduate School of Medicine, Suita, Osaka, Japan.