LBERT: Lexically aware Transformer-based Bidirectional Encoder Representation model for learning universal bio-entity relations.

Journal: Bioinformatics (Oxford, England)
Published Date:

Abstract

MOTIVATION: Natural Language Processing techniques are constantly being advanced to accommodate the influx of data as well as to provide exhaustive and structured knowledge dissemination. Within the biomedical domain, relation detection between bio-entities known as the Bio-Entity Relation Extraction (BRE) task has a critical function in knowledge structuring. Although recent advances in deep learning-based biomedical domain embedding have improved BRE predictive analytics, these works are often task selective or use external knowledge-based pre-/post-processing. In addition, deep learning-based models do not account for local syntactic contexts, which have improved data representation in many kernel classifier-based models. In this study, we propose a universal BRE model, i.e. LBERT, which is a Lexically aware Transformer-based Bidirectional Encoder Representation model, and which explores both local and global contexts representations for sentence-level classification tasks.

Authors

  • Neha Warikoo
    Institute of Biomedical Informatics, National Yang-Ming University, Taipei, Taiwan.
  • Yung-Chun Chang
    Graduate Institute of Data Science, College of Management, Taipei Medical University, Taipei, Taiwan.
  • Wen-Lian Hsu
    Institute of Information Science, Academia Sinica, Taiwan.