A Bi-level representation learning model for medical visual question answering.

Journal: Journal of biomedical informatics
Published Date:

Abstract

Medical Visual Question Answering (VQA) targets at answering questions related to given medical images and it contains tremendous potential in healthcare services. However, researches on medical VQA are still facing challenges, particularly on how to learn a fine-grained multimodal semantic representation from relatively small volume of data resources for answer prediction. Moreover, the long-tailed distribution labels of medical VQA data frequently result in poor performance of models. To this end, we propose a novel bi-level representation learning model with two reasoning modules to learn bi-level representations for the medical VQA task. One is sentence-level reasoning to learn sentence-level semantic representations from multimodal input. The other is token-level reasoning that employs an attention mechanism to generate a multimodal contextual vector by fusing image features and word embeddings. The contextual vector is used to filter irrelevant semantic representations from sentence-level reasoning to generate a fine-grained multimodal representation. Furthermore, a label-distribution-smooth margin loss is proposed to minimize generalization error bound of long-tailed distribution datasets by modifying margin bound of different labels in training set. Based on standard VQA-Rad dataset and PathVQA dataset, the proposed model achieves 0.7605 and 0.5434 on accuracy, 0.7741 and 0.5288 on F1-score, respectively, outperforming a set of state-of-the-art baseline models.

Authors

  • Yong Li
    Department of Surgical Sciences, Western Michigan University Homer Stryker M.D. School of Medicine, Kalamazoo, MI, United States.
  • Shaopei Long
    School of Computer Science, South China Normal University, Guangzhou, China. Electronic address: 1030829350@qq.com.
  • Zhenguo Yang
    Department of Computer Science, Guangdong University of Technology, Guangzhou, China; Department of Computer Science, City University of Hong Kong, Hong Kong, China. Electronic address: zhengyang5-c@my.cityu.edu.hk.
  • Heng Weng
    Department of Big Medical Data, Health Construction Administration Center, The Second Affiliated Hospital of Guangzhou University of Chinese Medicine, Guangzhou, China. ww128@qq.com.
  • Kun Zeng
    College of Physical Science and Technology, Department of Electronic Science, Fujian Provincial Key Laboratory of Plasma and Magnetic Resonance, Xiamen University, Xiamen, 361005, China.
  • Zhenhua Huang
    School of Computer Science, South China Normal University, Guangzhou, China. Electronic address: huangzhenhua@m.scnu.edu.cn.
  • Fu Lee Wang
    School of Science and Technology, Hong Kong Metropolitan University, Hong Kong SAR, China. Electronic address: pwang@hkmu.edu.hk.
  • Tianyong Hao
    School of Information Science and Technology, Guangdong University of Foreign Studies, Guangzhou, China. haoty@gdufs.edu.cn.