Enhancing Chinese Character Representation With Lattice-Aligned Attention.

Journal: IEEE transactions on neural networks and learning systems
Published Date:

Abstract

Word-character lattice models have been proved to be effective for some Chinese natural language processing (NLP) tasks, in which word boundary information is fused into character sequences. However, due to the inherently unidirectional sequential nature, prior approaches have only learned sequential interactions of character-word instances but fail to capture fine-grained correlations in word-character spaces. In this article, we propose a lattice-aligned attention network (LAN) that aims to model dense interactions over word-character lattice structure for enhancing character representations. By carefully combining cross-lattice module, gated word-character semantic fusion unit, and self-lattice attention module, the network can explicitly capture fine-grained correlations across different spaces (e.g., word-to-character and character-to-character), thus significantly improving model performance. Experimental results on three Chinese NLP benchmark tasks demonstrate that LAN obtains state-of-the-art results compared to several competitive approaches.

Authors

  • Shan Zhao
    Department of Mathematics, University of Alabama, Tuscaloosa, AL 35487-0350, USA.
  • Minghao Hu
  • Zhiping Cai
    College of Computer, National University of Defense Technology, Changsha, China. Electronic address: zpcai@nudt.edu.cn.
  • Zhanjun Zhang
    State Key Laboratory of Cognitive Neuroscience and Learning & IDG/McGovern Institute for Brain Research, Beijing Normal University, Beijing 100875, China.
  • Tongqing Zhou
  • Fang Liu
    The First Clinical Medical College of Gannan Medical University, Ganzhou 341000, Jiangxi Province, China.