Generative commonsense knowledge subgraph retrieval for open-domain dialogue response generation.

Journal: Neural networks : the official journal of the International Neural Network Society
PMID:

Abstract

Grounding on a commonsense knowledge subgraph can help the model generate more informative and diverse dialogue responses. Prior Traverse-based works explicitly retrieve a subgraph from the external knowledge base (eKB). Notably, the available knowledge is strictly restricted by the eKB. To break this restriction, Generative Retrieval methods externalize knowledge from the language model. However, they always generate boring knowledge due to their one-pass externalization procedure. This work proposes a novel TiLM Traverse in Language Model (TiLM), which uses three 'Chain-of-Thought' sub-tasks, i.e., Query Entity Production, Topic Entity Prediction, and Knowledge Subgraph Completion, to build a high-quality knowledge subgraph to ground the next Response Generation without explicitly accessing the eKB in inference. Experimental results on both Chinese and English datasets demonstrate TiLM's outstanding performance even only with a small scale of parameters.

Authors

  • Sixing Wu
    National Pilot School of Software, Yunnan University, Kunming, 650504, Yunnan, China; Engineering Research Center of Cyberspace, Yunnan University, Kunming, 650504, Yunnan, China. Electronic address: wusixing@ynu.edu.cn.
  • Jiong Yu
    The State Key Laboratory for Diagnosis and Treatment of Infectious Diseases, First Affiliated Hospital, College of Medicine, Zhejiang University; Collaborative Innovation Center for Diagnosis and Treatment of Infectious Diseases, Hangzhou, Zhejiang, China (mainland).
  • Jiahao Chen
    The First School of Clinical Medicine, Zhejiang Chinese Medical University, Hangzhou, 310006, China.
  • Wei Zhou
    Department of Eye Function Laboratory, Eye Hospital, China Academy of Chinese Medical Sciences, Beijing, China.