Retrieval In Decoder benefits generative models for explainable complex question answering.

Journal: Neural networks : the official journal of the International Neural Network Society
Published Date:

Abstract

Large-scale Language Models (LLMs) utilizing the Chain-of-Thought prompting demonstrate exceptional performance in a variety of tasks. However, the persistence of factual hallucinations remains a significant challenge in practical applications. Prevailing retrieval-augmented methods treat the retriever and generator as separate components, which inadvertently restricts the generator's capabilities to those of the retriever through intensive supervised training. In this work, we propose an unsupervised Retrieval In Decoder framework for multi-granularity decoding called RID, which integrates retrieval directly into the decoding process of generative models. It dynamically adjusts decoding granularity based on retrieval outcomes, and duly corrects the decoding direction through its direct impact on the next token. Moreover, we introduce a reinforcement learning-driven knowledge distillation method for adaptive explanation generation to better apply to Small-scale Language Models (SLMs). The experimental results across six public benchmarks surpass popular LLMs and existing retrieval-augmented methods, which demonstrates the effectiveness of RID in models of different scales and verifies its applicability and scalability.

Authors

  • JianZhou Feng
    Yanshan Univ, Sch Informat Sci & Engn,Software Engn Key Lab Hebei Prov, Qinhuangdao, Hebei, China.
  • Qin Wang
    Department of Pharmacy, Affiliated Hospital of Nantong University, Nantong, China.
  • Huaxiao Qiu
    School of Information Science and Engineering, Yanshan University, Qinhuangdao, 066004, China. Electronic address: qiuhuaxiao@stumail.ysu.edu.cn.
  • Lirong Liu
    School of Science. Yanshan University, Qinhuangdao, 066004, China. Electronic address: 1310103683@stumail.ysu.edu.cn.