A Survey on Causal Reinforcement Learning.

Journal: IEEE transactions on neural networks and learning systems
PMID:

Abstract

While reinforcement learning (RL) achieves tremendous success in sequential decision-making problems of many domains, it still faces key challenges of data inefficiency and the lack of interpretability. Interestingly, many researchers have leveraged insights from the causality literature recently, bringing forth flourishing works to unify the merits of causality and address well the challenges from RL. As such, it is of great necessity and significance to collate these causal RL (CRL) works, offer a review of CRL methods, and investigate the potential functionality from causality toward RL. In particular, we divide the existing CRL approaches into two categories according to whether their causality-based information is given in advance or not. We further analyze each category in terms of the formalization of different models, ranging from the Markov decision process (MDP), partially observed MDP (POMDP), multiarmed bandits (MABs), imitation learning (IL), and dynamic treatment regime (DTR). Each of them represents a distinct type of causal graphical illustration. Moreover, we summarize the evaluation matrices and open sources, while we discuss emerging applications, along with promising prospects for the future development of CRL.

Authors

  • Yan Zeng
    Department of Urology, Tongji Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, China.
  • Ruichu Cai
    Faculty of Computer Science, Guangdong University of Technology, Guangzhou, People's Republic of China. Electronic address: cairuichu@gmail.com.
  • Fuchun Sun
    Tsinghua National Laboratory for Information Science and Technology, Department of Computer Science and Technology, Tsinghua University, Beijing, China.
  • Libo Huang
  • Zhifeng Hao
    School of Computer Science and Technology, Guangdong University of Technology, Guangzhou 510006, China. zfhao@gdut.edu.cn.