End-to-End Autonomous Exploration with Deep Reinforcement Learning and Intrinsic Motivation.

Journal: Computational intelligence and neuroscience
PMID:

Abstract

Developing artificial intelligence (AI) agents is challenging for efficient exploration in visually rich and complex environments. In this study, we formulate the exploration question as a reinforcement learning problem and rely on intrinsic motivation to guide exploration behavior. Such intrinsic motivation is driven by curiosity and is calculated based on episode memory. To distribute the intrinsic motivation, we use a count-based method and temporal distance to generate it synchronously. We tested our approach in 3D maze-like environments and validated its performance in exploration tasks through extensive experiments. The experimental results show that our agent can learn exploration ability from raw sensory input and accomplish autonomous exploration across different mazes. In addition, the learned policy is not biased by stochastic objects. We also analyze the effects of different training methods and driving forces on exploration policy.

Authors

  • Xiaogang Ruan
    Faculty of Information Technology, Beijing University of Technology, Beijing, China.
  • Peng Li
    WuXi AppTec Co, Shanghai, China.
  • Xiaoqing Zhu
    Faculty of Information Technology, Beijing University of Technology, Beijing, China.
  • Hejie Yu
    Faculty of Information Technology, Beijing University of Technology, Beijing, China.
  • Naigong Yu
    Faculty of Information Technology, Beijing University of Technology, Beijing, China.