Learning and Reusing Quadruped Robot Movement Skills from Biological Dogs for Higher-Level Tasks.

Journal: Sensors (Basel, Switzerland)
Published Date:

Abstract

In the field of quadruped robots, the most classic motion control algorithm is based on model prediction control (MPC). However, this method poses challenges as it necessitates the precise construction of the robot's dynamics model, making it difficult to achieve agile movements similar to those of a biological dog. Due to these limitations, researchers are increasingly turning to model-free learning methods, which significantly reduce the difficulty of modeling and engineering debugging and simultaneously reduce real-time optimization computational burden. Inspired by the growth process of humans and animals, from learning to walk to fluent movements, this article proposes a hierarchical reinforcement learning framework for the motion controller to learn some higher-level tasks. First, some basic motion skills can be learned from motion data captured from a biological dog. Then, with these learned basic motion skills as a foundation, the quadruped robot can focus on learning higher-level tasks without starting from low-level kinematics, which saves redundant training time. By utilizing domain randomization techniques during the training process, the trained policy function can be directly transferred to a physical robot without modification, and the resulting controller can perform more biomimetic movements. By implementing the method proposed in this article, the agility and adaptability of the quadruped robot can be maximally utilized to achieve efficient operations in complex terrains.

Authors

  • Qifeng Wan
    State Key Laboratory for Turbulence and Complex Systems, Department of Advanced Manufacturing and Robotics, College of Engineering, Peking University, Beijing 100871, China.
  • Aocheng Luo
    State Key Laboratory for Turbulence and Complex Systems, Department of Advanced Manufacturing and Robotics, College of Engineering, Peking University, Beijing 100871, China.
  • Yan Meng
    School of Electronic and Information Engineering, Beijing Jiaotong University, Beijing 100044, China.
  • Chong Zhang
    Department of Big Data Management and Application, School of International Economics and Management, Beijing Technology and Business University, Beijing 100048, China.
  • Wanchao Chi
    Tencent Robotics X, Shenzhen 518057, China.
  • Shenghao Zhang
    Tencent Robotics X, Shenzhen 518057, China.
  • Yuzhen Liu
    The Department of Obstetrics and Gynecology, Affiliated Hospital of Weifang Medical University, Weifang 261042, People's Republic of China.
  • Qiuguo Zhu
    Institute of Cyber-Systems and Control, College of Control Science and Engineering, Zhejiang University, Hangzhou 310027, China.
  • Shihan Kong
    Department of Advanced Manufacturing and Robotics, College of Engineering, Peking University, Beijing 100871, China.
  • Junzhi Yu
    Institute of Automation, Chinese Academy of Sciences, Beijing 100190, PR China. Electronic address: junzhi.yu@ia.ac.cn.