TY - JOUR
T1 - Imitation-Enhanced Reinforcement Learning With Privileged Smooth Transition for Hexapod Locomotion
AU - Zhang, Zhelin
AU - Liu, Tie
AU - Ding, Liang
AU - Wang, Haoyu
AU - Xu, Peng
AU - Yang, Huaiguang
AU - Gao, Haibo
AU - Deng, Zongquan
AU - Pajarinen, Joni
N1 - Publisher Copyright:
© 2016 IEEE.
PY - 2025/1
Y1 - 2025/1
N2 - Deep reinforcement learning (DRL) methods have shown significant promise in controlling the movement of quadruped robots. However, for systems like hexapod robots, which feature a higher-dimensional action space, it remains challenging for an agent to devise an effective control strategy directly. Currently, no hexapod robots have demonstrated highly dynamic motion. To address this, we propose imitation-enhanced reinforcement learning (IERL), a two-stage approach enabling hexapod robots to achieve dynamic motion through direct control using RL methods. Initially, imitation learning (IL) replicates a basic positional control method, creating a pre-trained policy for basic locomotion. Subsequently, the parameters from this model are utilized as the starting point for the reinforcement learning process to train the agent. Moreover, we incorporate a smooth transition (ST) method to make IERL overcome the changes in network inputs between two stages, and adaptable to various complex network architectures incorporating latent features. Extensive simulations and real-world experiments confirm that our method effectively tackles the high-dimensional action space challenges of hexapod robots, significantly enhancing learning efficiency and enabling more natural, efficient, and dynamic movements compared to existing methods.
AB - Deep reinforcement learning (DRL) methods have shown significant promise in controlling the movement of quadruped robots. However, for systems like hexapod robots, which feature a higher-dimensional action space, it remains challenging for an agent to devise an effective control strategy directly. Currently, no hexapod robots have demonstrated highly dynamic motion. To address this, we propose imitation-enhanced reinforcement learning (IERL), a two-stage approach enabling hexapod robots to achieve dynamic motion through direct control using RL methods. Initially, imitation learning (IL) replicates a basic positional control method, creating a pre-trained policy for basic locomotion. Subsequently, the parameters from this model are utilized as the starting point for the reinforcement learning process to train the agent. Moreover, we incorporate a smooth transition (ST) method to make IERL overcome the changes in network inputs between two stages, and adaptable to various complex network architectures incorporating latent features. Extensive simulations and real-world experiments confirm that our method effectively tackles the high-dimensional action space challenges of hexapod robots, significantly enhancing learning efficiency and enabling more natural, efficient, and dynamic movements compared to existing methods.
KW - Hexapod robot
KW - Imitation learning
KW - Locomotion control
KW - Reinforcement learning
UR - http://www.scopus.com/inward/record.url?scp=85209563718&partnerID=8YFLogxK
U2 - 10.1109/LRA.2024.3497754
DO - 10.1109/LRA.2024.3497754
M3 - Article
AN - SCOPUS:85209563718
SN - 2377-3766
VL - 10
SP - 350
EP - 357
JO - IEEE Robotics and Automation Letters
JF - IEEE Robotics and Automation Letters
IS - 1
ER -