To read this content please select one of the options below:

A memory and attention-based reinforcement learning for musculoskeletal robots with prior knowledge of muscle synergies

Xiaona Wang (Institute of Automation, Chinese Academy of Sciences, Beijing, China and School of Artificial Intelligence, University of the Chinese Academy of Sciences, Beijing, China)
Jiahao Chen (Institute of Automation, Chinese Academy of Sciences, Beijing, China)
Hong Qiao (The State Key Laboratory of Multimodal Artificial Intelligence System, Institute of Automation, Chinese Academy of Sciences, Beijing, China)

Robotic Intelligence and Automation

ISSN: 2754-6969

Article publication date: 15 April 2024

21

Abstract

Purpose

Limited by the types of sensors, the state information available for musculoskeletal robots with highly redundant, nonlinear muscles is often incomplete, which makes the control face a bottleneck problem. The aim of this paper is to design a method to improve the motion performance of musculoskeletal robots in partially observable scenarios, and to leverage the ontology knowledge to enhance the algorithm’s adaptability to musculoskeletal robots that have undergone changes.

Design/methodology/approach

A memory and attention-based reinforcement learning method is proposed for musculoskeletal robots with prior knowledge of muscle synergies. First, to deal with partially observed states available to musculoskeletal robots, a memory and attention-based network architecture is proposed for inferring more sufficient and intrinsic states. Second, inspired by muscle synergy hypothesis in neuroscience, prior knowledge of a musculoskeletal robot’s muscle synergies is embedded in network structure and reward shaping.

Findings

Based on systematic validation, it is found that the proposed method demonstrates superiority over the traditional twin delayed deep deterministic policy gradients (TD3) algorithm. A musculoskeletal robot with highly redundant, nonlinear muscles is adopted to implement goal-directed tasks. In the case of 21-dimensional states, the learning efficiency and accuracy are significantly improved compared with the traditional TD3 algorithm; in the case of 13-dimensional states without velocities and information from the end effector, the traditional TD3 is unable to complete the reaching tasks, while the proposed method breaks through this bottleneck problem.

Originality/value

In this paper, a novel memory and attention-based reinforcement learning method with prior knowledge of muscle synergies is proposed for musculoskeletal robots to deal with partially observable scenarios. Compared with the existing methods, the proposed method effectively improves the performance. Furthermore, this paper promotes the fusion of neuroscience and robotics.

Keywords

Acknowledgements

Funding: This work was supported in part by the Major Project of Science and Technology Innovation 2030 C Brain Science and Brain-Inspired Intelligence under Grant 2021ZD0200408; in part by the National Natural Science Foundation of China under Grant 62203439, Grant 91948303, and Grant 62173326; in part by the Strategic Priority Research Program of Chinese Academy of Sciences under Grant XDB32050100; and in part by the Major Program of the National Natural Science Foundation of China under Grant T2293720, Grant T2293723, and Grant T2293724.

Citation

Wang, X., Chen, J. and Qiao, H. (2024), "A memory and attention-based reinforcement learning for musculoskeletal robots with prior knowledge of muscle synergies", Robotic Intelligence and Automation, Vol. ahead-of-print No. ahead-of-print. https://doi.org/10.1108/RIA-11-2023-0172

Publisher

:

Emerald Publishing Limited

Copyright © 2024, Emerald Publishing Limited

Related articles