TY - JOUR
T1 - Cooperative Multi-Robot Hierarchical Reinforcement Learning
AU - Setyawan, Gembong Edhi
AU - Hartono, Pitoyo
AU - Sawada, Hideyuki
N1 - Funding Information:
This work was supported by JSPS Grants-in-Aid for Scientific Research on Innovative Areas (Research in a proposed research area) 18H05473 and 18H05895.
Publisher Copyright:
© 2022,International Journal of Advanced Computer Science and Applications. All Rights Reserved.
PY - 2022
Y1 - 2022
N2 - Recent advances in multi-robot deep reinforcement learning have made it possible to perform efficient exploration in problem space, but it remains a significant challenge in many complex domains. To alleviate this problem, a hierarchical approach has been designed in which agents can operate at many levels to complete tasks more efficiently. This paper proposes a novel technique called Multi-Agent Hierarchical Deep Deterministic Policy Gradient that combines the benefits of multiple robot systems with the hierarchical system used in Deep Reinforcement Learning. Here, agents acquire the ability to decompose a problem into simpler subproblems with varying time scales. Furthermore, this study develops a framework to formulate tasks into multiple levels. The upper levels function to learn policies for defining lower levels’ subgoals, whereas the lowest level depicts robot’s learning policies for primitive actions in the real environment. The proposed method is implemented and validated in a modified Multiple Particle Environment (MPE) scenario.
AB - Recent advances in multi-robot deep reinforcement learning have made it possible to perform efficient exploration in problem space, but it remains a significant challenge in many complex domains. To alleviate this problem, a hierarchical approach has been designed in which agents can operate at many levels to complete tasks more efficiently. This paper proposes a novel technique called Multi-Agent Hierarchical Deep Deterministic Policy Gradient that combines the benefits of multiple robot systems with the hierarchical system used in Deep Reinforcement Learning. Here, agents acquire the ability to decompose a problem into simpler subproblems with varying time scales. Furthermore, this study develops a framework to formulate tasks into multiple levels. The upper levels function to learn policies for defining lower levels’ subgoals, whereas the lowest level depicts robot’s learning policies for primitive actions in the real environment. The proposed method is implemented and validated in a modified Multiple Particle Environment (MPE) scenario.
KW - Hierarchical deep reinforcement learning
KW - Multi-robot system
KW - Path-finding
KW - Task decomposition
UR - http://www.scopus.com/inward/record.url?scp=85139306773&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85139306773&partnerID=8YFLogxK
U2 - 10.14569/IJACSA.2022.0130904
DO - 10.14569/IJACSA.2022.0130904
M3 - Article
AN - SCOPUS:85139306773
VL - 13
SP - 35
EP - 44
JO - International Journal of Advanced Computer Science and Applications
JF - International Journal of Advanced Computer Science and Applications
SN - 2158-107X
IS - 9
ER -