Cooperative Multi-Robot Hierarchical Reinforcement Learning

Gembong Edhi Setyawan, Pitoyo Hartono, Hideyuki Sawada

Research output: Contribution to journalArticlepeer-review


Recent advances in multi-robot deep reinforcement learning have made it possible to perform efficient exploration in problem space, but it remains a significant challenge in many complex domains. To alleviate this problem, a hierarchical approach has been designed in which agents can operate at many levels to complete tasks more efficiently. This paper proposes a novel technique called Multi-Agent Hierarchical Deep Deterministic Policy Gradient that combines the benefits of multiple robot systems with the hierarchical system used in Deep Reinforcement Learning. Here, agents acquire the ability to decompose a problem into simpler subproblems with varying time scales. Furthermore, this study develops a framework to formulate tasks into multiple levels. The upper levels function to learn policies for defining lower levels’ subgoals, whereas the lowest level depicts robot’s learning policies for primitive actions in the real environment. The proposed method is implemented and validated in a modified Multiple Particle Environment (MPE) scenario.

Original languageEnglish
Pages (from-to)35-44
Number of pages10
JournalInternational Journal of Advanced Computer Science and Applications
Issue number9
Publication statusPublished - 2022


  • Hierarchical deep reinforcement learning
  • Multi-robot system
  • Path-finding
  • Task decomposition

ASJC Scopus subject areas

  • Computer Science(all)


Dive into the research topics of 'Cooperative Multi-Robot Hierarchical Reinforcement Learning'. Together they form a unique fingerprint.

Cite this