TY - GEN
T1 - Distributed Multi-Agent Deep Reinforcement Learning for Robust Coordination against Noise
AU - Motokawa, Yoshinari
AU - Sugawara, Toshiharu
N1 - Funding Information:
This study was partly supported by JSPS KAKENHI Grant Numbers 17KT0044 and 20H04245.
Publisher Copyright:
© 2022 IEEE.
PY - 2022
Y1 - 2022
N2 - In multi-agent systems, noise reduction techniques are considerable for improving the overall system reliability as agents are required to rely on limited environmental information to develop cooperative and coordinated behaviors with the surrounding agents. However, previous studies have often applied centralized noise reduction methods to build robust and versatile coordination in noisy multi-agent environments, while distributed and decentralized autonomous agents are more plausible for real-world application. In this paper, we introduce a distributed attentional actor architecture model for a multi-agent system (DA3-X), using which we demonstrate that agents with DA3-X can selectively learn the noisy environment and behave cooperatively. We experimentally evaluate the effectiveness of DA3-X by comparing learning methods with and without DA3-X and show that agents with DA3-X can achieve better performance than baseline agents. Furthermore, we visualize heatmaps of attentional weights from the DA3-X to analyze how the decision-making process and coordinated behavior are influenced by noise.
AB - In multi-agent systems, noise reduction techniques are considerable for improving the overall system reliability as agents are required to rely on limited environmental information to develop cooperative and coordinated behaviors with the surrounding agents. However, previous studies have often applied centralized noise reduction methods to build robust and versatile coordination in noisy multi-agent environments, while distributed and decentralized autonomous agents are more plausible for real-world application. In this paper, we introduce a distributed attentional actor architecture model for a multi-agent system (DA3-X), using which we demonstrate that agents with DA3-X can selectively learn the noisy environment and behave cooperatively. We experimentally evaluate the effectiveness of DA3-X by comparing learning methods with and without DA3-X and show that agents with DA3-X can achieve better performance than baseline agents. Furthermore, we visualize heatmaps of attentional weights from the DA3-X to analyze how the decision-making process and coordinated behavior are influenced by noise.
KW - Alter-exploration problem
KW - Attention mechanism
KW - Cooperation
KW - Coordination
KW - Distributed system
KW - Multi-agent deep reinforcement learning
KW - Noise reduction
UR - http://www.scopus.com/inward/record.url?scp=85140784231&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85140784231&partnerID=8YFLogxK
U2 - 10.1109/IJCNN55064.2022.9892253
DO - 10.1109/IJCNN55064.2022.9892253
M3 - Conference contribution
AN - SCOPUS:85140784231
T3 - Proceedings of the International Joint Conference on Neural Networks
BT - 2022 International Joint Conference on Neural Networks, IJCNN 2022 - Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2022 International Joint Conference on Neural Networks, IJCNN 2022
Y2 - 18 July 2022 through 23 July 2022
ER -