|
SPECIAL ISSUE: ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING TECHNOLOGIES
Hierarchical method for cooperative multi-agent reinforcement learning in Markov decision processes
V. È. Bol'shakov, A. N. Alfimtsev Bauman Moscow State Technical University, Moscow, Russia
Abstract:
In the rapidly evolving field of reinforcement learning, combination of hierarchical and multi-agent learning methods presents unique challenges and opens up new opportunities. This paper discusses a combination of multi-level hierarchical learning with subgoal discovery and multi-agent reinforcement learning with hindsight experience replay. Combining these approaches leads to the creation of Multi-Agent Subgoal Hierarchy Algorithm (MASHA) that allows multiple agents to learn efficiently in complex environments, including environments with sparse rewards. We demonstrate the results of the proposed approach in one of these environments inside the StarCraft II strategy game, in addition to making comparisons with other existing approaches. The proposed algorithm is developed in the paradigm of centralized learning with decentralized execution, which makes it possible to achieve a balance between coordination and autonomy of agents.
Keywords:
multi-agent reinforcement learning, hierarchical learning, subgoal discovery, hindsight experience replay, centralized learning with decentralized execution, sparse rewards.
Citation:
V. È. Bol'shakov, A. N. Alfimtsev, “Hierarchical method for cooperative multi-agent reinforcement learning in Markov decision processes”, Dokl. RAN. Math. Inf. Proc. Upr., 514:2 (2023), 250–261; Dokl. Math., 108:suppl. 2 (2023), S382–S392
Linking options:
https://www.mathnet.ru/eng/danma470 https://www.mathnet.ru/eng/danma/v514/i2/p250
|
Statistics & downloads: |
Abstract page: | 48 | References: | 13 |
|