A Review of Cooperative Multi-Agent Deep Reinforcement Learning
release_s2umqzxmqrhntkev3f6k554cv4
by
Afshin OroojlooyJadid, Davood Hajinezhad
2021
Abstract
Deep Reinforcement Learning has made significant progress in multi-agent
systems in recent years. In this review article, we have focused on presenting
recent approaches on Multi-Agent Reinforcement Learning (MARL) algorithms. In
particular, we have focused on five common approaches on modeling and solving
cooperative multi-agent reinforcement learning problems: (I) independent
learners, (II) fully observable critic, (III) value function factorization,
(IV) consensus, and (IV) learn to communicate. First, we elaborate on each of
these methods, possible challenges, and how these challenges were mitigated in
the relevant papers. If applicable, we further make a connection among
different papers in each category. Next, we cover some new emerging research
areas in MARL along with the relevant recent papers. Due to the recent success
of MARL in real-world applications, we assign a section to provide a review of
these applications and corresponding articles.
Also, a list of available environments for MARL research is provided in this
survey. Finally, the paper is concluded with proposals on the possible research
directions.
In text/plain
format
Archived Files and Locations
application/pdf 836.8 kB
file_g2deg6c6yzb7lnrjgckl7vnrga
|
arxiv.org (repository) web.archive.org (webarchive) |
1908.03963v4
access all versions, variants, and formats of this works (eg, pre-prints)