Authors
Jacopo Castellini, Frans A Oliehoek, Rahul Savani, Shimon Whiteson
Publication date
2019/5/8
Conference
Proceedings of the 18th International Conference on Autonomous Agents and MultiAgent Systems
Pages
1862-1864
Publisher
International Foundation for Autonomous Agents and Multiagent Systems
Description
Recent years have seen the application of deep reinforcement learning techniques to cooperative multi-agent systems, with great empirical success. However, given the lack of theoretical insight, it remains unclear what the employed neural networks are learning, or how we should enhance their representational power to address the problems on which they fail. In this work, we empirically investigate the representational power of various network architectures on a series of one-shot games. Despite their simplicity, these games capture many of the crucial problems that arise in the multi-agent setting, such as an exponential number of joint actions or the lack of an explicit coordination mechanism. Our results quantify how well various approaches can represent the requisite value functions, and help us identify issues that can impede good performance.
Total citations
20182019202020212022202320241111151864
Scholar articles
J Castellini, FA Oliehoek, R Savani, S Whiteson - arXiv preprint arXiv:1902.07497, 2019