Authors
Adrian Šošić, Wasiur R KhudaBukhsh, Abdelhak M Zoubir, Heinz Koeppl
Publication date
2017
Conference
16th Conference on Autonomous Agents and Multi-Agent Systems (AAMAS)
Pages
1413–1421
Publisher
https://dl.acm.org/doi/10.5555/3091125.3091320
Description
Inverse reinforcement learning (IRL) has become a useful tool for learning behavioral models from demonstration data. However, IRL remains mostly unexplored for multi-agent systems. In this paper, we show how the principle of IRL can be extended to homogeneous large-scale problems, inspired by the collective swarming behavior of natural systems. In particular, we make the following contributions to the field: 1) We introduce the swarMDP framework, a sub-class of decentralized partially observable Markov decision processes endowed with a swarm characterization. 2) Exploiting the inherent homogeneity of this framework, we reduce the resulting multi-agent IRL problem to a single-agent one by proving that the agent-specific value functions in this model coincide. 3) To solve the corresponding control problem, we propose a novel heterogeneous learning scheme that is particularly tailored to the swarm setting. Results on two example systems demonstrate that our framework is able to produce meaningful local reward models from which we can replicate the observed global system dynamics.
Total citations
201620172018201920202021202220232024231312111110218
Scholar articles
A Šošić, WR KhudaBukhsh, AM Zoubir, H Koeppl - arXiv preprint arXiv:1602.05450, 2016
A Šošic, WR KhudaBukhsh, AM Zoubir, H Koeppl - arXiv preprint arXiv:1602.05450, 2016