Authors
Junling Hu, Michael P Wellman
Publication date
1998/7/24
Journal
ICML
Volume
98
Pages
242-250
Description
In this paper, we adopt general-sum stochastic games as a framework for multiagent reinforcement learning. Our work extends previous work by Littman on zero-sum stochastic games to a broader framework. We design a multiagent Q-learning method under this framework, and prove that it converges to a Nash equilibrium under specified conditions. This algorithm is useful for finding the optimal strategy when there exists a unique Nash equilibrium in the game. When there exist multiple Nash equilibria in the game, this algorithm should be combined with other learning techniques to find optimal strategies.
Total citations
199920002001200220032004200520062007200820092010201120122013201420152016201720182019202020212022202320241225305082587356704248423632423729182331677175655225