Dynamic programming for partially observable stochastic games
AAAI'04 Proceedings of the 19th national conference on Artifical intelligence
A framework for sequential planning in multi-agent settings
Journal of Artificial Intelligence Research
Generalized point based value iteration for interactive POMDPs
AAAI'08 Proceedings of the 23rd national conference on Artificial intelligence - Volume 1
Representing Bayesian games without a common prior
Proceedings of the 9th International Conference on Autonomous Agents and Multiagent Systems: volume 1 - Volume 1
Model identification in interactive influence diagrams using mutual information
Web Intelligence and Agent Systems
QueryPOMDP: POMDP-based communication in multiagent systems
EUMAS'11 Proceedings of the 9th European conference on Multi-Agent Systems
Hi-index | 0.01 |
We analyze the asymptotic behavior of agents engaged in an infinite horizon partially observable stochastic game as formalized by the interactive POMDP framework. We show that when agents' initial beliefs satisfy a truth compatibility condition, their behavior converges to a subjective Ε-equilibrium in a finite time, and subjective equilibrium in the limit. This result is a generalization of a similar result in repeated games, to partially observable stochastic games. However, it turns out that the equilibrating process is difficult to demonstrate computationally because of the difficulty in coming up with initial beliefs that are both natural and satisfy the truth compatibility condition. Our results, therefore, shed some negative light on using equilibria as a solution concept for decision making in partially observable stochastic games.