Planning and acting in partially observable stochastic domains
Artificial Intelligence
On the difficulty of achieving equilibrium in interactive POMDPs
AAAI'06 proceedings of the 21st national conference on Artificial intelligence - Volume 2
Point-based dynamic programming for DEC-POMDPs
AAAI'06 proceedings of the 21st national conference on Artificial intelligence - Volume 2
A particle filtering based approach to approximating interactive POMDPs
AAAI'05 Proceedings of the 20th national conference on Artificial intelligence - Volume 2
A framework for sequential planning in multi-agent settings
Journal of Artificial Intelligence Research
Perseus: randomized point-based value iteration for POMDPs
Journal of Artificial Intelligence Research
Anytime point-based approximations for large POMDPs
Journal of Artificial Intelligence Research
Improving anytime point-based value iteration using principled point selections
IJCAI'07 Proceedings of the 20th international joint conference on Artifical intelligence
Taming decentralized POMDPs: towards efficient policy computation for multiagent settings
IJCAI'03 Proceedings of the 18th international joint conference on Artificial intelligence
Monte Carlo sampling methods for approximating interactive POMDPs
Journal of Artificial Intelligence Research
A partition-based first-order probabilistic logic to represent interactive beliefs
SUM'11 Proceedings of the 5th international conference on Scalable uncertainty management
Generalized and bounded policy iteration for finitely-nested interactive POMDPs: scaling up
Proceedings of the 11th International Conference on Autonomous Agents and Multiagent Systems - Volume 2
IJCAI'13 Proceedings of the Twenty-Third international joint conference on Artificial Intelligence
Hi-index | 0.00 |
We develop a point based method for solving finitely nested interactive POMDPs approximately. Analogously to point based value iteration (PBVI) in POMDPs, we maintain a set of belief points and form value functions composed of those value vectors that are optimal at these points. However, as we focus on muItiagent settings, the beliefs are nested and computation of the value vectors relies on predicted actions of others. Consequently, we develop a novel interactive gen eralization of PBVI applicable to muItiagent settings.