The Complexity of Decentralized Control of Markov Decision Processes
Mathematics of Operations Research
Point-based dynamic programming for DEC-POMDPs
AAAI'06 proceedings of the 21st national conference on Artificial intelligence - Volume 2
Memory-bounded dynamic programming for DEC-POMDPs
IJCAI'07 Proceedings of the 20th international joint conference on Artifical intelligence
Taming decentralized POMDPs: towards efficient policy computation for multiagent settings
IJCAI'03 Proceedings of the 18th international joint conference on Artificial intelligence
Bounded policy iteration for decentralized POMDPs
IJCAI'05 Proceedings of the 19th international joint conference on Artificial intelligence
Exploiting symmetries for single- and multi-agent Partially Observable Stochastic Domains
Artificial Intelligence
Hi-index | 0.01 |
We present a memory-bounded approximate algorithm for solving infinite-horizon decentralized partially observable Markov decision processes (DEC-POMDPs). In particular, we improve upon the bounded policy iteration (BPI) approach, which searches for a locally optimal stochastic finite state controller, by accompanying reachability analysis on controller nodes. As a result, the algorithm has different optimization criteria for the reachable and the unreachable nodes, and it is more effective in the search for an optimal policy. Through experiments on benchmark problems, we show that our algorithm is competitive to the recent nonlinear optimization approach, both in the solution time and the policy quality.