Learning models of other agents using influence diagrams
UM '99 Proceedings of the seventh international conference on User modeling
A Calculus of Communicating Systems
A Calculus of Communicating Systems
The Complexity of Decentralized Control of Markov Decision Processes
Mathematics of Operations Research
Refinement and coarsening of Bayesian networks
UAI '90 Proceedings of the Sixth Annual Conference on Uncertainty in Artificial Intelligence
Equivalence notions and model minimization in Markov decision processes
Artificial Intelligence - special issue on planning with uncertainty and incomplete information
Exact solutions of interactive POMDPs using behavioral equivalence
AAMAS '06 Proceedings of the fifth international joint conference on Autonomous agents and multiagent systems
The FlightGear flight simulator
ATEC '04 Proceedings of the annual conference on USENIX Annual Technical Conference
Formal models and algorithms for decentralized decision making under uncertainty
Autonomous Agents and Multi-Agent Systems
Graphical models for interactive POMDPs: representations and solutions
Autonomous Agents and Multi-Agent Systems
Lossless clustering of histories in decentralized POMDPs
Proceedings of The 8th International Conference on Autonomous Agents and Multiagent Systems - Volume 1
Approximate solutions of interactive dynamic influence diagrams using model clustering
AAAI'07 Proceedings of the 22nd national conference on Artificial intelligence - Volume 1
AAAI'07 Proceedings of the 22nd national conference on Artificial intelligence - Volume 2
A framework for sequential planning in multi-agent settings
Journal of Artificial Intelligence Research
Anytime point-based approximations for large POMDPs
Journal of Artificial Intelligence Research
Journal of Artificial Intelligence Research
HUGIN: a shell for building Bayesian belief universes for expert systems
IJCAI'89 Proceedings of the 11th international joint conference on Artificial intelligence - Volume 2
Memory-bounded dynamic programming for DEC-POMDPs
IJCAI'07 Proceedings of the 20th international joint conference on Artifical intelligence
Taming decentralized POMDPs: towards efficient policy computation for multiagent settings
IJCAI'03 Proceedings of the 18th international joint conference on Artificial intelligence
Multi-agent influence diagrams for representing and solving games
IJCAI'01 Proceedings of the 17th international joint conference on Artificial intelligence - Volume 2
Planning and acting in partially observable stochastic domains
Artificial Intelligence
Equivalence relations in fully and partially observable Markov decision processes
IJCAI'09 Proceedings of the 21st international jont conference on Artifical intelligence
Artificial Intelligence: A Modern Approach
Artificial Intelligence: A Modern Approach
History-dependent graphical multiagent models
Proceedings of the 9th International Conference on Autonomous Agents and Multiagent Systems: volume 1 - Volume 1
Evaluating influence diagrams using LIMIDs
UAI'00 Proceedings of the Sixteenth conference on Uncertainty in artificial intelligence
Graphical models for game theory
UAI'01 Proceedings of the Seventeenth conference on Uncertainty in artificial intelligence
Improved use of partial policies for identifying behavioral equivalence
Proceedings of the 11th International Conference on Autonomous Agents and Multiagent Systems - Volume 2
Learning Communication in Interactive Dynamic Influence Diagrams
WI-IAT '12 Proceedings of the The 2012 IEEE/WIC/ACM International Joint Conferences on Web Intelligence and Intelligent Agent Technology - Volume 02
Opponent modeling in a PGM framework
Proceedings of the 2013 international conference on Autonomous agents and multi-agent systems
Incremental clustering and expansion for faster optimal planning in decentralized POMDPs
Journal of Artificial Intelligence Research
Hi-index | 0.00 |
We focus on the problem of sequential decision making in partially observable environments shared with other agents of uncertain types having similar or conflicting objectives. This problem has been previously formalized by multiple frameworks one of which is the interactive dynamic influence diagram (I-DID), which generalizes the well-known influence diagram to the multiagent setting. I-DIDs are graphical models and may be used to compute the policy of an agent given its belief over the physical state and others' models, which changes as the agent acts and observes in the multiagent setting. As we may expect, solving I-DIDs is computationally hard. This is predominantly due to the large space of candidate models ascribed to the other agents and its exponential growth over time. We present two methods for reducing the size of the model space and stemming its exponential growth. Both these methods involve aggregating individual models into equivalence classes. Our first method groups together behaviorally equivalent models and selects only those models for updating which will result in predictive behaviors that are distinct from others in the updated model space. The second method further compacts the model space by focusing on portions of the behavioral predictions. Specifically, we cluster actionally equivalent models that prescribe identical actions at a single time step. Exactly identifying the equivalences would require us to solve all models in the initial set. We avoid this by selectively solving some of the models, thereby introducing an approximation. We discuss the error introduced by the approximation, and empirically demonstrate the improved efficiency in solving I-DIDs due to the equivalences.