Parallel and distributed computation: numerical methods
Parallel and distributed computation: numerical methods
A survey of algorithmic methods for partially observed Markov decision processes
Annals of Operations Research
Acting optimally in partially observable stochastic domains
AAAI'94 Proceedings of the twelfth national conference on Artificial intelligence (vol. 2)
Exploiting structure for planning and control
Exploiting structure for planning and control
Generalized prioritized sweeping
NIPS '97 Proceedings of the 1997 conference on Advances in neural information processing systems 10
Reinforcement learning with hierarchies of machines
NIPS '97 Proceedings of the 1997 conference on Advances in neural information processing systems 10
Multi-time models for temporally abstract planning
NIPS '97 Proceedings of the 1997 conference on Advances in neural information processing systems 10
Hierarchical control and learning for markov decision processes
Hierarchical control and learning for markov decision processes
Decomposition techniques for planning in stochastic domains
IJCAI'95 Proceedings of the 14th international joint conference on Artificial intelligence - Volume 2
Hierarchical solution of Markov decision processes using macro-actions
UAI'98 Proceedings of the Fourteenth conference on Uncertainty in artificial intelligence
A new decomposition technique for solving Markov decision processes
Proceedings of the 2001 ACM symposium on Applied computing
Decision-Theoretic Control of Planetary Rovers
Revised Papers from the International Seminar on Advances in Plan-Based Control of Robotic Agents,
Towards Stochastic Constraint Programming: A Study of Online Multi-choice Knapsack with Deadlines
CP '01 Proceedings of the 7th International Conference on Principles and Practice of Constraint Programming
Nearly deterministic abstractions of Markov decision processes
Eighteenth national conference on Artificial intelligence
Mobile Robotics Planning Using Abstract Markov Decision Processes
ICTAI '99 Proceedings of the 11th IEEE International Conference on Tools with Artificial Intelligence
Performance models for large scale multiagent systems: using distributed POMDP building blocks
AAMAS '03 Proceedings of the second international joint conference on Autonomous agents and multiagent systems
Planning and programming with first-order markov decision processes: insights and challenges
TARK '01 Proceedings of the 8th conference on Theoretical aspects of rationality and knowledge
Automated resource-driven mission phasing techniques for constrained agents
Proceedings of the fourth international joint conference on Autonomous agents and multiagent systems
Dynamic preferences in multi-criteria reinforcement learning
ICML '05 Proceedings of the 22nd international conference on Machine learning
Analyzing characteristics of task structures to develop GPGP coordination mechanisms
AAMAS '06 Proceedings of the fifth international joint conference on Autonomous agents and multiagent systems
Solving large TÆMS problems efficiently by selective exploration and decomposition
Proceedings of the 6th international joint conference on Autonomous agents and multiagent systems
Transfer in variable-reward hierarchical reinforcement learning
Machine Learning
A Cooperative Distributed Problem Solving Technique for Large Markov Decision Processes
Proceedings of the 2006 conference on ECAI 2006: 17th European Conference on Artificial Intelligence August 29 -- September 1, 2006, Riva del Garda, Italy
Hierarchical reinforcement learning with the MAXQ value function decomposition
Journal of Artificial Intelligence Research
Accelerating reinforcement learning by composing solutions of automatically identified subtasks
Journal of Artificial Intelligence Research
Restricted value iteration: theory and algorithms
Journal of Artificial Intelligence Research
Probabilistic hybrid action models for predicting concurrent percept-driven robot behavior
Journal of Artificial Intelligence Research
Bounding the suboptimality of reusing subproblems
IJCAI'99 Proceedings of the 16th international joint conference on Artificial intelligence - Volume 2
Resource-driven mission-phasing techniques for constrained agents in stochastic environments
Journal of Artificial Intelligence Research
Robust combination of local controllers
UAI'01 Proceedings of the Seventeenth conference on Uncertainty in artificial intelligence
Topological value iteration algorithms
Journal of Artificial Intelligence Research
Map partitioning to approximate an exploration strategy in mobile robotics
Multiagent and Grid Systems
Hi-index | 0.00 |
This paper presents two new approaches to decomposing and solving large Markov decision problems (MDPs), a partial decoupling method and a complete decoupling method. In these approaches, a large, stochastic decision problem is divided into smaller pieces. The first approach builds a cache of policies for each part of the problem independently, and then combines the pieces in a separate, light-weight step. A second approach also divides the problem into smaller pieces, but information is communicated between the different problem pieces, allowing intelligent decisions to be made about which piece requires the most attention. Both approaches can be used to find optimal policies or approximately optimal policies with provable bounds. These algorithms also provide a framework for the efficient transfer of knowledge across problems that share similar structure.