Macro-operators: a weak method for learning
Artificial Intelligence - Lecture notes in computer science 178
The complexity of Markov decision processes
Mathematics of Operations Research
Linear programming and network flows (2nd ed.)
Linear programming and network flows (2nd ed.)
A model for reasoning about persistence and causation
Computational Intelligence
Planning under time constraints in stochastic domains
Artificial Intelligence - Special volume on planning and scheduling
Finite State Markovian Decision Processes
Finite State Markovian Decision Processes
Decomposition Techniques for Planning in Stochastic Domains
Decomposition Techniques for Planning in Stochastic Domains
Exploiting structure in policy construction
IJCAI'95 Proceedings of the 14th international joint conference on Artificial intelligence - Volume 2
Solving very large weakly coupled Markov decision processes
AAAI '98/IAAI '98 Proceedings of the fifteenth national/tenth conference on Artificial intelligence/Innovative applications of artificial intelligence
Tree based discretization for continuous state space reinforcement learning
AAAI '98/IAAI '98 Proceedings of the fifteenth national/tenth conference on Artificial intelligence/Innovative applications of artificial intelligence
A new decomposition technique for solving Markov decision processes
Proceedings of the 2001 ACM symposium on Applied computing
Self-Similar Layered Hidden Markov Models
PKDD '01 Proceedings of the 5th European Conference on Principles of Data Mining and Knowledge Discovery
Decision-Theoretic Control of Planetary Rovers
Revised Papers from the International Seminar on Advances in Plan-Based Control of Robotic Agents,
Towards Stochastic Constraint Programming: A Study of Online Multi-choice Knapsack with Deadlines
CP '01 Proceedings of the 7th International Conference on Principles and Practice of Constraint Programming
Value iteration working with belief subset
Eighteenth national conference on Artificial intelligence
Mobile Robotics Planning Using Abstract Markov Decision Processes
ICTAI '99 Proceedings of the 11th IEEE International Conference on Tools with Artificial Intelligence
Performance models for large scale multiagent systems: using distributed POMDP building blocks
AAMAS '03 Proceedings of the second international joint conference on Autonomous agents and multiagent systems
Reinforcement learning based on local state feature learning and policy adjustment
Information Sciences—Informatics and Computer Science: An International Journal - Special issue: Introduction to multimedia and mobile agents
Planning, learning and coordination in multiagent decision processes
TARK '96 Proceedings of the 6th conference on Theoretical aspects of rationality and knowledge
Causal Graph Based Decomposition of Factored MDPs
The Journal of Machine Learning Research
Economic hierarchical Q-learning
AAAI'08 Proceedings of the 23rd national conference on Artificial intelligence - Volume 2
Restricted value iteration: theory and algorithms
Journal of Artificial Intelligence Research
Hybrid BDI-POMDP framework for multiagent teaming
Journal of Artificial Intelligence Research
Policy recognition in the abstract hidden Markov model
Journal of Artificial Intelligence Research
Computing near optimal strategies for stochastic investment planning problems
IJCAI'99 Proceedings of the 16th international joint conference on Artificial intelligence - Volume 2
Bounding the suboptimality of reusing subproblems
IJCAI'99 Proceedings of the 16th international joint conference on Artificial intelligence - Volume 2
An overview of planning under uncertainty
Artificial intelligence today
Analogical replay for efficient conditional planning
AAAI'97/IAAI'97 Proceedings of the fourteenth national conference on artificial intelligence and ninth conference on Innovative applications of artificial intelligence
Resource-driven mission-phasing techniques for constrained agents in stochastic environments
Journal of Artificial Intelligence Research
Solving efficiently Decentralized MDPs with temporal and resource constraints
Autonomous Agents and Multi-Agent Systems
Distributed planning in hierarchical factored MDPs
UAI'02 Proceedings of the Eighteenth conference on Uncertainty in artificial intelligence
A clustering approach to solving large stochastic matching problems
UAI'01 Proceedings of the Seventeenth conference on Uncertainty in artificial intelligence
Structured reachability analysis for Markov decision processes
UAI'98 Proceedings of the Fourteenth conference on Uncertainty in artificial intelligence
Hierarchical solution of Markov decision processes using macro-actions
UAI'98 Proceedings of the Fourteenth conference on Uncertainty in artificial intelligence
Flexible decomposition algorithms for weakly coupled Markov decision problems
UAI'98 Proceedings of the Fourteenth conference on Uncertainty in artificial intelligence
On the complexity of solving Markov decision problems
UAI'95 Proceedings of the Eleventh conference on Uncertainty in artificial intelligence
Coordinating teams in uncertain environments: a hybrid BDI-POMDP approach
ProMAS'04 Proceedings of the Second international conference on Programming Multi-Agent Systems
Map partitioning to approximate an exploration strategy in mobile robotics
Multiagent and Grid Systems
Hi-index | 0.00 |
This paper is concerned with modeling planning problems involving uncertainty as discrete-time, finite-state stochastic automata Solving planning problems is reduced to computing policies for Markov decision processes. Classical methods for solving Markov decision processes cannot cope with the size of the state spaces for typical problems encountered in practice. As an alternative, we investigate methods that decompose global planning problems into a number of local problems solve the local problems separately and then combine the local solutions to generate a global solution. We present algorithms that decompose planning problems into smaller problems given an arbitrary partition of the state space. The local problems are interpreted as Markov decision processes and solutions to the local problems are interpreted as policies restricted to the subsets of the state space defined by the partition. One algorithm relies on constructing and solving an abstract version of the original decision problem. A second algorithm iteratively approximates parameters of the local problems to converge to an optimal solution. We show how properties of a specified partition affect the time and storage required for these algorithms.