Taxonomic syntax for first order inference
Journal of the ACM (JACM)
Planning under time constraints in stochastic domains
Artificial Intelligence - Special volume on planning and scheduling
Machine Learning
Learning action strategies for planning domains
Artificial Intelligence
Stochastic dynamic programming with factored representations
Artificial Intelligence
Relational reinforcement learning
Machine Learning - Special issue on inducive logic programming
Markov Decision Processes: Discrete Stochastic Dynamic Programming
Markov Decision Processes: Discrete Stochastic Dynamic Programming
Learning Logical Definitions from Relations
Machine Learning
Machine Learning
Probabilistic Planning in the Graphplan Framework
ECP '99 Proceedings of the 5th European Conference on Planning: Recent Advances in AI Planning
Human Problem Solving
Max-norm projections for factored MDPs
IJCAI'01 Proceedings of the 17th international joint conference on Artificial intelligence - Volume 1
Symbolic dynamic programming for first-order MDPs
IJCAI'01 Proceedings of the 17th international joint conference on Artificial intelligence - Volume 1
Observations on cognitive judgments
AAAI'91 Proceedings of the ninth National conference on Artificial intelligence - Volume 2
Model minimization in Markov decision processes
AAAI'97/IAAI'97 Proceedings of the fourteenth national conference on artificial intelligence and ninth conference on Innovative applications of artificial intelligence
Integrating Guidance into Relational Reinforcement Learning
Machine Learning
Exploiting first-order regression in inductive policy selection
UAI '04 Proceedings of the 20th conference on Uncertainty in artificial intelligence
Learning Control Knowledge for Forward Search Planning
The Journal of Machine Learning Research
Practical solution techniques for first-order MDPs
Artificial Intelligence
An Inductive Logic Programming Approach to Statistical Relational Learning
Proceedings of the 2005 conference on An Inductive Logic Programming Approach to Statistical Relational Learning
Learning measures of progress for planning domains
AAAI'05 Proceedings of the 20th national conference on Artificial intelligence - Volume 3
Journal of Artificial Intelligence Research
Learning symbolic models of stochastic domains
Journal of Artificial Intelligence Research
Using learned policies in heuristic-search planning
IJCAI'07 Proceedings of the 20th international joint conference on Artifical intelligence
Team programming in Golog under partial observability
IJCAI'07 Proceedings of the 20th international joint conference on Artifical intelligence
Approximate policy iteration using large-margin classifiers
IJCAI'03 Proceedings of the 18th international joint conference on Artificial intelligence
Inference and Learning in Planning (Extended Abstract)
DS '09 Proceedings of the 12th International Conference on Discovery Science
Learning Linear Ranking Functions for Beam Search with Application to Planning
The Journal of Machine Learning Research
Game-theoretic agent programming in Golog under partial observability
KI'06 Proceedings of the 29th annual German conference on Artificial intelligence
Automatic induction of bellman-error features for probabilistic planning
Journal of Artificial Intelligence Research
Unique state and automatical action abstracting based on logical MDPs with negation
ICNC'06 Proceedings of the Second international conference on Advances in Natural Computation - Volume Part II
Imitation learning in relational domains: a functional-gradient boosting approach
IJCAI'11 Proceedings of the Twenty-Second international joint conference on Artificial Intelligence - Volume Volume Two
An Ensemble Architecture for Learning Complex Problem-Solving Techniques from Demonstration
ACM Transactions on Intelligent Systems and Technology (TIST)
Hi-index | 0.00 |
We select policies for large Markov Decision Processes (MDPs) with compact first-order representations. We find policies that generalize well as the number of objects in the domain grows, potentially without bound. Existing dynamic-programming approaches based on flat, propositional, or first-order representations either are impractical here or do not naturally scale as the number of objects grows without bound. We implement and evaluate an alternative approach that induces first-order policies using training data constructed by solving small problem instances using PGraphplan (Blurn & Langford, 1999). Our policies are represented as ensembles of decision lists, using a taxonomic concept language. This approach extends the work of Martin and Geffner (2000) to stochastic domains, ensemble learning, and a wider variety of problems. Empirically, we find "good" policies for several stochastic first-order MDPs that are beyond the scope of previous approaches. We also discuss the application of this work to the relational reinforcement-learning problem.