Learning to coordinate without sharing information
AAAI '94 Proceedings of the twelfth national conference on Artificial intelligence (vol. 1)
HTN planning: complexity and expressivity
AAAI'94 Proceedings of the twelfth national conference on Artificial intelligence (vol. 2)
Reinforcement learning with hierarchies of machines
NIPS '97 Proceedings of the 1997 conference on Advances in neural information processing systems 10
Theory for coordinating concurrent hierarchical planning agents using summary information
AAAI '99/IAAI '99 Proceedings of the sixteenth national conference on Artificial intelligence and the eleventh Innovative applications of artificial intelligence conference innovative applications of artificial intelligence
The MAXQ Method for Hierarchical Reinforcement Learning
ICML '98 Proceedings of the Fifteenth International Conference on Machine Learning
Intra-Option Learning about Temporally Abstract Actions
ICML '98 Proceedings of the Fifteenth International Conference on Machine Learning
AAMAS '04 Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems - Volume 1
Conflict estimation of abstract plans for multiagent systems
Proceedings of the 6th international joint conference on Autonomous agents and multiagent systems
Multi-agent Planning in Sokoban
CEEMAS '07 Proceedings of the 5th international Central and Eastern European conference on Multi-Agent Systems and Applications V
Hi-index | 0.00 |
This paper proposes a learning method to select the most appropriate abstract plans during hierarchical planning in the context of multi-agent systems (MAS). In hierarchical planning, a plan is first created at the most abstract level, and is then refined to a more concrete plan, level by level. Thus, selecting an appropriate plan at the abstract level is very important because the selected plan restricts the scope of lower concrete-level plans. This restriction can enable agents to create plans efficiently, but if all the plans under the selected plan contain serious and difficult-to-resolve conflicts with other agents' plans, the resulting plan does not work well or is of low quality. We propose a method in which, from the conflict pattern among agents' plans, an agent learns which abstract plans will cause conflicts with less probability or which conflicts are easy to resolve, thus inducing probabilistically higher-utility concrete plans after conflict resolution. We also show some experimental results to evaluate our method, with the results suggesting structures of resources where tasks are executed.