Anytime coordination for progressive planning agents
AAAI '99/IAAI '99 Proceedings of the sixteenth national conference on Artificial intelligence and the eleventh Innovative applications of artificial intelligence conference innovative applications of artificial intelligence
Satisficing Equilibria: A Non-Classical Theory of Games and Decisions
Autonomous Agents and Multi-Agent Systems
Multi-criteria Reinforcement Learning
ICML '98 Proceedings of the Fifteenth International Conference on Machine Learning
UCP-Networks: A Directed Graphical Representation of Conditional Utilities
UAI '01 Proceedings of the 17th Conference in Uncertainty in Artificial Intelligence
Degree of Local Cooperation and Its Implication on Global Utility
AAMAS '04 Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems - Volume 2
A polynomial algorithm for decentralized Markov decision processes with temporal constraints
Proceedings of the fourth international joint conference on Autonomous agents and multiagent systems
Algebraic Markov decision processes
IJCAI'05 Proceedings of the 19th international joint conference on Artificial intelligence
Vector Valued Markov Decision Process for robot platooning
Proceedings of the 2008 conference on ECAI 2008: 18th European Conference on Artificial Intelligence
Bounded decentralised coordination over multiple objectives
The 10th International Conference on Autonomous Agents and Multiagent Systems - Volume 1
Vector-Value Markov Decision Process for multi-objective stochastic path planning
International Journal of Hybrid Intelligent Systems
Hi-index | 0.00 |
Multi-Objective Multiagent Planning (MOMAP) addresses the problem of resolving conflicts between individual agent interests and the group interests. In this paper, we address this problem by presenting a formal framework to represent objective relationships, a decision model using a Vector-Valued Decentralized Markov Decision Process (2V-DEC-MDP) and an algorithm to solve the resulting 2V-DEC-MDP. The formal framework of a Vector-Valued MDP considered uses the value function which returns a vector representing the individual and the group interests. An optimal policy in such contexts is not clear but in this approach we develop a regret-based technique to find a good tradeoff between the group and individual interests. To do that, the approach we present uses Egalitarian Social Welfare orderings that allow an agent to consider during its local optimization the satisfaction of all criteria and reducing their differences. The obtained result is a good balance between individual and group satisfactions where the local policies can lead to more global satisfying behaviors in some settings. This result is illustrated in many examples and compared to alternate local policies.