Acting optimally in partially observable stochastic domains
AAAI'94 Proceedings of the twelfth national conference on Artificial intelligence (vol. 2)
Leveraging agent properties to assure survivability of distributed multi-agent systems
AAMAS '03 Proceedings of the second international joint conference on Autonomous agents and multiagent systems
A Framework to Control Emergent Survivability of Multi Agent Systems
AAMAS '04 Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems - Volume 1
Hi-index | 0.00 |
A survivable agent system depends on the incorporation of many recovery features. However, the optimal use of these features requires the ability to assess the actual state of the agent system accurately at a given time. This paper describes an approach for the estimation of the state of an agent system using Partially-Observable Markov Decision Processes (POMDPS). POMDPS are dependent on a model of the agent system - components, environment, sensors, and the actuators that can correct problems. Based on this model, we define a state estimation for each component (asset) in the agent system. We model a survivable agent system as a POMDP that takes into account both environmental threats and observations from sensors. We describe the process of updating the state estimation as time passes, as sensor inputs are received, and as actuators affect changes. This state estimation process has been deployed within the agent system that runs the Ultralog application and tested using Ultralog's survivability tests on a full-scale (1000+) agent system. This test successfully ran a long-running logistics application in an unstable environment with high failure rates.