Markov Decision Processes: Discrete Stochastic Dynamic Programming
Markov Decision Processes: Discrete Stochastic Dynamic Programming
ICALP '96 Proceedings of the 23rd International Colloquium on Automata, Languages and Programming
Cooperative exploration in the electronic marketplace
AAAI'05 Proceedings of the 20th national conference on Artificial intelligence - Volume 1
Social interaction under uncertainty in multi agent systems
AAAI'08 Proceedings of the 23rd national conference on Artificial intelligence - Volume 3
Collaborative multi agent physical search with Probabilistic knowledge
IJCAI'09 Proceedings of the 21st international jont conference on Artifical intelligence
Physical search problems with probabilistic knowledge
Artificial Intelligence
Hi-index | 0.00 |
This paper considers the problem of an agent searching for a resource or a tangible good in a physical environment, where at each stage of its search it observes one source where this good can be found. The cost of acquiring the resource or good at a given source is uncertain (a-priori), and the agent can observe its true value only when physically arriving at the source. Sample applications involving this type of search include agents in exploration and patrol missions (e.g., an agent seeking to find the best location to deploy sensing equipment along its path). The uniqueness of these settings is that the expense of observing the source on each step of the process derives from the last source the agent explored. We analyze three variants of the problem, differing in their objective: minimizing the total expected cost, maximizing the success probability given an initial budget, and minimizing the budget necessary to obtain a given success probability. For each variant, we first introduce and analyze the problem with a single agent, either providing a polynomial solution to the problem or proving it is NP-Complete. We also introduce an innovative fully polynomial time approximation scheme algorithm for the minimum budget variant. Finally, the results for the single agent case are generalized to multi-agent settings.