Revisiting Asimov's First Law: A Response to the Call to Arms
ATAL '01 Revised Papers from the 8th International Workshop on Intelligent Agents VIII
A prototype infrastructure for distributed robot-agent-person teams
AAMAS '03 Proceedings of the second international joint conference on Autonomous agents and multiagent systems
The TRAINS Project: A Case Study in Defining a Conversational Planning Agent
The TRAINS Project: A Case Study in Defining a Conversational Planning Agent
Exploiting belief bounds: practical POMDPs for personal assistant agents
Proceedings of the fourth international joint conference on Autonomous agents and multiagent systems
The DEFACTO system: training tool for incident commanders
IAAI'05 Proceedings of the 17th conference on Innovative applications of artificial intelligence - Volume 3
Journal of Artificial Intelligence Research
Towards adjustable autonomy for the real world
Journal of Artificial Intelligence Research
Planning and acting in partially observable stochastic domains
Artificial Intelligence
Science fiction in computer science education
Proceedings of the 43rd ACM technical symposium on Computer Science Education
Hi-index | 0.00 |
In the March 1942 issue of "Astounding Science Fiction", Isaac Asimov for the first time enumerated his three laws of robotics. Decades later, researchers in agents and multiagent systems have begun to examine these laws for providing a useful set of guarantees on deployed agent systems. Motivated by unexpected failures or behavior degradations in complex mixed agent-human teams, this paper for the first time focuses on applying Asimov's first two laws to provide behavioral guarantees in such teams. However, operationalizing these laws in the context of such mixed agent-human teams raises three novel issues. First, while the laws were originally written for interaction of an individual robot and an individual human, clearly, our systems must operate in a team context. Second, key notions in these laws (e.g. causing "harm" to humans) are specified in very abstract terms and must be specified in concrete terms in implemented systems. Third, since removed from science-fiction, agents or humans may not have perfect information about the world, they must act based on these laws despite uncertainty of information. Addressing this uncertainty is a key thrust of this paper, and we illustrate that agents must detect and overcome such states of uncertainty while ensuring adherence to Asimov's laws. We illustrate the results of two different domains that each have different approaches to operationalizing Asimov's laws.