Intelligence without robots: a reply to Brooks
AI Magazine
Machine Learning
Introduction to Reinforcement Learning
Introduction to Reinforcement Learning
The agent-based perspective on imitation
Imitation in animals and artifacts
Imitation in animals and artifacts
Software agents that learn through observation
AAMAS '06 Proceedings of the fifth international joint conference on Autonomous agents and multiagent systems
A survey of robot learning from demonstration
Robotics and Autonomous Systems
Confidence-based robot policy learning from demonstration
Confidence-based robot policy learning from demonstration
International Journal of Robotics Research
A Developmental Roadmap for Learning by Imitation in Robots
IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics
Imitation with ALICE: learning to imitate corresponding actions across dissimilar embodiments
IEEE Transactions on Systems, Man, and Cybernetics, Part A: Systems and Humans
Multiagent hierarchical learning from demonstration
IJCAI'11 Proceedings of the Twenty-Second international joint conference on Artificial Intelligence - Volume Volume Three
Hi-index | 0.00 |
Learning by observation can be of key importance whenever agents sharing similar features want to learn from each other. This paper presents an agent architecture that enables software agents to learn by direct observation of the actions executed by expert agents while they are performing a task. This is possible because the proposed architecture displays information that is essential for observation, making it possible for software agents to observe each other. The agent architecture supports a learning process that covers all aspects of learning by observation, such as discovering and observing experts, learning from the observed data, applying the acquired knowledge and evaluating the agent's progress. The evaluation provides control over the decision to obtain new knowledge or apply the acquired knowledge to new problems. We combine two methods for learning from the observed information. The first one, the recall method, uses the sequence on which the actions were observed to solve new problems. The second one, the classification method, categorizes the information in the observed data and determines to which set of categories the new problems belong. Results show that agents are able to learn in conditions where common supervised learning algorithms fail, such as when agents do not know the results of their actions a priori or when not all the effects of the actions are visible. The results also show that our approach provides better results than other learning methods since it requires shorter learning periods.