Self-calibration of environmental camera for mobile robot navigation
Robotics and Autonomous Systems
Autonomous robot navigation in human-centered environments based on 3D data fusion
EURASIP Journal on Applied Signal Processing
Learning semantic scene models from observing activity in visual surveillance
IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics
Hi-index | 0.00 |
In this paper, we present a multi-agent system for easy and quick robot deployment of robots in different environments, without prior software development and without the need of maps of the environment where the robot will work. Our system, consisting on a network of intelligent cameras and autonomous robots, is able to manage local interactions amongst the intelligent cameras spread across the environment, so that they will self-organize to support robots' navigation and even inform the robots about the existence of events that require their presence. Our design is scalable, robust, and flexible to the environment, so that we get robots able to navigate safely attending events that require their presence. In this paper, we show the performance of a first implementation of our system, which was validated in real world experiments, and upon which we will continue working.