3D Scene Interpretation Through Computer Vision from the Coordinated Analysis of Images Obtained by

Research Project DPI2007-66556-C03-03 (2007-2010)

Interpretacion de Escenas 3D Mediante Vision Artificial a Partir del Analisis Coordinado de Imagenes Adquiridas por un Equipo de Robots Moviles

This project aims at designing new computer vision techniques that allow the automatic interpretation of 3D scenes from the analysis of images provided by a team of mobile robots. The new schemes will take advantage of the different images correspondihg to a same region delivered by each member of the team by means of a coordinated inspection. Such a interpretation is expected to contribute to increasing the performance of cooperative tasks developed by the team of robots. In this way, the accuracy of both the self-localization of mobile robots in the space and the global 3D model of the environment is desired as a result of those techniques.

nao3

In particular, this project nas the following specific objectives:

  • Computer Vision Goals: new algorithms with the purpose that robots automatically obtain a manageable description of the objects in a 3D scene and their spatial and topological interrelations. 3D scene interpretation by using the description previously obtained.
  • Localization Goals: simultaneous self-localization and mapping techniques with 6 degrees of freedom (SLAM 3D) over unstructured environments. Previous scene interpretation will be especially considered for the accomplishment of this objective.
  • Interaction Goals: new techniques for the collaborative exploration of unknown environments through a team of mobile robots, by taking into account previously proposed simulation algorithms.

The team of mobile robots is constituted by a NAO V3 robot, a Pioneer P2-AT robot, a Pioneer P3-AT robot and three Koala robots. The first is a humanoid robot. The other five are wheeled all-terrain robots. Each Koa”a will mount a binocular Color Bumblebee camera. On the other hand, the P2-AT is endowed with a trinocular Color Digiclops camera, and the P3-AT mounts a binocular Color Bumblebee-2 camera. All 3D cameras will be controlled by embedded computers. All robots will communicate with each other and with external computers through a wireless network.

Read More

Coordinated Exploration of Wide-Area Environments with Multiple Robots Through Vision-Based 3D SLAM

Research Project DPI2004-07993-C03-03 (2004-2007)

Exploración Coordinada de Entornos Extensos con Múltiples Robots Mediante SLAM 3D Basado en Visión

This project aims at designing and implementing new exploration strategies that allow a team of mobile robots to deploy collaboratively in order to obtain processable three-dimensional models of wide-area, unknown environments. Every robot will map its surroundings and simultaneously determine its position and orientation in space by processing visual information obtained by means of an off-the-shelf stereo camera. The accuracy of both the robot’s pose and the locally obtained 3D models will be continuously improved by integrating information gathered by separate robots. In this way, a consistent global 3D model of the environment is expected as a result of the exploration process.

visual_slam

The team of explorers is constituted by three Koala robots and a Pioneer P2-AT robot. The four robots are all-terrair. Each Koala will mount a binocular Color Bumblebee camera, one ultrasound-based range sensor and a ring of infrared proximity sensons. On the other hand, the P2-AT is endowed with a ring of ultrasound-based range sensors and a trinocular Color Digiclops camera. All 3D cameras will be controlled by embedded computers. All robots will communicate with each other and with external computers through a Bluetooth wireless network.

This project is a part of a larger coordinated project entitled: Development and Integration of Perception and Actuation Techniques in Teams of Mobile Robots (“Desarrollo e Integracion de Tecnicas de Percepcion y Actuacion en Grupos de Robots Moviles”) DPI2004-07993-C03.

Read More