Visual correction for mobile robot homing

https://doi.org/10.1016/j.robot.2004.08.005Get rights and content

Abstract

We present a method to send a mobile robot to locations specified by images previously taken from these positions, which sometimes has been referred as homing. Classically this has been carried out using the fundamental matrix, but the fundamental matrix is ill conditioned with planar scenes, which are quite usual in man made environments. Many times in robot homing, small baseline images with high disparity due to rotation are compared, where the fundamental matrix also gives bad results. We use a monocular vision system and we compute motion through an homography obtained from automatically matched lines. In this work we compare the use of the homography and the fundamental matrix and we propose the correction of motion directly from the parameters of the 2D homography, which only needs one calibration parameter. It is shown that it is robust, sufficiently accurate and simple.

Introduction

Robot navigation normally has involved the use of commands to move the robot to the desired position. These commands include the required position and orientation, which suppose good measurement of the motion made by the robot. However, odometry errors or slipping and mechanical drifts may make the desired position not to be reached. Therefore, the use of an additional perception system is mandatory. Vision is perhaps the most broadly researched perception system.

Using vision, some autonomous vehicles are able to execute tasks based on landmarks which give global localization (Map-Based Navigation). Others carry out specific tasks, building maps of the environment simultaneously (Map-Building-Based Navigation). Our system can be classified in a third group as Map-less navigation [2], because the robot can autonomously navigate without prepared landmarks or complex map-building systems. In our work the target positions are specified with images taken and memorized in the teaching phase. In the playback phase the motion correction to reach to the target position is computed from a projective transformation, which is obtained by processing the current image and the stored reference image. The geometric features extracted and matched are lines which have some advantages with respect to points [3], specially in man made environments.

This kind of navigation using a representation of the route with a sequence of images has been previously considered by a correlation based matching of the current and the reference images [7]. Extracting and matching geometric information from images is not currently time costly and geometric based approaches are less sensitive to noise or illumination changes than others. Of course, the data of the target images to memorize is lower in our proposal, since only extracted lines are stored. Other authors [9] also use vertical lines to correct robot motion, but using a calibrated trinocular vision system.

The recovering of motion from geometric features has been presented using the epipolar geometry [1]. Rotation and direction of translation are initially computed from the essential matrix. In addition, the steps to collision are also computed using a third image. However, there are situations where the fundamental matrix is not meaningful (small translations or planar scenes) and other models are needed to obtain motion [5], [8]. We recover motion from homography solving matches of lines automatically. To match them, we use image information and the constraints imposed by the projective transformation. Besides that, robust statistical techniques are considered, which make the complete process useful in real applications. Some experiments show results in typical situations of visual robot homing (Section 6).

Section snippets

Motion from two images

In this work, motion information is obtained from the previously stored image and the current image, taken both with the same camera. We first discuss the ways to compute motion from two images of a man made environment, where straight lines and planar surfaces are plentiful.

Two perspective images can be geometrically linked by linear algebraic relations: the fundamental matrix and the homography. An homography relates points or lines in one image belonging to a plane of the scene with points

Computation of 2D homography

Our approach takes straight lines in the image as key features, because they are plentiful in man made environments. The straight lines have a simple mathematical representation, they can be extracted more accurately than points and they can be used in cases where there are partial occlusions. After extracting the lines, automatic computation of correspondences and homographies is carried out as previously presented [4]. Thus, the extracted lines are initially putatively matched to the weighted

2D homography and planar motion

Let us suppose vertical images of vertical planar scenes. The relative motion between the reference system attached to the first camera position and the reference attached to the planar scene can be written as a function of the distance d from the origin to the plane and the angle ϕ of the plane with respect to the camera reference system. Similarly, the transformation between both camera locations can be written as a function of the rotation θ and the translation (tx,tz). Developing Eq. (1)

Performance of approximate motion from homography

We have built a simulator of scenes and motions and we have compared the homography parameters obtained in different situations in relation to the motion. The comparison has been made in three situations: Rotation with no translation, translation with no rotation, and rotation with translation. As told (Section 4), the rotation can be approached from μ, ρ or from the eigenvalues and the relative translation along the z-axis (which indicates the advance of the robot when it is looking ahead) can

Experimental results

We have performed several experiments to evaluate the accuracy with real images (Fig. 4) computing the motion from the μ parameter. We compare it with the rotation computed through the fundamental matrix. The homography has been automatically computed using lines as proposed in [4] and the fundamental matrix could be obtained from at least two homographies [10], but we have used the “image-matching” software [15] to have a more standard benchmark to compare results. The ground truth has been

Conclusions

For teaching by doing in robotic tasks, we have argued the computation of visual motion based on homographies. These tasks, also referred to as homing, have many times been solved using the fundamental matrix, but the correction from homographies behaves better in many practical situations. Additionally, homography based on vertical lines is sufficient for robots moving indoors. Several experiments, in simulation and with controlled images, have been made to show the performance of homography

Acknowledgment

This work was supported by projects DPI2000-1272 and DPI2003-07986.

C. Sagüés received the MSc and PhD degrees from the University of Zaragoza, Spain. During the course of his PhD he was working on force and distance sensors for robots. Since 1994 he has been with the Department of Computer Science and Systems Engineering as an assistant professor and currently he is the head teacher. He gives lectures on control engineering, robotics for BSc students and he also gives lectures on mobile robots and 3D computer vision for MSc students. His current research

References (15)

There are more references available in the full text version of this article.

Cited by (40)

  • A fast robot homing approach using sparse image waypoints

    2012, Image and Vision Computing
    Citation Excerpt :

    In the case of local visual homing, numerous studies have emphasized on the accuracy of guiding a robot between consecutive image clues. In general, two classes of approaches are developed: holistic methods [1,2] and correspondence-based methods [3–22]. In a holistic homing method, robot's view is warped in order to generate the most similar image to the waypoint image [1,2] because no correspondence is detected between two images.

  • Visual control through the trifocal tensor for nonholonomic robots

    2010, Robotics and Autonomous Systems
    Citation Excerpt :

    This geometric model is robust and well defined with short baseline. A well-known hybrid method is referred in [9,10] and more examples of visual control based on homographies are cited in [11–16]. However, the performance of a homography-based control can be affected if there is no dominant plane in the scene.

  • Vision-Based Yaw Bias Estimation for a Quadrotor UAV in Indoor Environment

    2018, IEEE International Conference on Control and Automation, ICCA
  • An image-based Accurate Alignment for Substation Inspection Robot

    2018, 2018 International Conference on Power System Technology, POWERCON 2018 - Proceedings
View all citing articles on Scopus

C. Sagüés received the MSc and PhD degrees from the University of Zaragoza, Spain. During the course of his PhD he was working on force and distance sensors for robots. Since 1994 he has been with the Department of Computer Science and Systems Engineering as an assistant professor and currently he is the head teacher. He gives lectures on control engineering, robotics for BSc students and he also gives lectures on mobile robots and 3D computer vision for MSc students. His current research interest include visual robot navigation and computer vision, being the author of several conference and journal papers related to these topics.

J.J. Guerrero graduated in electrical engineering from the University of Zaragoza in 1989. He obtained the PhD in 1996 from the same institution, and currently he is assistant professor at the Department of Computer Science and Systems Engineering. He gives lectures on control engineering, robotics and 3D computer vision. His research interest are in the area of computer vision, particularly in 3D visual perception, photogrammetry, robotics and vision based navigation.

View full text