Vision-based localization for mobile robots

https://doi.org/10.1016/S0921-8890(01)00138-5Get rights and content

Abstract

Robust self-localization and repositioning strategies are essential capabilities for robots operating in highly dynamic environments. Environments are considered highly dynamic, if objects of interest move continuously and quickly, and if chances of hitting or getting hit by other robots are quite significant. An outstanding example for such an environment is provided by RoboCup.

Vision system designs for robots used in RoboCup are based on several approaches, aimed at fitting both the mechanical characteristics of the players and the strategies and operations that the different roles or playing situations may require. This paper reviews three approaches to vision-based self-localization used in the RoboCup middle-size league competition and describes the results they achieve in the robot soccer environment for which they have been designed.

Introduction

To perform complex navigation tasks and to coordinate its movement with other agents in indoor environments, an autonomous mobile robot needs knowledge about the world in which it is moving. This enables it to recognize the main landmarks and determine its position with respect to the environment. In fact, answering the question “Where am I?” is one of the most crucial and, often, critical tasks (both in terms of quality and time complexity) for an autonomous mobile robot [6], [12]. The self-localization problem is usually solved with a combination of relative and absolute position measurement techniques [7]. Odometry and inertial navigation (dead reckoning) systems are usually integrated with methods employing active beacons, natural or artificial landmarks as well as environment model matching, in order to compensate for wheel slippage and other positioning errors [13]. Natural and artificial landmarks are distinct features that a robot can recognize from its sensory input. In general, landmarks have a fixed well-known position within an absolute reference system, relative to which a robot can localize itself. The main task in self-localization is then to recognize such landmarks reliably by processing the sensory input to compute the robot’s position [4], [26], [30].

The literature concerning computer vision reports on a large number of natural and artificial landmark positioning systems integrated in autonomous robot architectures [3], [8]. Several authors have tried to find a good correlation between what the robot perceives through a camera and its internal representation of the environment [21], [27]. In indoor environments, edges of doors or walls are commonly used as landmarks to perform this correlation, but often such patterns are too generic to allow an accurate localization in a more complex environment. Researchers have often focused on the choice of optimal shapes or patterns as landmarks so that robust localization can be obtained from a single [23], [24] or from multiple images [20]. Furthermore, considerable work has been devoted to the treatment of error in sensor observations [5]. Sometimes it can be useful to evaluate alternative techniques that can recognize landmarks which can be specially designed to carry both human- and robot-oriented visual information, though they might also be complemented by other natural signs which are often found in office and industrial environments [2], [28]. Often these methods do not yield high localization accuracy, but there is a number of practical situations, e.g., docking and navigating in small openings, where coarser robot positioning accuracy suffices.

The RoboCup environment provides researchers with a common environment in which it is possible to test and compare different visual self-localization methods reliably. Self-localization in RoboCup is very important to achieve coordination between players, both to defend the team’s penalty area and to make goal-oriented decisions [19]. Although the world of RoboCup is simpler than many real-life indoor environments, it is a very dynamic world in which the robot has to react quickly to the changes caused by the continuous movement of the robots and the ball. Most robots rely on odometry for real-time self-localization. However, odometry fails immediately when a robot hits other robots. The appeal of the RoboCup challenge is actually due to its highly dynamic environment, in which robots usually obtain only partial views of the scene, since reference markers are often occluded by other robots.

To reposition themselves, robots in RoboCup use different kinds of sensors, such as cameras, laser range finders, sonars, and so on. While self-localization based on laser range finders provides a very accurate and robust estimation of robot position, visual sensors are more flexible; the number and configuration of cameras can be chosen according to which other tasks are performed by the vision system and to the role that the robot has within the team. Most distance sensors require that the field of play be surrounded by walls. This limits their future use, when RoboCup eventually moves towards a more natural, wall-free environment.

In RoboCup, according to the current rules, the appearance of relevant objects is well known and it is therefore possible to provide robots with some geometric models of the field or of its subparts. Visual self-localization is typically performed by matching and comparing the robot’s view of the world with a model. This comparison can often be used to refine and compensate for the errors introduced by odometric data, but in some cases a robot must be able to localize itself without any help from sensory systems other than that of vision.

Several kinds of vision systems have been applied in RoboCup. While most robots use traditional single-camera vision systems, it is becoming quite common to adopt more complex solutions. For example, the problem of having only a limited field of view can be solved using a camera pointing upwards to a concave mirror to obtain a global vision system. Sometimes, different vision system designs and localization strategies are used for robots playing on the same team, which allows for more specialized robots to be built according to their roles but makes sharing data among robots more difficult.

This paper reviews three approaches to visual self-localization used in the RoboCup middle-size (F-2000) competition, describing and comparing the results they achieve in the RoboCup test-bed environment. The paper is organized as follows: a self-localization method based on an omnidirectional vision system is illustrated in Section 2, a self-localization based only on the comparison between a pair of images from a binocular vision system is proposed in Section 3, and a self-localization based on the Monte Carlo localization technique is described in Section 4. Section 5 reports some results of the above described techniques. Finally some conclusions from which possible future work can stem are drawn in Section 6.

Section snippets

Robust vision-based localization with monocular vision

A lot of successful work on self-localization in indoor or bounded environments with laser range finders has been presented in the past. As already mentioned in Section 1, these methods are very accurate, fast and robust. The main drawback in the use of laser range finders or more generally in the use of data from exclusive distance-measuring sensors is the almost complete lack of information about the objects from which the distance measurements originate. Only by extracting geometric features

Binocular vision and geometric reasoning

Unlike the method described in the previous section, that relies on data coming from an omnidirectional vision sensor, the self-localization method described in this section has been designed for use on a robot goalkeeper supplied with a vision system based on two wide-angle cameras. The fields of view of the two cameras overlap in a region of about 20° which corresponds to the center of the zone that lies immediately in front of the robot. This configuration makes it possible to obtain a

Monocular vision and vision-based Monte Carlo localization

Monte Carlo localization (MCL) has recently become a very popular framework for solving the self-localization problem in mobile robots [15], [16]. This method is very reliable and robust with respect to noise, especially if the robots are equipped with laser range finders or sonar sensors. However, in environments like the popular RoboCup domain [22], providing a laser scanner for each robot is difficult or impossible and sonar data is extremely noisy due to the highly dynamic environment.

Evaluation and discussion

The algorithm described in Section 2.1 was implemented and evaluated on the goalkeeper robot of the Attempto team Tübingen.
Calibration. Fig. 9 presents a derived mapping of the position of pixels on a radial line in the camera image where the type classification changes (e.g. from FIELD to WALL) against the distance to this point in the environment. Two different methods are applied for generating an appropriate mapping. If a robot for whose camera system the mapping is to be obtained is

Conclusions

The RoboCup competition has proved to be an interesting research platform for the study of autonomous intelligent systems which can interact with the real physical world. In particular, RoboCup is particularly challenging for vision systems, that need to integrate efficiency and accuracy to be able to deal with the real-time requirements of the dynamic and rapidly changing environment in which the competition takes place. All three systems described in the paper were extensively tested on the

Acknowledgements

The authors from the University of Ulm would like to thank all the student members of the RoboCup team The Ulm Sparrows for their great commitment and support, and G. Palm, the head of the department for his encouragement, many valuable discussions, and for providing the necessary resources. Financial support for the Ulm team was provided by the University administration and the University Society Ulm.

The projects developed at the University of Parma that are described in this paper have been

Giovanni Adorni, formerly with the University of Parma, is a Full Professor at the Department of Communications, Computer and System Science of the University of Genoa. His research interests include artificial intelligence, computer vision, robotics and soft computing. He is a member of AAAI, IAS and AI∗IA.

References (33)

  • D. Burschka, S.A. Blum, Identification of 3D reference structures for video-based localization, in: Computer Vision —...
  • K. Chung, Markov Chains with Stationary Transition Probabilities, Springer, Berlin,...
  • T. Dean, M. Boddy, An analysis of time-dependent planning, in: Proceedings of AAAI-88, St. Paul, MN AAAI Press, Menlo...
  • F. Dellaert, W. Burgard, D. Fox, S. Thrun, Using the condensation algorithm for robust, vision-based mobile robot...
  • H.S. Dulimaria et al.

    Mobile robot localization in indoor environment

    Pattern Recognition

    (1997)
  • T. Einsele, Real-time self-localization in unknown indoor environment using a panorama laser range finder, in:...
  • Cited by (43)

    • Automatic guidance of underground mining vehicles using laser sensors

      2012, Tunnelling and Underground Space Technology
      Citation Excerpt :

      The method exhibits high positioning accuracy. Nonetheless, the physical environmental factors in the underground tunnel, such as light, make actual image acquisition difficult (Beliveau et al., 1996; Adorni et al., 2001). The laser-based autonomous vehicle navigation system consists of three main pieces of hardware with supporting software:

    • MAP approach for vision-based self-localization of mobile robot

      2008, Zidonghua Xuebao/Acta Automatica Sinica
    • An AI-based Object Detection Approach for Robotic Competitions

      2023, International Conference on Electrical, Computer, Communications and Mechatronics Engineering, ICECCME 2023
    • GPS-based localization of autonomous vehicles

      2021, Autonomous Driving and Advanced Driver-Assistance Systems (ADAS): Applications, Development, Legal Issues, and Testing
    View all citing articles on Scopus

    Giovanni Adorni, formerly with the University of Parma, is a Full Professor at the Department of Communications, Computer and System Science of the University of Genoa. His research interests include artificial intelligence, computer vision, robotics and soft computing. He is a member of AAAI, IAS and AI∗IA.

    Stefano Cagnoni is an Assistant Professor at the Department of Computer Engineering of the University of Parma. His research interests include computer vision, pattern recognition, robotics and soft computing. He is a member of IEEE-CS, IAPR and AI∗IA.

    Stefan Enderle is a Ph.D. candidate at the Neural Information Processing Department at the University of Ulm. His research interests include sensor interpretation and fusion, map building and self-localization in robotics, and software development for autonomous mobile robots.

    Gerhard K. Kraetzschmar is a Senior Research Scientist and Assistant Professor at the Neural Information Processing Department at the University of Ulm. His research interests include autonomous mobile robots, neurosymbolic integration, learning and memory, and robot control and agent architectures. He is a member of IEEE, AAAI, ACM, IAS, and GI.

    Monica Mordonini is an Assistant Professor at the Department of Computer Engineering of the University of Parma. Her research interests include computer vision, robotics and soft computing.

    Michael Plagge is a Ph.D. candidate at the Department of Computer Architecture at the University of Tübingen. His research interests include multisensor fusion, fast and robust world modeling and tracking of dynamic objects with autonomous mobile robots.

    Marcus Ritter was a M.S. student in Computer Science at the University of Ulm. He is now an Engineer at Wonderbits.

    Stefan Sablatnög is a Ph.D. candidate at the Neural Information Processing Department at the University of Ulm. His research interests include spatial representation and reasoning in robotics and software development for autonomous mobile robots. He is a member of IEEE.

    Andreas Zell is a Full Professor at the Department of Computer Architecture at the University of Tübingen. His research interests cover a wide range of fields in robotics, neural networks, evolutionary computation and genetic programming, bioinformatics and multimedia-aided learning and teaching. He is a member of IEEE and GI.

    View full text