Ground landmark-based localisation for mobile robots using a calibrated camera
by Lili Ma; Mohamed Alborati; Xiaohai Li
International Journal of Computational Vision and Robotics (IJCVR), Vol. 12, No. 3, 2022

Abstract: This paper discusses a non-iterative vision-based localisation scheme for an autonomous mobile robot designed to work in parking lots. The robot is equipped with a single calibrated camera. Commonly-available features such as yellow-coloured road lines are used as visual landmarks. Given the orientation of a yellow line, the orientation of the robot is computed based on the detected information of the yellow line on the image plane. The robot's position is calculated by further assuming that the position of one point on the yellow line is also known. Besides theoretical analysis, we also experimented the proposed scheme on a low-cost physical robot in a parking lot. Two issues pertaining to real-time implementation of the scheme on the robot are addressed: determination of image size that balances processing speed and feature extraction accuracy; effect of feature extraction error on the localisation accuracy.

Online publication date: Tue, 03-May-2022

The full text of this article is only available to individual subscribers or to users at subscribing institutions.

 
Existing subscribers:
Go to Inderscience Online Journals to access the Full Text of this article.

Pay per view:
If you are not a subscriber and you just want to read the full contents of this article, buy online access here.

Complimentary Subscribers, Editors or Members of the Editorial Board of the International Journal of Computational Vision and Robotics (IJCVR):
Login with your Inderscience username and password:

    Username:        Password:         

Forgotten your password?


Want to subscribe?
A subscription gives you complete access to all articles in the current issue, as well as to all articles in the previous three years (where applicable). See our Orders page to subscribe.

If you still need assistance, please email subs@inderscience.com