dc.contributor.author |
Alatise, Mary B.
|
|
dc.contributor.author |
Hancke, Gerhard P.
|
|
dc.date.accessioned |
2017-11-15T12:55:23Z |
|
dc.date.available |
2017-11-15T12:55:23Z |
|
dc.date.issued |
2017-09-21 |
|
dc.description.abstract |
Using a single sensor to determine the pose estimation of a device cannot give accurate
results. This paper presents a fusion of an inertial sensor of six degrees of freedom (6-DoF) which
comprises the 3-axis of an accelerometer and the 3-axis of a gyroscope, and a vision to determine a
low-cost and accurate position for an autonomous mobile robot. For vision, a monocular vision-based
object detection algorithm speeded-up robust feature (SURF) and random sample consensus
(RANSAC) algorithms were integrated and used to recognize a sample object in several images
taken. As against the conventional method that depend on point-tracking, RANSAC uses an iterative
method to estimate the parameters of a mathematical model from a set of captured data which
contains outliers. With SURF and RANSAC, improved accuracy is certain; this is because of their
ability to find interest points (features) under different viewing conditions using a Hessain matrix.
This approach is proposed because of its simple implementation, low cost, and improved accuracy.
With an extended Kalman filter (EKF), data from inertial sensors and a camera were fused to estimate
the position and orientation of the mobile robot. All these sensors were mounted on the mobile robot
to obtain an accurate localization. An indoor experiment was carried out to validate and evaluate the
performance. Experimental results show that the proposed method is fast in computation, reliable
and robust, and can be considered for practical applications. The performance of the experiments
was verified by the ground truth data and root mean square errors (RMSEs). |
en_ZA |
dc.description.department |
Electrical, Electronic and Computer Engineering |
en_ZA |
dc.description.librarian |
am2017 |
en_ZA |
dc.description.sponsorship |
The National Researcher Foundation grant funded by the South African government in collaboration with the University of Pretoria. |
en_ZA |
dc.description.uri |
http://www.mdpi.com/journal/sensors |
en_ZA |
dc.identifier.citation |
Alatise, M.B. & Hancke, G.P. 2017, 'Pose estimation of a mobile robot based on fusion of IMU data and vision data using an extended Kalman filter', Sensors, vol. 17, no. 10, art. 2164, pp. 1-22. |
en_ZA |
dc.identifier.isbn |
10.3390/s17102164 |
|
dc.identifier.issn |
1424-8220 (onlne) |
|
dc.identifier.uri |
http://hdl.handle.net/2263/63182 |
|
dc.language.iso |
en |
en_ZA |
dc.publisher |
MDPI Publishing |
en_ZA |
dc.rights |
© 2017 by the authors; licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution license (http://creativecommons.org/licenses/by/4.0/). |
en_ZA |
dc.subject |
Pose estimation |
en_ZA |
dc.subject |
Mobile robot |
en_ZA |
dc.subject |
Inertial sensors |
en_ZA |
dc.subject |
Vision |
en_ZA |
dc.subject |
Object |
en_ZA |
dc.subject |
Speeded-up robust feature (SURF) |
en_ZA |
dc.subject |
Random sample consensus (RANSAC) |
en_ZA |
dc.subject |
Extended Kalman filter (EKF) |
en_ZA |
dc.subject |
Kalman filters |
en_ZA |
dc.subject |
Position and orientations |
en_ZA |
dc.subject |
Object detection algorithms |
en_ZA |
dc.subject |
Autonomous mobile robot |
en_ZA |
dc.subject |
Robots |
en_ZA |
dc.subject |
Object detection |
en_ZA |
dc.subject |
Mean square error |
en_ZA |
dc.subject |
Iterative methods |
en_ZA |
dc.subject |
Inertial navigation systems |
en_ZA |
dc.subject |
Estimation |
en_ZA |
dc.subject |
Degrees of freedom (mechanics) |
en_ZA |
dc.subject |
Bandpass filters |
en_ZA |
dc.title |
Pose estimation of a mobile robot based on fusion of IMU data and vision data using an extended Kalman filter |
en_ZA |
dc.type |
Article |
en_ZA |