Simultaneous localisation and mapping using a single camera becomes difficult when erratic motions violate predictive motion models. This problem needs to be addressed when visual SLAM algorithms are transferred from robots or mobile vehicles onto hand-held or wearable devices. In this paper we describe a novel SLAM extension to a camera localisation algorithm based on particle filtering which provides resilience to erratic motion. The mapping component is based on auxiliary unscented Kalman filters coupled to the main particle filter via measurement covariances. This coupling allows the system to survive unpredictable motions such as camera shake, and enables a return to full SLAM operation once normal motion resumes. We present results demonstrating the effectiveness of the approach when operating within a desktop environment.
|Translated title of the contribution||Real-time Visual SLAM with Resilience to Erratic Motion|
|Title of host publication||IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 17-22 June 2006|
|Pages||1244 - 1249|
|Number of pages||6|
|Publication status||Published - Jun 2006|