Bibliographic Metadata

Title
Implementing a Time-of-Flight Camera Interface for Visual Simultaneous Localization and Mapping / von Peter Jojic
AuthorJojic, Peter
CensorVincze, Markus ; Gemeiner, Peter
Published2008
DescriptionVI, 88 Bl. : Ill., graph. Darst.
Institutional NoteWien, Techn. Univ., Dipl.-Arb., 2008
LanguageEnglish
Document typeThesis (Diplom)
Keywords (DE)Autonome mobile Roboter / Simultane Lokalisierung und Kartenerstellung / Navigation / 3D Kamera / 3D Sensor
Keywords (EN)Autonomous mobile robots / Simultaneous Localization and Mapping / SLAM / 3D scene analysis / tracking / 3D camera / Time-of-Flight principle / range-imaging camera / 3D sensor
URNurn:nbn:at:at-ubtuw:1-24923 Persistent Identifier (URN)
Restriction-Information
 The work is publicly available
Files
Implementing a Time-of-Flight Camera Interface for Visual Simultaneous Localization and Mapping [7.22 mb]
Links
Reference
Classification
Abstract (German)

x.

Abstract (English)

To navigate successfully in an unknown environment, mobile robots have to know their location, and they need a map of the scene.

These two necessities cannot be separated and for navigation purposes they have to be solved simultaneously. The combination of these tasks is known within the robotics community as Simultaneous Localization and Mapping (SLAM).

Different sensors can be used to solve SLAM, but we think that a camera is the most appealing option, this is because it provides dense information content. Using the standard single perspective-projective camera as the only SLAM sensor has two major disadvantages. First, the depth information is immediately lost. To estimate the robot's location and positions' of scene landmarks, the camera has to move and perceive the environment from several different views. Second, the features lying at occlusion boundaries can not be distinctively rejected. However, false features can cause SLAM to collapse.

In this thesis, a recently developed Time-of-Flight (ToF) camera is used as the only sensor input for SLAM. The ToF sensor provides 2D images as the standard perspective-projective camera, but it can also measure the position of the scene features directly. Presented in this work is a new interface for a vision SLAM framework, which incorporates ToF sensor readings in real-time. However, the ToF cameras suffer from several noise effects, e.g. scattering, mixed pixels etc. We present how these various noise effects influence the previously mentioned localization and mapping problem.

Initially the experimental results for the selected vision SLAM framework using the ToF camera performed well, when enough near distant features have been available. In case new features were not detectable, SLAM usually gets instable or lost.

To tackle the problem of false scene landmarks lying at occlusion boundaries a concept is presented. The idea of this concept is to straightforwardly use the measured 3D information to analyze the cornerness of a landmark. Simulated results show that landmarks can be identified using the analysis based on the eigen decomposition, and this can improve the real-time feature initialization.

Stats
The PDF-Document has been downloaded 49 times.