Conference Paper

Monocular SLAM for Visual Odometry

Catalonia Univ., Barcelona
DOI: 10.1109/WISP.2007.4447564 Conference: Intelligent Signal Processing, 2007. WISP 2007. IEEE International Symposium on
Source: IEEE Xplore

ABSTRACT The ego-motion online estimation process from a video input is often called visual odometry. Typically optical flow and structure from motion (SFM) techniques have been used for visual odometry. Monocular simultaneous localization and mapping (SLAM) techniques implicitly estimate camera ego-motion while incrementally build a map of the environment. However in monocular SLAM, when the number of features in the system state increases, the computational cost grows rapidly; consequently maintaining frame rate operation becomes impractical. In this paper monocular SLAM is proposed for map-based visual odometry. The number of features is bounded removing features dynamically from the system state, for maintaining a stable processing time. In the other hand if features are removed then previous visited sites can not be recognized, nevertheless in an odometry context this could not be a problem. A method for feature initialization and a simple method for recovery metric scale are proposed. The experimental results using real image sequences show that the scheme presented in this paper is promising.

0 Bookmarks
 · 
162 Views
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: In a visual odometry system, location of a mobile robot is automatically estimated (localized) from video. When the video is captured by an "upward" camera fixed to an indoor mobile robot, a panorama image of the ceiling (ceiling map) is generated by using a visual motion between two adjacent frames in the video. Similarly, location of another robot can be estimated on the ceiling map by using a visual motion between the current frame and the previously generated ceiling map. Under the assumption that the robot goes straight or rotates around a fixed point, there is no problem on the localization as far as the floor is flat. However, when there is debris on the floor, the estimated location contains error. In this paper, we reduce this error by utilizing visual motions in video from the "forward" camera fixed to the robot. This is a visual compensation of motions in the "upward" camera's video with those in the "forward" camera's video. It was experimentally confirmed that the maximum absolute value of the error was reduced to approximately 11%.
    Scientific research and essays 02/2011; 6:131-135. · 0.32 Impact Factor
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: For mobile robots working on giant turbine blades, pose estimation of the robots is very important in their tasks. Vision based scheme utilizes only the visual information of the robot's surrounding environment, thus becoming a good candidate scheme for pose estimation. A most important task in this scheme is to detect feature points in sucessive image frames and to match them. In this paper, an improve pose estimation algorithm based on SIFT (Scale Invariant Feature Transform) is presented considering the characteristics of local images of the turbine blades, and the pose estimation problem and condition. The improvement includes pre-subsampling the image, which reduces the computation, and bidirectional matching, which improves the precision. RANSAC (Random Sample Consensus) method is used to get a better estimation of the robot's pose. The experiment platform is built and the experimental results show the validity of the proposed algorithm.
    2009 IASTED Int. Conf. on Robotics, Telematics and Applications, Beijing, China; 10/2009
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: In the SLAM (Simultaneous Localization and Mapping) technology, an environmental map is generated by a mobile robot. When it results in failure, it is necessary to inspect the scene. This localization and browsing require transmission of video signal to a remote place. In the system in this paper, an indoor mobile robot has two cameras. One is the "upward" which captures scenery of ceiling. The other is "forward" for scenery in front of the robot. Video signals from the cameras are encoded and transmitted from the robot to a remote server. It causes a problem that data size is too huge to be transmitted. To cope with this problem, the Functionally Layered Coding (FLC) was reported. In the existing FLC, visual motions are estimated by using the rotation invariant phase only correlation (RI-POC) technique. It can estimate two kinds of motions -translation and rotation. However, it requires doubled computational complexity and many components to be transmitted. In this paper, we analyze relation between kinetic movements of a robot and visual motions observed in videos, and propose to replace RI-POC by a simple POC. It was confirmed that the proposed method reduced data size for transmission to 61.6%.
    International Journal of the Physical Sciences. 01/2011; 5:2652-2657.