Conference Paper

A probabilistic representation of LiDAR range data for efficient 3D object detection

Dept. of Electr., Comput., & Syst. Eng., Rensselaer Polytech. Inst., Troy, NY
DOI: 10.1109/CVPRW.2008.4563033 Conference: Computer Vision and Pattern Recognition Workshops, 2008. CVPRW '08. IEEE Computer Society Conference on
Source: IEEE Xplore

ABSTRACT We present a novel approach to 3D object detection in scenes scanned by LiDAR sensors, based on a probabilistic representation of free, occupied, and hidden space that extends the concept of occupancy grids from robot mapping algorithms. This scene representation naturally handles LiDAR sampling issues, can be used to fuse multiple LiDAR data sets, and captures the inherent uncertainty of the data due to occlusions and clutter. Using this model, we formulate a hypothesis testing methodology to determine the probability that given 3D objects are present in the scene. By propagating uncertainty in the original sample points, we are able to measure confidence in the detection results in a principled way. We demonstrate the approach in examples of detecting objects that are partially occluded by scene clutter such as camouflage netting.

8 Reads
  • Source
    • "Recently, there has been some investigation into object detection and recognition using 3D LIDAR data, mostly for stationary objects like buildings and trees. Some examples include object recognition in 3D LIDAR data with recurrent neural network [1], coarse-to-fine scheme for object indexing/ verification and rotationally invariant semi-local spin image features for 3D object representation [2], probabilistic representation of LIDAR range data for efficient 3D object detection [3], 3D object recognition from range images using local feature histograms [4], and DARPA URGENT object recognition based on strip-cueing [6]. However, to the best of our knowledge, little work has been done for detection and recognition of moving objects (i.e., persons and vehicles) in motion 3D data. "
    [Show abstract] [Hide abstract]
    ABSTRACT: This paper presents methods for 3D object detection and multi-object (or multi-agent) behavior recognition using a sequence of 3D point clouds of a scene taken over time. This motion 3D data can be collected using different sensors and techniques such as flash LIDAR (Light Detection And Ranging), stereo cameras, time-of-flight cameras, or spatial phase imaging sensors. Our goal is to segment objects from the D point cloud data in order to construct tracks of multiple objects (i.e., persons and vehicles) and then classify the multi-object tracks as one of a set of known behaviors, such as “A person drives a car and gets out”. A track is a sequence of object locations changing over time and is the compact object-level information we use and obtain from the motion 3D data. Leveraging the rich structure of dynamic 3D data makes many visual learning problems better posed and more tractable. Our behavior recognition method is based on combining the Dynamic Time Warping-based behavior distances from the multiple object-level tracks using a normalized car-centric coordinate system to recognize the interactive behavior of those multiple objects. We apply our techniques for behavior recognition on data collected using a LIDAR sensor, with promising results.
    01/2011; DOI:10.1109/CVPRW.2011.5981808
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: Building a model of the environment is essential for mobile robotics. It allows the robot to reason about its surroundings and plan actions according to its intentions. To enable safe motion planning it is vital to anticipate object movements. This paper presents an improved formulation for occupancy filtering. Our approach is closely related to the Bayesian Occupancy Filter (BOF) presented in. The basic idea of occupancy filters is to represent the environment as a 2-dimensional grid of cells holding information about their state of occupancy and velocity. To improve the accuracy of predictions, prior knowledge about the motion preferences is used, derived from map data that can be obtained from navigation systems. In combination with a physically accurate transition model, it is possible to estimate the environment dynamics. Experiments show that this yields reliable estimates even for occluded regions.
    Intelligent Vehicles Symposium, 2009 IEEE; 07/2009
  • [Show abstract] [Hide abstract]
    ABSTRACT: This paper proposes a novel method for reconstruction of the target three-dimensional (3D) objects based on motion vision. The main contribution of this paper is twofold: (1) we build a new mathematical model to express the position relationship of the camera, the image plane, the ground plane and the target 3D objects at different times. (2) We propose and implement a 3D reconstruction algorithm refer to the mathematical model using geometric theory. With the result of that, we know the distance from the camera to the objects and the height of the objects in the real world. The experimental indicates have shown the accuracy of the approach presented in this paper.
Show more


8 Reads
Available from