Conference Paper

A variational framework for image segmentation combining motion estimation and shape regularization

Dept. of Comput. Sci., Univ. of California, Los Angeles, CA, USA
DOI: 10.1109/CVPR.2003.1211337 Conference: Computer Vision and Pattern Recognition, 2003. Proceedings. 2003 IEEE Computer Society Conference on, Volume: 1
Source: DBLP


Based on a geometric interpretation of the optic flow constraint equation, we propose a conditional probability on the spatio-temporal image gradient. We consistently derive a variational approach for the segmentation of the image domain into regions of homogeneous motion. The proposed energy functional extends the Mumford-Shah functional from gray value segmentation to motion segmentation. It depends on the spatio-temporal image gradient calculated from only two consecutive images of an image sequence. Moreover, it depends on motion vectors for a set of regions and a boundary separating these regions. In contrast to most alternative approaches, the problems of motion estimation and motion segmentation are jointly solved by minimizing a single functional. Numerical evaluation with both explicit and implicit (level set based) representations of the boundary shows the strengths and limitations of our approach.

Full-text preview

Available from:
  • Source
    • "Other methods use motion projection to link segments, i.e., the position of a segment in a future frame is estimated from its current position and motion features [6] [7] [8] [9]. Cremers (2003) tracked motion segments by jointly solving motion segmentation and motion estimation by minimizing a single energy functional [10]. "
    [Show abstract] [Hide abstract]
    ABSTRACT: Model-free tracking is important for solving tasks such as moving-object tracking and action recognition in cases where no prior object knowledge is available. For this purpose, we extend the concept of spatially synchronous dynamics in spin-lattice models to the spatiotemporal domain to track segments within an image sequence. The method is related to synchronization processes in neural networks and based on superparamagnetic clustering of data. Spin interactions result in the formation of clusters of correlated spins, providing an automatic labeling of corresponding image regions. The algorithm obeys detailed balance. This is an important property as it allows for consistent spin-transfer across subsequent frames, which can be used for segment tracking. Therefore, in the tracking process the correct equilibrium will always be found, which is an important advance as compared with other more heuristic tracking procedures. In the case of long image sequences, i.e., movies, the algorithm is augmented with a feedback mechanism, further stabilizing segment tracking.
    Sensors 11/2009; 9(11). DOI:10.3390/s91109355 · 2.25 Impact Factor
  • Source
    • "With the advantage of being implicit, intrinsic and parameter-free, level sets track moving interfaces through either model-free (Paragios and Deriche, 2002) or either model-based (Cremers, 2003) methods. "
    [Show abstract] [Hide abstract]
    ABSTRACT: In this paper, we propose a novel mathematical model to address segmentation from an aerial and satellite imagery. Segmentation is performed through the consistent recovery of the zero iso-surfaces of a level set function towards image's foreground (desired object for extraction) and background discrimination. The level set method is embedded in a geodesic active contours variational formulation. Geodesic active contours technique is an advancement of the classical snakes and active contours that can handle the limitation to change their topology. The proposed mathematical model consists of a functional with two basic energies; one that forms a region-based energetic module for the evolving interface in the level set space and an another one that takes into account the a priori knowledge of the geometry of the desired for extraction objects. With such a functional shape information is embedded into the level set based segmentation scheme. Shape information can cope with missing or misleading information in the input images due to noise, clutter and occlusion. The shape priors were developed to segment objects of familiar shape in a given image. Promising results demonstrate the potentials of our approach, like in cases where the desired object for extraction were buildings.
  • Source
    • "Recently some research works have been done on the motion based image segmentation [4] [6] [12]. These kinds of work do the optical flow estimation and image segmentation simultaneously. "
    [Show abstract] [Hide abstract]
    ABSTRACT: In this paper, a novel algorithm is proposed for the vision-based object tracking by autonomous vehicles. To estimate the velocity of the tracked object, the algorithm fuses the information captured by the vehicle’s on-board sensors such as the cameras and inertial motion sensors. Optical flow vectors, color features, stereo pair disparities are used as optical features while the vehicle’s inertial measurements are used to determine the cameras’ motion. The algorithm determines the velocity and position of the target in the world coordinate which are then tracked by the vehicle. In order to formulate this tracking algorithm, it is necessary to use a proper model which describes the dynamic information of the tracked object. However due to the complex nature of the moving object, it is necessary to have robust and adaptive dynamic models. Here, several simple and basic linear dynamic models are selected and combined to approximate the unpredictable, complex or highly nonlinear dynamic properties of the moving target. With these basic linear dynamic models, a detailed description of the three-dimensional (3D) target tracking scheme using the Interacting Multiple Models (IMM) along with an Extended Kalman Filter is presented. The final state of the target is estimated as a weighted combination of the outputs from each different dynamic model. Performance of the proposed fusion based IMM tracking algorithm is demonstrated through extensive experimental results.
    Computer Vision and Image Understanding 01/2008; 109(1-109):1-21. DOI:10.1016/j.cviu.2006.12.001 · 1.54 Impact Factor
Show more