| Real-time 3-D Motion Estimation | |
|
People
|
A. Chiuso, P. Favaro, X. Feng, H. Jin, P. Perona, S. Soatto |
|
References
|
|
|
Code
|
Real-time code for shape and motion estimation (zipped
Visual C++6.0 workspace)
For academic purpose only. |
|
Synopsis
|
We are in the process of building a real-time vision
system for use as a flexible sensor for intelligent control applications
in complex and unknown environments. The system is envisioned to support
a variety of applications in different disciplines ranging from vehicle
navigation and exploration in unknown and unstructured environments to
the realization of autonomous intelligent robotic agents and real-time
special effects editing; it will track a large number of point and line
features in the 2D image plane; it will include a recursive estimator for
calculating the position of these features in 3D space, and feedback loops
from this estimate to the 2D image trackers. We can currently track about
30 point-features and estimate their position and motion in 3D at nearly
frame-rate. In order to achieve a considerably higher performance
much attention will be devoted to the design of hardware-implementable
robust optical flow and feature tracking algorithms. The appearance/disappearance
of features and correspondence across frames will be handled via feedback
loops from the 3D estimator.
Despite the wealth of algorithms to estimate 3-D structure
from image motion (SFM), none of them has proven effective on real-world
sequences without user intervention. While the geometry of SFM is by now
fairly well understood, the role of noise has been studied only superficially,
and the interplay between geometry and
Real scenes are rarely static. Therefore, the problem of segmenting a visual scene into portions which move according to the same motion model is extremely important towards the feasibility of vision as a sensor for control systems operating in non-trivial environments. At this stage, various consistency criteria will be tested: (i) metric (norm of the estimation residual), (ii) graph-theoretic (partitioning and normalized cuttings) and (iii) statistical. The interplay between image-plane 2D segmentation, and 3D motion segmentation will also be investigated. The integration of the feature tracking, structure from motion, and segmentation modules into the complete system presents some novel issues that have both practical and theoretical relevance. The feed-forward of local inconsistency in the feature tracking provides hypotheses for the segmentation. Predictions in the segmentation (high-level) are fed back to the feature selection (low-level) to both validate the measurements, and to guide the selection of new features for robustness and speed. Information is also fed to the structure from motion module. The overall system will thus implement a complex control system where information is represented at different levels of granularity. |