Visual Data Fusion for Objects Localization by Active Vision

13 years 5 months ago
Visual Data Fusion for Objects Localization by Active Vision
Visual sensors provide exclusively uncertain and partial knowledge of a scene. In this article, we present a suitable scene knowledge representation that makes integration and fusion of new, uncertain and partial sensor measures possible. It is based on a mixture of stochastic and set membership models. We consider that, for a large class of applications, an approximated representation is sufficient to build a preliminary map of the scene. Our approximation mainly results in ellipsoidal calculus by means of a normal assumption for stochastic laws and ellipsoidal over or inner bounding for uniform laws. These approximations allow us to build an efficient estimation process integrating visual data on line. Based on this estimation scheme, optimal exploratory motions of the camera can be automatically determined. Real time experimental results validating our approach are finally given. 1 Overview Whatever the application, a main issue for automated visual systems is to model the environme...
François Chaumette, Grégory Flandin
Added 16 Oct 2009
Updated 16 Oct 2009
Type Conference
Year 2002
Where ECCV
Authors François Chaumette, Grégory Flandin
Comments (0)