Notification:
We are currently experiencing intermittent issues impacting performance. We apologize for the inconvenience.
By Topic

A model for visual flow-field cueing and self-motion estimation

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$31 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

3 Author(s)
Zacharias, G.L. ; Bolt Beranek & Newman Inc., Cambridge, MA, USA ; Caglayan, A.K. ; Sinacori, J.B.

A computational model for visual flow-field cueing and self-motion estimation is developed and simulated. The model is predicated on the notion that the pilot makes noisy, sampled measurements on the spatially distributed visual flow-field surrounding him and, on the basis of these measurements, generates estimates of his own linear and angular terrain-relative velocities which optimally satisfy, in a least-squares sense, the visual kinematic flow constraints. The least-squares formulation is applicable to general observer motions and viewing geometries; it is projection-plane independent and rational in its treatment of redundant and noisy flow cues. A subsidiary but significant output of the model is an `impact time' map, and observer-centered spatially scaled replica of the viewed surface. Simulations are presented to demonstrate the parametric sensitivity and ability to model relevant human visual performance data. Preliminary simulation results of human visual performance are encouraging.

Published in:

Systems, Man and Cybernetics, IEEE Transactions on  (Volume:SMC-15 ,  Issue: 3 )