Predicting human visuomotor behaviour in a driving task |
| |
Authors: | Leif Johnson Brian Sullivan Mary Hayhoe Dana Ballard |
| |
Affiliation: | 1.Department of Computer Science, University of Texas at Austin, TX, USA;2.Department of Psychology, University of Texas at Austin, TX, USA;3.Smith-Kettlewell Eye Research Institute, San Francisco, CA, USA |
| |
Abstract: | The sequential deployment of gaze to regions of interest is an integral part of human visual function. Owing to its central importance, decades of research have focused on predicting gaze locations, but there has been relatively little formal attempt to predict the temporal aspects of gaze deployment in natural multi-tasking situations. We approach this problem by decomposing complex visual behaviour into individual task modules that require independent sources of visual information for control, in order to model human gaze deployment on different task-relevant objects. We introduce a softmax barrier model for gaze selection that uses two key elements: a priority parameter that represents task importance per module, and noise estimates that allow modules to represent uncertainty about the state of task-relevant visual information. Comparisons with human gaze data gathered in a virtual driving environment show that the model closely approximates human performance. |
| |
Keywords: | visual attention eye movements reward top-down control state uncertainty |
|
|