Welcome to the new version of CaltechAUTHORS. Login is currently restricted to library staff. If you notice any issues, please email coda@library.caltech.edu
Published June 2000 | Published + Accepted Version
Journal Article Open

A saliency-based search mechanism for overt and covert shifts of visual attention

Abstract

Most models of visual search, whether involving overt eye movements or covert shifts of attention, are based on the concept of a saliency map, that is, an explicit two-dimensional map that encodes the saliency or conspicuity of objects in the visual environment. Competition among neurons in this map gives rise to a single winning location that corresponds to the next attended target. Inhibiting this location automatically allows the system to attend to the next most salient location. We describe a detailed computer implementation of such a scheme, focusing on the problem of combining information across modalities, here orientation, intensity and color information, in a purely stimulus-driven manner. The model is applied to common psychophysical stimuli as well as to a very demanding visual search task. Its successful performance is used to address the extent to which the primate visual system carries out visual search via one or more such saliency maps and how this can be tested.

Additional Information

Received 27 May 1999; received in revised form 19 July 1999. © 2000 Elsevier Science Ltd. We thank Dr Toet from the TNO Human Factors Research Institute, The Netherlands, for providing us with the database of military images and human search times on these images. This research was supported by NSF-ERC, NIMH and ONR.

Attached Files

Published - 1-s2.0-S0042698999001637-main.pdf

Accepted Version - 188.pdf

Accepted Version - 188.ps

Files

188.pdf
Files (33.3 MB)
Name Size Download all
md5:01c1793c188dab88a225c69a45dce4c6
9.8 MB Preview Download
md5:5abc42c2bc710eafa64a2b5c7f34313a
20.6 MB Download
md5:98be9da47d02d557fbdd57a10fc326fe
3.0 MB Preview Download

Additional details

Created:
September 15, 2023
Modified:
October 23, 2023