Human visual object categorization can be described by models with low memory capacity
Abstract
Studies of high-level models of visual object categorization have left unresolved issues of neurobiological relevance, including how features are extracted from the image and the role played by memory capacity in categorization performance. We compared the ability of a comprehensive set of models to match the categorization performance of human observers while explicitly accounting for the models' numbers of free parameters. The most successful models did not require a large memory capacity, suggesting that a sparse, abstracted representation of category properties may underlie categorization performance. This type of representation––different from classical prototype abstraction––could also be extracted directly from two-dimensional images via a biologically plausible early-vision model, rather than relying on experimenter-imposed features.
Additional Information
Received 31 October 2001; received in revised form 20 August 2002. c2003 Elsevier Ltd. This work was supported by a Predoctoral Fellowship from the Howard Hughes Medical Institute to R.J. Peters. Additional support was provided by the Engineering Research Centers Program of the National Science Foundation under Award Number EEC-9402726, by the NIMH and by the W.M. Keck Foundation Fund for Discovery in Basic Medical Research at Caltech.Attached Files
Published - 450.pdf
Files
Name | Size | Download all |
---|---|---|
md5:41e9705d721bb65b9482446af572e802
|
427.7 kB | Preview Download |
Additional details
- Eprint ID
- 40490
- Resolver ID
- CaltechAUTHORS:20130816-103220181
- Howard Hughes Medical Institute (HHMI) Predoctoral Fellowship
- Engineering Research Centers Programs of the NSF
- EEC-9402726
- NIMH
- W.M. Keck Foundation Fund for Discovery in Basic Medical Research
- Created
-
2008-01-11Created from EPrint's datestamp field
- Updated
-
2021-11-09Created from EPrint's last_modified field
- Caltech groups
- Koch Laboratory (KLAB)