Welcome to the new version of CaltechAUTHORS. Login is currently restricted to library staff. If you notice any issues, please email coda@library.caltech.edu
Published March 2016 | Supplemental Material
Journal Article Open

A systematic comparison between visual cues for boundary detection

Abstract

The detection of object boundaries is a critical first step for many visual processing tasks. Multiple cues (we consider luminance, color, motion and binocular disparity) available in the early visual system may signal object boundaries but little is known about their relative diagnosticity and how to optimally combine them for boundary detection. This study thus aims at understanding how early visual processes inform boundary detection in natural scenes. We collected color binocular video sequences of natural scenes to construct a video database. Each scene was annotated with two full sets of ground-truth contours (one set limited to object boundaries and another set which included all edges). We implemented an integrated computational model of early vision that spans all considered cues, and then assessed their diagnosticity by training machine learning classifiers on individual channels. Color and luminance were found to be most diagnostic while stereo and motion were least. Combining all cues yielded a significant improvement in accuracy beyond that of any cue in isolation. Furthermore, the accuracy of individual cues was found to be a poor predictor of their unique contribution for the combination. This result suggested a complex interaction between cues, which we further quantified using regularization techniques. Our systematic assessment of the accuracy of early vision models for boundary detection together with the resulting annotated video dataset should provide a useful benchmark towards the development of higher-level models of visual processing.

Additional Information

© 2016 Elsevier Ltd. Under an Elsevier user license. Received 21 August 2014, Revised 17 November 2015, Accepted 17 November 2015, Available online 2 March 2016. This work was supported by ONR grant (N000141110743), DARPA young faculty award (N66001-14-1-4037) and NSF early career award (IIS-1252951) to TS. Additional support was provided by the Center for Computation and Visualization (CCV) at Brown University.

Attached Files

Supplemental Material - 1-s2.0-S0042698915003685-mmc1.pdf

Files

1-s2.0-S0042698915003685-mmc1.pdf
Files (16.2 MB)
Name Size Download all
md5:a0f33a338178a401a3ec66889fb80280
16.2 MB Preview Download

Additional details

Created:
August 22, 2023
Modified:
October 17, 2023