Context Embedding Networks
- Creators
-
Kim, Kun Ho
-
Mac Aodha, Oisin
-
Perona, Pietro
Abstract
Low dimensional embeddings that capture the main variations of interest in collections of data are important for many applications. One way to construct these embeddings is to acquire estimates of similarity from the crowd. Similarity is a multi-dimensional concept that varies from individual to individual. However, existing models for learning crowd embeddings typically make simplifying assumptions such as all individuals estimate similarity using the same criteria, the list of criteria is known in advance, or that the crowd workers are not influenced by the data that they see. To overcome these limitations we introduce Context Embedding Networks (CENs). In addition to learning interpretable embeddings from images, CENs also model worker biases for different attributes along with the visual context i.e. the attributes highlighted by a set of images. Experiments on three noisy crowd annotated datasets show that modeling both worker bias and visual context results in more interpretable embeddings compared to existing approaches.
Additional Information
© 2018 IEEE. We thank Google for supporting the Visipedia project and AWS Research Credits for their donation.Attached Files
Accepted Version - f5e42ed707dc068a9c95b245d9834855c023.pdf
Files
Name | Size | Download all |
---|---|---|
md5:7bb72d5d3a24e56e2a86fdd359905f10
|
5.7 MB | Preview Download |
Additional details
- Eprint ID
- 87328
- Resolver ID
- CaltechAUTHORS:20180622-113357959
- Amazon Web Services
- Created
-
2018-06-22Created from EPrint's datestamp field
- Updated
-
2021-11-15Created from EPrint's last_modified field