Welcome to the new version of CaltechAUTHORS. Login is currently restricted to library staff. If you notice any issues, please email coda@library.caltech.edu
Published May 10, 2021 | Submitted
Report Open

Matched sample selection with GANs for mitigating attribute confounding

Abstract

Measuring biases of vision systems with respect to protected attributes like gender and age is critical as these systems gain widespread use in society. However, significant correlations between attributes in benchmark datasets make it difficult to separate algorithmic bias from dataset bias. To mitigate such attribute confounding during bias analysis, we propose a matching approach that selects a subset of images from the full dataset with balanced attribute distributions across protected attributes. Our matching approach first projects real images onto a generative adversarial network (GAN)'s latent space in a manner that preserves semantic attributes. It then finds image matches in this latent space across a chosen protected attribute, yielding a dataset where semantic and perceptual attributes are balanced across the protected attribute. We validate projection and matching strategies with qualitative, quantitative, and human annotation experiments. We demonstrate our work in the context of gender bias in multiple open-source facial-recognition classifiers and find that bias persists after removing key confounders via matching. Code and documentation to reproduce the results here and apply the methods to new data is available at https://github.com/csinva/matching-with-gans.

Additional Information

Attribution 4.0 International (CC BY 4.0). Code and documentation to reproduce the results here and apply the methods to new data is available at github.com/csinva/matching-withgans. We additionally release all collected annotations and computed intermediate outputs. The authors would like to thank Luis Goncalves for very useful discussions and comments. Additionaly, we would like to thank De'Aira Bryant, Nashlie Sephus, Wei Xia, Yuanjun Xiong and the rest of the faces team and fairness team at Amazon for thoughtful feedback and discussions.

Attached Files

Submitted - 2103.13455.pdf

Files

2103.13455.pdf
Files (24.1 MB)
Name Size Download all
md5:940f82170c7e318207b0f7a741e615e1
24.1 MB Preview Download

Additional details

Created:
August 20, 2023
Modified:
October 23, 2023