Discriminative Clustering by Regularized Information Maximization
Abstract
Is there a principled way to learn a probabilistic discriminative classifier from an unlabeled data set? We present a framework that simultaneously clusters the data and trains a discriminative classifier. We call it Regularized Information Maximization (RIM). RIM optimizes an intuitive information-theoretic objective function which balances class separation, class balance and classifier complexity. The approach can flexibly incorporate different likelihood functions, express prior assumptions about the relative size of different classes and incorporate partial labels for semi-supervised learning. In particular, we instantiate the framework to unsupervised, multi-class kernelized logistic regression. Our empirical evaluation indicates that RIM outperforms existing methods on several real data sets, and demonstrates that RIM is an effective model selection method.
Additional Information
©2010 Neural Information Processing Systems. We thank Alex Smola for helpful comments and discussion, and Thanos Siapas for providing the neural tetrode data. This research was partially supported by NSF grant IIS-0953413, a gift from Microsoft Corporation, and ONR MURI Grant N00014-06-1-0734.Attached Files
Published - 4154-discriminative-clustering-by-regularized-information-maximization.pdf
Files
Name | Size | Download all |
---|---|---|
md5:514030e9ccd0de315dd394c0116ce5a6
|
982.2 kB | Preview Download |
Additional details
- Eprint ID
- 65825
- Resolver ID
- CaltechAUTHORS:20160331-165410183
- NSF
- IIS-0953413
- Microsoft Corporation
- Office of Naval Research (ONR)
- N00014-06-1-0734
- Created
-
2016-04-01Created from EPrint's datestamp field
- Updated
-
2020-03-09Created from EPrint's last_modified field
- Series Name
- Advances in Neural Information Processing Systems
- Series Volume or Issue Number
- 22