Published January 1, 1997
| Submitted
Technical Report
Open
Validation of Average Error Rate Over Classifiers
- Creators
- Bax, Eric
Chicago
Abstract
We examine methods to estimate the average and variance of test error rates over a set of classifiers. We begin with the process of drawing a classifier at random for each example. Given validation data, the average test error rate can be estimated as if validating a single classifier. Given the test example inputs, the variance can be computed exactly. Next, we consider the process of drawing a classifier at random and using it on all examples. Once again, the expected test error rate can be validated as if validating a single classifier. However, the variance must be estimated by validating all classifers, which yields loose or uncertain bounds.
Additional Information
© 1997 California Institute of Technology. Thanks to Zehra Cataltepe and Joseph Sill for their instructive conversations and helpful pointers. Thanks to Dr. Yaser Abu-Mostafa for teaching - the results in this paper were inspired by his class on learning theory. Thanks to Dr. Joel Franklin for advice and guidance. Also, thanks to an anonymous referee for invaluable advice on the presentation of these results.Attached Files
Submitted - CSTR97.pdf
Submitted - postscript.ps
Files
CSTR97.pdf
Files
(453.1 kB)
Name | Size | Download all |
---|---|---|
md5:7229945db552010e81e761eea4000d33
|
152.2 kB | Preview Download |
md5:55f8a5e06cdd50bbe3b20c7021f41727
|
301.0 kB | Download |
Additional details
- Eprint ID
- 26825
- Resolver ID
- CaltechCSTR:1997.cs-tr-97-17
- Created
-
2001-04-30Created from EPrint's datestamp field
- Updated
-
2019-10-03Created from EPrint's last_modified field
- Caltech groups
- Computer Science Technical Reports