Investigating Generalization by Controlling Normalized Margin
Abstract
Weight norm ‖w‖ and margin γ participate in learning theory via the normalized margin γ/‖w‖. Since standard neural net optimizers do not control normalized margin, it is hard to test whether this quantity causally relates to generalization. This paper designs a series of experimental studies that explicitly control normalized margin and thereby tackle two central questions. First: does normalized margin always have a causal effect on generalization? The paper finds that no -- networks can be produced where normalized margin has seemingly no relationship with generalization, counter to the theory of Bartlett et al. (2017). Second: does normalized margin ever have a causal effect on generalization? The paper finds that yes -- in a standard training setup, test performance closely tracks normalized margin. The paper suggests a Gaussian process model as a promising explanation for this behavior.
Additional Information
© 2022 by the authors. The authors are grateful to the anonymous reviewers for their helpful comments. This material is based upon work supported by the National Science Foundation Graduate Research Fellowship under Grant No. DGE-1745301. This material was also supported by the following grants: NSF #1918865; ONR N00014-21-1-2483.Attached Files
Published - farhang22a.pdf
Submitted - 2205.03940.pdf
Supplemental Material - farhang22a-supp.zip
Files
Additional details
- Eprint ID
- 115571
- Resolver ID
- CaltechAUTHORS:20220714-212426792
- NSF Graduate Research Fellowship
- DGE-1745301
- NSF
- CCF-1918865
- Office of Naval Research (ONR)
- N00014-21-1-2483
- Created
-
2022-07-15Created from EPrint's datestamp field
- Updated
-
2023-06-02Created from EPrint's last_modified field