Welcome to the new version of CaltechAUTHORS. Login is currently restricted to library staff. If you notice any issues, please email coda@library.caltech.edu
Published August 17, 2022 | Submitted
Report Open

Engineering flexible machine learning systems by traversing functionally invariant paths in weight space

Abstract

Deep neural networks achieve human-like performance on a variety of perceptual and decision-making tasks. However, networks perform poorly when confronted with changing tasks or goals, and broadly fail to match the flexibility and robustness of human intelligence. Here, we develop a mathematical and algorithmic framework that enables flexible and continuous training of neural networks on a range of objectives by constructing path connected sets of networks that achieve equivalent functional performance on a given machine learning task. We view the weight space of a neural network as a curved Riemannian manifold and move a network along a functionally invariant path in weight space while searching for networks that satisfy secondary objectives. A path-sampling algorithm trains computer vision and natural language processing networks with millions of weight parameters to learn a series of classification tasks without performance loss while accommodating secondary objectives including network sparsification, incremental task learning, and increased adversarial robustness. Broadly, we conceptualize a neural network as a mathematical object that can be iteratively transformed into distinct configurations by the path-sampling algorithm to define a sub-manifold of networks that can be harnessed to achieve user goals.

Additional Information

Attribution-NonCommercial-NoDerivatives 4.0 International (CC BY-NC-ND 4.0).

Attached Files

Submitted - 2205.00334.pdf

Files

2205.00334.pdf
Files (3.4 MB)
Name Size Download all
md5:5fd3df01b82cc03e0d89c14b860d1a66
3.4 MB Preview Download

Additional details

Created:
August 20, 2023
Modified:
December 13, 2023