Welcome to the new version of CaltechAUTHORS. Login is currently restricted to library staff. If you notice any issues, please email coda@library.caltech.edu
Published February 2019 | Accepted Version
Book Section - Chapter Open

Inverse Abstraction of Neural Networks Using Symbolic Interpolation

Abstract

Neural networks in real-world applications have to satisfy critical properties such as safety and reliability. The analysis of such properties typically requires extracting information through computing pre-images of the network transformations, but it is well-known that explicit computation of pre-images is intractable. We introduce new methods for computing compact symbolic abstractions of pre-images by computing their overapproximations and underapproximations through all layers. The abstraction of pre-images enables formal analysis and knowledge extraction without affecting standard learning algorithms. We use inverse abstractions to automatically extract simple control laws and compact representations for pre-images corresponding to unsafe outputs. We illustrate that the extracted abstractions are interpretable and can be used for analyzing complex properties.

Additional Information

© 2019 Association for the Advancement of Artificial Intelligence. The work is supported by DARPA Assured Autonomy, NSF CNS-1830399 and the VeHICaL project (NSF grant #1545126).

Attached Files

Accepted Version - dgm19-aiaa.pdf

Files

dgm19-aiaa.pdf
Files (580.6 kB)
Name Size Download all
md5:0939dab5dd0f3dc910150c87c69163d5
580.6 kB Preview Download

Additional details

Created:
August 19, 2023
Modified:
October 18, 2023