arXiv Analytics

Sign in

arXiv:2405.15006 [cs.LG]AbstractReferencesReviewsResources

Path-metrics, pruning, and generalization

Antoine Gonon, Nicolas Brisebarre, Elisa Riccietti, Rémi Gribonval

Published 2024-05-23Version 1

Analyzing the behavior of ReLU neural networks often hinges on understanding the relationships between their parameters and the functions they implement. This paper proves a new bound on function distances in terms of the so-called path-metrics of the parameters. Since this bound is intrinsically invariant with respect to the rescaling symmetries of the networks, it sharpens previously known bounds. It is also, to the best of our knowledge, the first bound of its kind that is broadly applicable to modern networks such as ResNets, VGGs, U-nets, and many more. In contexts such as network pruning and quantization, the proposed path-metrics can be efficiently computed using only two forward passes. Besides its intrinsic theoretical interest, the bound yields not only novel theoretical generalization bounds, but also a promising proof of concept for rescaling-invariant pruning.

Related articles: Most relevant | Search more
arXiv:2202.03841 [cs.LG] (Published 2022-02-08)
Width is Less Important than Depth in ReLU Neural Networks
arXiv:2306.13264 [cs.LG] (Published 2023-06-23)
FedSelect: Customized Selection of Parameters for Fine-Tuning during Personalized Federated Learning
arXiv:2112.09181 [cs.LG] (Published 2021-12-16, updated 2023-03-16)
Approximation of functions with one-bit neural networks