arXiv Analytics

Sign in

arXiv:2010.14495 [cs.LG]AbstractReferencesReviewsResources

Are wider nets better given the same number of parameters?

Anna Golubeva, Behnam Neyshabur, Guy Gur-Ari

Published 2020-10-27Version 1

Empirical studies demonstrate that the performance of neural networks improves with increasing number of parameters. In most of these studies, the number of parameters is increased by increasing the network width. This begs the question: Is the observed improvement due to the larger number of parameters, or is it due to the larger width itself? We compare different ways of increasing model width while keeping the number of parameters constant. We show that for models initialized with a random, static sparsity pattern in the weight tensors, network width is the determining factor for good performance, while the number of weights is secondary, as long as trainability is ensured. As a step towards understanding this effect, we analyze these models in the framework of Gaussian Process kernels. We find that the distance between the sparse finite-width model kernel and the infinite-width kernel at initialization is indicative of model performance.

Related articles: Most relevant | Search more
arXiv:2006.12467 [cs.LG] (Published 2020-06-22)
Limits to Depth Efficiencies of Self-Attention
arXiv:2203.09255 [cs.LG] (Published 2022-03-17)
On the Spectral Bias of Convolutional Neural Tangent and Gaussian Process Kernels
arXiv:2306.13264 [cs.LG] (Published 2023-06-23)
FedSelect: Customized Selection of Parameters for Fine-Tuning during Personalized Federated Learning