arXiv:1903.07378 [cs.LG]AbstractReferencesReviewsResources
On-line learning dynamics of ReLU neural networks using statistical physics techniques
Published 2019-03-18Version 1
We introduce exact macroscopic on-line learning dynamics of two-layer neural networks with ReLU units in the form of a system of differential equations, using techniques borrowed from statistical physics. For the first experiments, numerical solutions reveal similar behavior compared to sigmoidal activation researched in earlier work. In these experiments the theoretical results show good correspondence with simulations. In ove-rrealizable and unrealizable learning scenarios, the learning behavior of ReLU networks shows distinctive characteristics compared to sigmoidal networks.
Comments: Accepted contribution: ESANN 2019, 6 pages European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning 2019
Related articles: Most relevant | Search more
arXiv:1809.07122 [cs.LG] (Published 2018-09-19)
Capacity Control of ReLU Neural Networks by Basis-path Norm
arXiv:1903.02237 [cs.LG] (Published 2019-03-06)
Positively Scale-Invariant Flatness of ReLU Neural Networks
arXiv:2202.03841 [cs.LG] (Published 2022-02-08)
Width is Less Important than Depth in ReLU Neural Networks