arXiv Analytics

Sign in

arXiv:1709.08524 [cs.LG]AbstractReferencesReviewsResources

Generative learning for deep networks

Boris Flach, Alexander Shekhovtsov, Ondrej Fikar

Published 2017-09-25Version 1

Learning, taking into account full distribution of the data, referred to as generative, is not feasible with deep neural networks (DNNs) because they model only the conditional distribution of the outputs given the inputs. Current solutions are either based on joint probability models facing difficult estimation problems or learn two separate networks, mapping inputs to outputs (recognition) and vice-versa (generation). We propose an intermediate approach. First, we show that forward computation in DNNs with logistic sigmoid activations corresponds to a simplified approximate Bayesian inference in a directed probabilistic multi-layer model. This connection allows to interpret DNN as a probabilistic model of the output and all hidden units given the input. Second, we propose that in order for the recognition and generation networks to be more consistent with the joint model of the data, weights of the recognition and generator network should be related by transposition. We demonstrate in a tentative experiment that such a coupled pair can be learned generatively, modelling the full distribution of the data, and has enough capacity to perform well in both recognition and generation.

Related articles: Most relevant | Search more
arXiv:1904.08050 [cs.LG] (Published 2019-04-17)
Sparseout: Controlling Sparsity in Deep Networks
arXiv:1602.02644 [cs.LG] (Published 2016-02-08)
Generating Images with Perceptual Similarity Metrics based on Deep Networks
arXiv:2010.11029 [cs.LG] (Published 2020-10-21)
Learning Curves for Analysis of Deep Networks