arXiv Analytics

Sign in

arXiv:1905.11926 [cs.LG]AbstractReferencesReviewsResources

Network Deconvolution

Chengxi Ye, Matthew Evanusa, Hua He, Anton Mitrokhin, Thomas Goldstein, James A. Yorke, Cornelia Fermüller, Yiannis Aloimonos

Published 2019-05-28Version 1

Convolution is a central operation in Convolutional Neural Networks (CNNs), which applies a kernel or mask to overlapping regions shifted across the image. In this work we show that the underlying kernels are trained with highly correlated data, which leads to co-adaptation of model weights. To address this issue we propose what we call network deconvolution, a procedure that aims to remove pixel-wise and channel-wise correlations before the data is fed into each layer. We show that by removing this correlation we are able to achieve better convergence rates during model training with superior results without the use of batch normalization on the CIFAR-10, CIFAR-100, MNIST, Fashion-MNIST datasets, as well as against reference models from "model zoo" on the ImageNet standard benchmark.

Related articles: Most relevant | Search more
arXiv:1905.04967 [cs.LG] (Published 2019-05-13)
Implicit Filter Sparsification In Convolutional Neural Networks
arXiv:1708.08012 [cs.LG] (Published 2017-08-26)
Deep learning with convolutional neural networks for decoding and visualization of EEG pathology
arXiv:1812.11337 [cs.LG] (Published 2018-12-29)
Quantized Guided Pruning for Efficient Hardware Implementations of Convolutional Neural Networks