{ "id": "1509.08745", "version": "v1", "published": "2015-09-29T13:32:30.000Z", "updated": "2015-09-29T13:32:30.000Z", "title": "Compression of Deep Neural Networks on the Fly", "authors": [ "Guillaume Soulié", "Vincent Gripon", "Maëlys Robert" ], "comment": "Under review as a conference paper at SPL 2015. arXiv admin note: text overlap with arXiv:1412.6115 by other authors", "categories": [ "cs.LG", "cs.CV", "cs.NE" ], "abstract": "Because of their performance, deep neural networks are increasingly used for object recognition. They are particularly attractive because of their ability to 'absorb' great quantities of labeled data through millions of parameters. However, as the accuracy and the model sizes increase, so does the memory requirements of the classifiers. This prohibits their usage on resource limited hardware, including cell phones or other embedded devices. We introduce a novel compression method for deep neural networks that performs during the learning phase. It consists in adding an extra regularization term to the cost function of fully-connected layers. We combine this method with Product Quantization (PQ) of the trained weights for higher savings in memory and storage consumption. We evaluate our method on two data sets (MNIST and CIFAR10), on which we achieve significantly larger compression than state-of-the-art methods.", "revisions": [ { "version": "v1", "updated": "2015-09-29T13:32:30.000Z" } ], "analyses": { "keywords": [ "deep neural networks", "extra regularization term", "novel compression method", "achieve significantly larger compression", "model sizes increase" ], "tags": [ "conference paper" ], "note": { "typesetting": "TeX", "pages": 0, "language": "en", "license": "arXiv", "status": "editable", "adsabs": "2015arXiv150908745S" } } }