{ "id": "1901.04609", "version": "v1", "published": "2019-01-15T00:04:22.000Z", "updated": "2019-01-15T00:04:22.000Z", "title": "Tightening Mutual Information Based Bounds on Generalization Error", "authors": [ "Yuheng Bu", "Shaofeng Zou", "Venugopal V. Veeravalli" ], "comment": "Submitted to ISIT 2019", "categories": [ "cs.LG", "stat.ML" ], "abstract": "A mutual information based upper bound on the generalization error of a supervised learning algorithm is derived in this paper. The bound is constructed in terms of the mutual information between each individual training sample and the output of the learning algorithm, which requires weaker conditions on the loss function, but provides a tighter characterization of the generalization error than existing studies. Examples are further provided to demonstrate that the bound derived in this paper is tighter, and has a broader range of applicability. Application to noisy and iterative algorithms, e.g., stochastic gradient Langevin dynamics (SGLD), is also studied, where the constructed bound provides a tighter characterization of the generalization error than existing results.", "revisions": [ { "version": "v1", "updated": "2019-01-15T00:04:22.000Z" } ], "analyses": { "keywords": [ "generalization error", "tightening mutual information", "stochastic gradient langevin dynamics", "tighter characterization", "learning algorithm" ], "note": { "typesetting": "TeX", "pages": 0, "language": "en", "license": "arXiv", "status": "editable" } } }