{ "id": "1708.02511", "version": "v1", "published": "2017-08-08T15:01:55.000Z", "updated": "2017-08-08T15:01:55.000Z", "title": "Adversarial Divergences are Good Task Losses for Generative Modeling", "authors": [ "Gabriel Huang", "Gauthier Gidel", "Hugo Berard", "Ahmed Touati", "Simon Lacoste-Julien" ], "comment": "10 pages, workshop paper for PADL ICML 2017 workshop", "categories": [ "cs.LG", "stat.ML" ], "abstract": "Generative modeling of high dimensional data like images is a notoriously difficult and ill-defined problem. In particular, how to evaluate a learned generative model is unclear. In this paper, we argue that adversarial learning, pioneered with generative adversarial networks (GANs), provides an interesting framework to implicitly define more meaningful task losses for unsupervised tasks, such as for generating \"visually realistic\" images. By unifying GANs and structured prediction under the framework of statistical decision theory, we put into light links between recent advances in structured prediction theory and the choice of the divergence in GANs. We argue that the insights about the notions of \"hard\" and \"easy\" to learn losses can be analogously extended to adversarial divergences. We also discuss the attractive properties of adversarial divergences for generative modeling, and perform experiments to show the importance of choosing a divergence that reflects the final task.", "revisions": [ { "version": "v1", "updated": "2017-08-08T15:01:55.000Z" } ], "analyses": { "keywords": [ "generative model", "adversarial divergences", "task losses", "high dimensional data", "perform experiments" ], "note": { "typesetting": "TeX", "pages": 10, "language": "en", "license": "arXiv", "status": "editable" } } }