{ "id": "2306.17560", "version": "v1", "published": "2023-06-30T11:23:49.000Z", "updated": "2023-06-30T11:23:49.000Z", "title": "Class-Incremental Learning using Diffusion Model for Distillation and Replay", "authors": [ "Quentin Jodelet", "Xin Liu", "Yin Jun Phua", "Tsuyoshi Murata" ], "categories": [ "cs.LG", "cs.CV" ], "abstract": "Class-incremental learning aims to learn new classes in an incremental fashion without forgetting the previously learned ones. Several research works have shown how additional data can be used by incremental models to help mitigate catastrophic forgetting. In this work, following the recent breakthrough in text-to-image generative models and their wide distribution, we propose the use of a pretrained Stable Diffusion model as a source of additional data for class-incremental learning. Compared to competitive methods that rely on external, often unlabeled, datasets of real images, our approach can generate synthetic samples belonging to the same classes as the previously encountered images. This allows us to use those additional data samples not only in the distillation loss but also for replay in the classification loss. Experiments on the competitive benchmarks CIFAR100, ImageNet-Subset, and ImageNet demonstrate how this new approach can be used to further improve the performance of state-of-the-art methods for class-incremental learning on large scale datasets.", "revisions": [ { "version": "v1", "updated": "2023-06-30T11:23:49.000Z" } ], "analyses": { "keywords": [ "diffusion model", "distillation", "large scale datasets", "additional data samples", "generate synthetic samples" ], "note": { "typesetting": "TeX", "pages": 0, "language": "en", "license": "arXiv", "status": "editable" } } }