arXiv Analytics

Sign in

arXiv:2306.17560 [cs.LG]AbstractReferencesReviewsResources

Class-Incremental Learning using Diffusion Model for Distillation and Replay

Quentin Jodelet, Xin Liu, Yin Jun Phua, Tsuyoshi Murata

Published 2023-06-30Version 1

Class-incremental learning aims to learn new classes in an incremental fashion without forgetting the previously learned ones. Several research works have shown how additional data can be used by incremental models to help mitigate catastrophic forgetting. In this work, following the recent breakthrough in text-to-image generative models and their wide distribution, we propose the use of a pretrained Stable Diffusion model as a source of additional data for class-incremental learning. Compared to competitive methods that rely on external, often unlabeled, datasets of real images, our approach can generate synthetic samples belonging to the same classes as the previously encountered images. This allows us to use those additional data samples not only in the distillation loss but also for replay in the classification loss. Experiments on the competitive benchmarks CIFAR100, ImageNet-Subset, and ImageNet demonstrate how this new approach can be used to further improve the performance of state-of-the-art methods for class-incremental learning on large scale datasets.

Related articles: Most relevant | Search more
arXiv:2211.13449 [cs.LG] (Published 2022-11-24)
Fast Sampling of Diffusion Models via Operator Learning
arXiv:2409.09569 [cs.LG] (Published 2024-09-15)
Bias Begets Bias: The Impact of Biased Embeddings on Diffusion Models
arXiv:2309.13415 [cs.LG] (Published 2023-09-23)
Dream the Impossible: Outlier Imagination with Diffusion Models