arXiv Analytics

Sign in

arXiv:2403.13807 [cs.CV]AbstractReferencesReviewsResources

Editing Massive Concepts in Text-to-Image Diffusion Models

Tianwei Xiong, Yue Wu, Enze Xie, Yue Wu, Zhenguo Li, Xihui Liu

Published 2024-03-20Version 1

Text-to-image diffusion models suffer from the risk of generating outdated, copyrighted, incorrect, and biased content. While previous methods have mitigated the issues on a small scale, it is essential to handle them simultaneously in larger-scale real-world scenarios. We propose a two-stage method, Editing Massive Concepts In Diffusion Models (EMCID). The first stage performs memory optimization for each individual concept with dual self-distillation from text alignment loss and diffusion noise prediction loss. The second stage conducts massive concept editing with multi-layer, closed form model editing. We further propose a comprehensive benchmark, named ImageNet Concept Editing Benchmark (ICEB), for evaluating massive concept editing for T2I models with two subtasks, free-form prompts, massive concept categories, and extensive evaluation metrics. Extensive experiments conducted on our proposed benchmark and previous benchmarks demonstrate the superior scalability of EMCID for editing up to 1,000 concepts, providing a practical approach for fast adjustment and re-deployment of T2I diffusion models in real-world applications.

Comments: Project page: https://silentview.github.io/EMCID/ . Code: https://github.com/SilentView/EMCID
Categories: cs.CV, cs.LG
Related articles: Most relevant | Search more
arXiv:2211.01324 [cs.CV] (Published 2022-11-02)
eDiffi: Text-to-Image Diffusion Models with an Ensemble of Expert Denoisers
arXiv:2410.06025 [cs.CV] (Published 2024-10-08, updated 2024-10-10)
Sparse Repellency for Shielded Generation in Text-to-image Diffusion Models
arXiv:2407.18658 [cs.CV] (Published 2024-07-26)
Adversarial Robustification via Text-to-Image Diffusion Models