InvDiff: Invariant Guidance for Bias Mitigation in Diffusion Models
As one of the most successful generative models, diffusion models have demonstrated remarkable efficacy in synthesizing high-quality images. These models learn the underlying high-dimensional data distribution in an unsupervised manner. Despite their success, diffusion models are highly data-driven...
Gespeichert in:
Hauptverfasser: | , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | As one of the most successful generative models, diffusion models have
demonstrated remarkable efficacy in synthesizing high-quality images. These
models learn the underlying high-dimensional data distribution in an
unsupervised manner. Despite their success, diffusion models are highly
data-driven and prone to inheriting the imbalances and biases present in
real-world data. Some studies have attempted to address these issues by
designing text prompts for known biases or using bias labels to construct
unbiased data. While these methods have shown improved results, real-world
scenarios often contain various unknown biases, and obtaining bias labels is
particularly challenging. In this paper, we emphasize the necessity of
mitigating bias in pre-trained diffusion models without relying on auxiliary
bias annotations. To tackle this problem, we propose a framework, InvDiff,
which aims to learn invariant semantic information for diffusion guidance.
Specifically, we propose identifying underlying biases in the training data and
designing a novel debiasing training objective. Then, we employ a lightweight
trainable module that automatically preserves invariant semantic information
and uses it to guide the diffusion model's sampling process toward unbiased
outcomes simultaneously. Notably, we only need to learn a small number of
parameters in the lightweight learnable module without altering the pre-trained
diffusion model. Furthermore, we provide a theoretical guarantee that the
implementation of InvDiff is equivalent to reducing the error upper bound of
generalization. Extensive experimental results on three publicly available
benchmarks demonstrate that InvDiff effectively reduces biases while
maintaining the quality of image generation. Our code is available at
https://github.com/Hundredl/InvDiff. |
---|---|
DOI: | 10.48550/arxiv.2412.08480 |