Conditional Image Synthesis with Diffusion Models: A Survey
Conditional image synthesis based on user-specified requirements is a key component in creating complex visual content. In recent years, diffusion-based generative modeling has become a highly effective way for conditional image synthesis, leading to exponential growth in the literature. However, th...
Gespeichert in:
Hauptverfasser: | , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Conditional image synthesis based on user-specified requirements is a key
component in creating complex visual content. In recent years, diffusion-based
generative modeling has become a highly effective way for conditional image
synthesis, leading to exponential growth in the literature. However, the
complexity of diffusion-based modeling, the wide range of image synthesis
tasks, and the diversity of conditioning mechanisms present significant
challenges for researchers to keep up with rapid developments and understand
the core concepts on this topic. In this survey, we categorize existing works
based on how conditions are integrated into the two fundamental components of
diffusion-based modeling, i.e., the denoising network and the sampling process.
We specifically highlight the underlying principles, advantages, and potential
challenges of various conditioning approaches in the training, re-purposing,
and specialization stages to construct a desired denoising network. We also
summarize six mainstream conditioning mechanisms in the essential sampling
process. All discussions are centered around popular applications. Finally, we
pinpoint some critical yet still open problems to be solved in the future and
suggest some possible solutions. Our reviewed works are itemized at
https://github.com/zju-pi/Awesome-Conditional-Diffusion-Models. |
---|---|
DOI: | 10.48550/arxiv.2409.19365 |