Neural Assets: 3D-Aware Multi-Object Scene Synthesis with Image Diffusion Models
We address the problem of multi-object 3D pose control in image diffusion models. Instead of conditioning on a sequence of text tokens, we propose to use a set of per-object representations, Neural Assets, to control the 3D pose of individual objects in a scene. Neural Assets are obtained by pooling...
Gespeichert in:
Hauptverfasser: | , , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | We address the problem of multi-object 3D pose control in image diffusion
models. Instead of conditioning on a sequence of text tokens, we propose to use
a set of per-object representations, Neural Assets, to control the 3D pose of
individual objects in a scene. Neural Assets are obtained by pooling visual
representations of objects from a reference image, such as a frame in a video,
and are trained to reconstruct the respective objects in a different image,
e.g., a later frame in the video. Importantly, we encode object visuals from
the reference image while conditioning on object poses from the target frame.
This enables learning disentangled appearance and pose features. Combining
visual and 3D pose representations in a sequence-of-tokens format allows us to
keep the text-to-image architecture of existing models, with Neural Assets in
place of text tokens. By fine-tuning a pre-trained text-to-image diffusion
model with this information, our approach enables fine-grained 3D pose and
placement control of individual objects in a scene. We further demonstrate that
Neural Assets can be transferred and recomposed across different scenes. Our
model achieves state-of-the-art multi-object editing results on both synthetic
3D scene datasets, as well as two real-world video datasets (Objectron, Waymo
Open). |
---|---|
DOI: | 10.48550/arxiv.2406.09292 |