Neural Assets: 3D-Aware Multi-Object Scene Synthesis with Image Diffusion Models
We address the problem of multi-object 3D pose control in image diffusion models. Instead of conditioning on a sequence of text tokens, we propose to use a set of per-object representations, Neural Assets, to control the 3D pose of individual objects in a scene. Neural Assets are obtained by pooling...
Gespeichert in:
Hauptverfasser: | , , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Wu, Ziyi Rubanova, Yulia Kabra, Rishabh Hudson, Drew A Gilitschenski, Igor Aytar, Yusuf van Steenkiste, Sjoerd Allen, Kelsey R Kipf, Thomas |
description | We address the problem of multi-object 3D pose control in image diffusion
models. Instead of conditioning on a sequence of text tokens, we propose to use
a set of per-object representations, Neural Assets, to control the 3D pose of
individual objects in a scene. Neural Assets are obtained by pooling visual
representations of objects from a reference image, such as a frame in a video,
and are trained to reconstruct the respective objects in a different image,
e.g., a later frame in the video. Importantly, we encode object visuals from
the reference image while conditioning on object poses from the target frame.
This enables learning disentangled appearance and pose features. Combining
visual and 3D pose representations in a sequence-of-tokens format allows us to
keep the text-to-image architecture of existing models, with Neural Assets in
place of text tokens. By fine-tuning a pre-trained text-to-image diffusion
model with this information, our approach enables fine-grained 3D pose and
placement control of individual objects in a scene. We further demonstrate that
Neural Assets can be transferred and recomposed across different scenes. Our
model achieves state-of-the-art multi-object editing results on both synthetic
3D scene datasets, as well as two real-world video datasets (Objectron, Waymo
Open). |
doi_str_mv | 10.48550/arxiv.2406.09292 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2406_09292</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2406_09292</sourcerecordid><originalsourceid>FETCH-LOGICAL-a672-2127cdfe4f7d8bfcd9242dad50f2f6b58405d842b059474e3605d6c1857bae613</originalsourceid><addsrcrecordid>eNotz7tOwzAYhmEvDKhwAUz1DTg4jk9hi1oOlVpaqd0jH35TozRFdkLp3QOF6dO7fNKD0F1JC66FoPcmfcXPgnEqC1qzml2jzSuMyXS4yRmG_ICrOWlOJgFejd0Qydq-gxvw1kEPeHvuhz3kmPEpDnu8OJg3wPMYwpjjscero4cu36CrYLoMt_87Qbunx93shSzXz4tZsyRGKkZYyZTzAXhQXtvgfM0488YLGliQVmhOhdecWSpqrjhU8qelK7VQ1oAsqwma_t1eSO1HigeTzu0vrb3Qqm-XIUia</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Neural Assets: 3D-Aware Multi-Object Scene Synthesis with Image Diffusion Models</title><source>arXiv.org</source><creator>Wu, Ziyi ; Rubanova, Yulia ; Kabra, Rishabh ; Hudson, Drew A ; Gilitschenski, Igor ; Aytar, Yusuf ; van Steenkiste, Sjoerd ; Allen, Kelsey R ; Kipf, Thomas</creator><creatorcontrib>Wu, Ziyi ; Rubanova, Yulia ; Kabra, Rishabh ; Hudson, Drew A ; Gilitschenski, Igor ; Aytar, Yusuf ; van Steenkiste, Sjoerd ; Allen, Kelsey R ; Kipf, Thomas</creatorcontrib><description>We address the problem of multi-object 3D pose control in image diffusion
models. Instead of conditioning on a sequence of text tokens, we propose to use
a set of per-object representations, Neural Assets, to control the 3D pose of
individual objects in a scene. Neural Assets are obtained by pooling visual
representations of objects from a reference image, such as a frame in a video,
and are trained to reconstruct the respective objects in a different image,
e.g., a later frame in the video. Importantly, we encode object visuals from
the reference image while conditioning on object poses from the target frame.
This enables learning disentangled appearance and pose features. Combining
visual and 3D pose representations in a sequence-of-tokens format allows us to
keep the text-to-image architecture of existing models, with Neural Assets in
place of text tokens. By fine-tuning a pre-trained text-to-image diffusion
model with this information, our approach enables fine-grained 3D pose and
placement control of individual objects in a scene. We further demonstrate that
Neural Assets can be transferred and recomposed across different scenes. Our
model achieves state-of-the-art multi-object editing results on both synthetic
3D scene datasets, as well as two real-world video datasets (Objectron, Waymo
Open).</description><identifier>DOI: 10.48550/arxiv.2406.09292</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Computer Vision and Pattern Recognition ; Computer Science - Learning</subject><creationdate>2024-06</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,781,886</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2406.09292$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2406.09292$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Wu, Ziyi</creatorcontrib><creatorcontrib>Rubanova, Yulia</creatorcontrib><creatorcontrib>Kabra, Rishabh</creatorcontrib><creatorcontrib>Hudson, Drew A</creatorcontrib><creatorcontrib>Gilitschenski, Igor</creatorcontrib><creatorcontrib>Aytar, Yusuf</creatorcontrib><creatorcontrib>van Steenkiste, Sjoerd</creatorcontrib><creatorcontrib>Allen, Kelsey R</creatorcontrib><creatorcontrib>Kipf, Thomas</creatorcontrib><title>Neural Assets: 3D-Aware Multi-Object Scene Synthesis with Image Diffusion Models</title><description>We address the problem of multi-object 3D pose control in image diffusion
models. Instead of conditioning on a sequence of text tokens, we propose to use
a set of per-object representations, Neural Assets, to control the 3D pose of
individual objects in a scene. Neural Assets are obtained by pooling visual
representations of objects from a reference image, such as a frame in a video,
and are trained to reconstruct the respective objects in a different image,
e.g., a later frame in the video. Importantly, we encode object visuals from
the reference image while conditioning on object poses from the target frame.
This enables learning disentangled appearance and pose features. Combining
visual and 3D pose representations in a sequence-of-tokens format allows us to
keep the text-to-image architecture of existing models, with Neural Assets in
place of text tokens. By fine-tuning a pre-trained text-to-image diffusion
model with this information, our approach enables fine-grained 3D pose and
placement control of individual objects in a scene. We further demonstrate that
Neural Assets can be transferred and recomposed across different scenes. Our
model achieves state-of-the-art multi-object editing results on both synthetic
3D scene datasets, as well as two real-world video datasets (Objectron, Waymo
Open).</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Computer Vision and Pattern Recognition</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotz7tOwzAYhmEvDKhwAUz1DTg4jk9hi1oOlVpaqd0jH35TozRFdkLp3QOF6dO7fNKD0F1JC66FoPcmfcXPgnEqC1qzml2jzSuMyXS4yRmG_ICrOWlOJgFejd0Qydq-gxvw1kEPeHvuhz3kmPEpDnu8OJg3wPMYwpjjscero4cu36CrYLoMt_87Qbunx93shSzXz4tZsyRGKkZYyZTzAXhQXtvgfM0488YLGliQVmhOhdecWSpqrjhU8qelK7VQ1oAsqwma_t1eSO1HigeTzu0vrb3Qqm-XIUia</recordid><startdate>20240613</startdate><enddate>20240613</enddate><creator>Wu, Ziyi</creator><creator>Rubanova, Yulia</creator><creator>Kabra, Rishabh</creator><creator>Hudson, Drew A</creator><creator>Gilitschenski, Igor</creator><creator>Aytar, Yusuf</creator><creator>van Steenkiste, Sjoerd</creator><creator>Allen, Kelsey R</creator><creator>Kipf, Thomas</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20240613</creationdate><title>Neural Assets: 3D-Aware Multi-Object Scene Synthesis with Image Diffusion Models</title><author>Wu, Ziyi ; Rubanova, Yulia ; Kabra, Rishabh ; Hudson, Drew A ; Gilitschenski, Igor ; Aytar, Yusuf ; van Steenkiste, Sjoerd ; Allen, Kelsey R ; Kipf, Thomas</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a672-2127cdfe4f7d8bfcd9242dad50f2f6b58405d842b059474e3605d6c1857bae613</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Computer Vision and Pattern Recognition</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Wu, Ziyi</creatorcontrib><creatorcontrib>Rubanova, Yulia</creatorcontrib><creatorcontrib>Kabra, Rishabh</creatorcontrib><creatorcontrib>Hudson, Drew A</creatorcontrib><creatorcontrib>Gilitschenski, Igor</creatorcontrib><creatorcontrib>Aytar, Yusuf</creatorcontrib><creatorcontrib>van Steenkiste, Sjoerd</creatorcontrib><creatorcontrib>Allen, Kelsey R</creatorcontrib><creatorcontrib>Kipf, Thomas</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Wu, Ziyi</au><au>Rubanova, Yulia</au><au>Kabra, Rishabh</au><au>Hudson, Drew A</au><au>Gilitschenski, Igor</au><au>Aytar, Yusuf</au><au>van Steenkiste, Sjoerd</au><au>Allen, Kelsey R</au><au>Kipf, Thomas</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Neural Assets: 3D-Aware Multi-Object Scene Synthesis with Image Diffusion Models</atitle><date>2024-06-13</date><risdate>2024</risdate><abstract>We address the problem of multi-object 3D pose control in image diffusion
models. Instead of conditioning on a sequence of text tokens, we propose to use
a set of per-object representations, Neural Assets, to control the 3D pose of
individual objects in a scene. Neural Assets are obtained by pooling visual
representations of objects from a reference image, such as a frame in a video,
and are trained to reconstruct the respective objects in a different image,
e.g., a later frame in the video. Importantly, we encode object visuals from
the reference image while conditioning on object poses from the target frame.
This enables learning disentangled appearance and pose features. Combining
visual and 3D pose representations in a sequence-of-tokens format allows us to
keep the text-to-image architecture of existing models, with Neural Assets in
place of text tokens. By fine-tuning a pre-trained text-to-image diffusion
model with this information, our approach enables fine-grained 3D pose and
placement control of individual objects in a scene. We further demonstrate that
Neural Assets can be transferred and recomposed across different scenes. Our
model achieves state-of-the-art multi-object editing results on both synthetic
3D scene datasets, as well as two real-world video datasets (Objectron, Waymo
Open).</abstract><doi>10.48550/arxiv.2406.09292</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2406.09292 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2406_09292 |
source | arXiv.org |
subjects | Computer Science - Artificial Intelligence Computer Science - Computer Vision and Pattern Recognition Computer Science - Learning |
title | Neural Assets: 3D-Aware Multi-Object Scene Synthesis with Image Diffusion Models |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-18T15%3A09%3A41IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Neural%20Assets:%203D-Aware%20Multi-Object%20Scene%20Synthesis%20with%20Image%20Diffusion%20Models&rft.au=Wu,%20Ziyi&rft.date=2024-06-13&rft_id=info:doi/10.48550/arxiv.2406.09292&rft_dat=%3Carxiv_GOX%3E2406_09292%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |