STAG4D: Spatial-Temporal Anchored Generative 4D Gaussians
Recent progress in pre-trained diffusion models and 3D generation have spurred interest in 4D content creation. However, achieving high-fidelity 4D generation with spatial-temporal consistency remains a challenge. In this work, we propose STAG4D, a novel framework that combines pre-trained diffusion...
Gespeichert in:
Hauptverfasser: | , , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Zeng, Yifei Jiang, Yanqin Zhu, Siyu Lu, Yuanxun Lin, Youtian Zhu, Hao Hu, Weiming Cao, Xun Yao, Yao |
description | Recent progress in pre-trained diffusion models and 3D generation have
spurred interest in 4D content creation. However, achieving high-fidelity 4D
generation with spatial-temporal consistency remains a challenge. In this work,
we propose STAG4D, a novel framework that combines pre-trained diffusion models
with dynamic 3D Gaussian splatting for high-fidelity 4D generation. Drawing
inspiration from 3D generation techniques, we utilize a multi-view diffusion
model to initialize multi-view images anchoring on the input video frames,
where the video can be either real-world captured or generated by a video
diffusion model. To ensure the temporal consistency of the multi-view sequence
initialization, we introduce a simple yet effective fusion strategy to leverage
the first frame as a temporal anchor in the self-attention computation. With
the almost consistent multi-view sequences, we then apply the score
distillation sampling to optimize the 4D Gaussian point cloud. The 4D Gaussian
spatting is specially crafted for the generation task, where an adaptive
densification strategy is proposed to mitigate the unstable Gaussian gradient
for robust optimization. Notably, the proposed pipeline does not require any
pre-training or fine-tuning of diffusion networks, offering a more accessible
and practical solution for the 4D generation task. Extensive experiments
demonstrate that our method outperforms prior 4D generation works in rendering
quality, spatial-temporal consistency, and generation robustness, setting a new
state-of-the-art for 4D generation from diverse inputs, including text, image,
and video. |
doi_str_mv | 10.48550/arxiv.2403.14939 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2403_14939</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2403_14939</sourcerecordid><originalsourceid>FETCH-LOGICAL-a679-4ecb8db7989de9f6c839d04b39559bec9d5051592b897bcb268c3654c1d802813</originalsourceid><addsrcrecordid>eNotj71uwjAURr10qCgP0Am_QFI7thPfbhE_AQmJgezRtX0RkUKInBbB29MC0xk-6eg7jH1KkWprjPjCeG0vaaaFSqUGBe8M9nVZ6cU33w_402KX1HQazhE7Xvb-eI4UeEU9xb_xQlwveIW_49hiP36wtwN2I01fnLB6tazn62S7qzbzcptgXkCiyTsbXAEWAsEh91ZBENopMAYceQhGGGkgcxYK512WW69yo70MVmRWqgmbPbWP780Q2xPGW_Pf0Dwa1B0Emj-r</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>STAG4D: Spatial-Temporal Anchored Generative 4D Gaussians</title><source>arXiv.org</source><creator>Zeng, Yifei ; Jiang, Yanqin ; Zhu, Siyu ; Lu, Yuanxun ; Lin, Youtian ; Zhu, Hao ; Hu, Weiming ; Cao, Xun ; Yao, Yao</creator><creatorcontrib>Zeng, Yifei ; Jiang, Yanqin ; Zhu, Siyu ; Lu, Yuanxun ; Lin, Youtian ; Zhu, Hao ; Hu, Weiming ; Cao, Xun ; Yao, Yao</creatorcontrib><description>Recent progress in pre-trained diffusion models and 3D generation have
spurred interest in 4D content creation. However, achieving high-fidelity 4D
generation with spatial-temporal consistency remains a challenge. In this work,
we propose STAG4D, a novel framework that combines pre-trained diffusion models
with dynamic 3D Gaussian splatting for high-fidelity 4D generation. Drawing
inspiration from 3D generation techniques, we utilize a multi-view diffusion
model to initialize multi-view images anchoring on the input video frames,
where the video can be either real-world captured or generated by a video
diffusion model. To ensure the temporal consistency of the multi-view sequence
initialization, we introduce a simple yet effective fusion strategy to leverage
the first frame as a temporal anchor in the self-attention computation. With
the almost consistent multi-view sequences, we then apply the score
distillation sampling to optimize the 4D Gaussian point cloud. The 4D Gaussian
spatting is specially crafted for the generation task, where an adaptive
densification strategy is proposed to mitigate the unstable Gaussian gradient
for robust optimization. Notably, the proposed pipeline does not require any
pre-training or fine-tuning of diffusion networks, offering a more accessible
and practical solution for the 4D generation task. Extensive experiments
demonstrate that our method outperforms prior 4D generation works in rendering
quality, spatial-temporal consistency, and generation robustness, setting a new
state-of-the-art for 4D generation from diverse inputs, including text, image,
and video.</description><identifier>DOI: 10.48550/arxiv.2403.14939</identifier><language>eng</language><subject>Computer Science - Computer Vision and Pattern Recognition</subject><creationdate>2024-03</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2403.14939$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2403.14939$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Zeng, Yifei</creatorcontrib><creatorcontrib>Jiang, Yanqin</creatorcontrib><creatorcontrib>Zhu, Siyu</creatorcontrib><creatorcontrib>Lu, Yuanxun</creatorcontrib><creatorcontrib>Lin, Youtian</creatorcontrib><creatorcontrib>Zhu, Hao</creatorcontrib><creatorcontrib>Hu, Weiming</creatorcontrib><creatorcontrib>Cao, Xun</creatorcontrib><creatorcontrib>Yao, Yao</creatorcontrib><title>STAG4D: Spatial-Temporal Anchored Generative 4D Gaussians</title><description>Recent progress in pre-trained diffusion models and 3D generation have
spurred interest in 4D content creation. However, achieving high-fidelity 4D
generation with spatial-temporal consistency remains a challenge. In this work,
we propose STAG4D, a novel framework that combines pre-trained diffusion models
with dynamic 3D Gaussian splatting for high-fidelity 4D generation. Drawing
inspiration from 3D generation techniques, we utilize a multi-view diffusion
model to initialize multi-view images anchoring on the input video frames,
where the video can be either real-world captured or generated by a video
diffusion model. To ensure the temporal consistency of the multi-view sequence
initialization, we introduce a simple yet effective fusion strategy to leverage
the first frame as a temporal anchor in the self-attention computation. With
the almost consistent multi-view sequences, we then apply the score
distillation sampling to optimize the 4D Gaussian point cloud. The 4D Gaussian
spatting is specially crafted for the generation task, where an adaptive
densification strategy is proposed to mitigate the unstable Gaussian gradient
for robust optimization. Notably, the proposed pipeline does not require any
pre-training or fine-tuning of diffusion networks, offering a more accessible
and practical solution for the 4D generation task. Extensive experiments
demonstrate that our method outperforms prior 4D generation works in rendering
quality, spatial-temporal consistency, and generation robustness, setting a new
state-of-the-art for 4D generation from diverse inputs, including text, image,
and video.</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj71uwjAURr10qCgP0Am_QFI7thPfbhE_AQmJgezRtX0RkUKInBbB29MC0xk-6eg7jH1KkWprjPjCeG0vaaaFSqUGBe8M9nVZ6cU33w_402KX1HQazhE7Xvb-eI4UeEU9xb_xQlwveIW_49hiP36wtwN2I01fnLB6tazn62S7qzbzcptgXkCiyTsbXAEWAsEh91ZBENopMAYceQhGGGkgcxYK512WW69yo70MVmRWqgmbPbWP780Q2xPGW_Pf0Dwa1B0Emj-r</recordid><startdate>20240322</startdate><enddate>20240322</enddate><creator>Zeng, Yifei</creator><creator>Jiang, Yanqin</creator><creator>Zhu, Siyu</creator><creator>Lu, Yuanxun</creator><creator>Lin, Youtian</creator><creator>Zhu, Hao</creator><creator>Hu, Weiming</creator><creator>Cao, Xun</creator><creator>Yao, Yao</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20240322</creationdate><title>STAG4D: Spatial-Temporal Anchored Generative 4D Gaussians</title><author>Zeng, Yifei ; Jiang, Yanqin ; Zhu, Siyu ; Lu, Yuanxun ; Lin, Youtian ; Zhu, Hao ; Hu, Weiming ; Cao, Xun ; Yao, Yao</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a679-4ecb8db7989de9f6c839d04b39559bec9d5051592b897bcb268c3654c1d802813</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><toplevel>online_resources</toplevel><creatorcontrib>Zeng, Yifei</creatorcontrib><creatorcontrib>Jiang, Yanqin</creatorcontrib><creatorcontrib>Zhu, Siyu</creatorcontrib><creatorcontrib>Lu, Yuanxun</creatorcontrib><creatorcontrib>Lin, Youtian</creatorcontrib><creatorcontrib>Zhu, Hao</creatorcontrib><creatorcontrib>Hu, Weiming</creatorcontrib><creatorcontrib>Cao, Xun</creatorcontrib><creatorcontrib>Yao, Yao</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Zeng, Yifei</au><au>Jiang, Yanqin</au><au>Zhu, Siyu</au><au>Lu, Yuanxun</au><au>Lin, Youtian</au><au>Zhu, Hao</au><au>Hu, Weiming</au><au>Cao, Xun</au><au>Yao, Yao</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>STAG4D: Spatial-Temporal Anchored Generative 4D Gaussians</atitle><date>2024-03-22</date><risdate>2024</risdate><abstract>Recent progress in pre-trained diffusion models and 3D generation have
spurred interest in 4D content creation. However, achieving high-fidelity 4D
generation with spatial-temporal consistency remains a challenge. In this work,
we propose STAG4D, a novel framework that combines pre-trained diffusion models
with dynamic 3D Gaussian splatting for high-fidelity 4D generation. Drawing
inspiration from 3D generation techniques, we utilize a multi-view diffusion
model to initialize multi-view images anchoring on the input video frames,
where the video can be either real-world captured or generated by a video
diffusion model. To ensure the temporal consistency of the multi-view sequence
initialization, we introduce a simple yet effective fusion strategy to leverage
the first frame as a temporal anchor in the self-attention computation. With
the almost consistent multi-view sequences, we then apply the score
distillation sampling to optimize the 4D Gaussian point cloud. The 4D Gaussian
spatting is specially crafted for the generation task, where an adaptive
densification strategy is proposed to mitigate the unstable Gaussian gradient
for robust optimization. Notably, the proposed pipeline does not require any
pre-training or fine-tuning of diffusion networks, offering a more accessible
and practical solution for the 4D generation task. Extensive experiments
demonstrate that our method outperforms prior 4D generation works in rendering
quality, spatial-temporal consistency, and generation robustness, setting a new
state-of-the-art for 4D generation from diverse inputs, including text, image,
and video.</abstract><doi>10.48550/arxiv.2403.14939</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2403.14939 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2403_14939 |
source | arXiv.org |
subjects | Computer Science - Computer Vision and Pattern Recognition |
title | STAG4D: Spatial-Temporal Anchored Generative 4D Gaussians |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-26T16%3A14%3A28IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=STAG4D:%20Spatial-Temporal%20Anchored%20Generative%204D%20Gaussians&rft.au=Zeng,%20Yifei&rft.date=2024-03-22&rft_id=info:doi/10.48550/arxiv.2403.14939&rft_dat=%3Carxiv_GOX%3E2403_14939%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |