Human Multi-View Synthesis from a Single-View Model:Transferred Body and Face Representations
Generating multi-view human images from a single view is a complex and significant challenge. Although recent advancements in multi-view object generation have shown impressive results with diffusion models, novel view synthesis for humans remains constrained by the limited availability of 3D human...
Gespeichert in:
Hauptverfasser: | , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Generating multi-view human images from a single view is a complex and
significant challenge. Although recent advancements in multi-view object
generation have shown impressive results with diffusion models, novel view
synthesis for humans remains constrained by the limited availability of 3D
human datasets. Consequently, many existing models struggle to produce
realistic human body shapes or capture fine-grained facial details accurately.
To address these issues, we propose an innovative framework that leverages
transferred body and facial representations for multi-view human synthesis.
Specifically, we use a single-view model pretrained on a large-scale human
dataset to develop a multi-view body representation, aiming to extend the 2D
knowledge of the single-view model to a multi-view diffusion model.
Additionally, to enhance the model's detail restoration capability, we
integrate transferred multimodal facial features into our trained human
diffusion model. Experimental evaluations on benchmark datasets demonstrate
that our approach outperforms the current state-of-the-art methods, achieving
superior performance in multi-view human synthesis. |
---|---|
DOI: | 10.48550/arxiv.2412.03011 |