Style Aligned Image Generation via Shared Attention

Large-scale Text-to-Image (T2I) models have rapidly gained prominence across creative fields, generating visually compelling outputs from textual prompts. However, controlling these models to ensure consistent style remains challenging, with existing methods necessitating fine-tuning and manual inte...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Hertz, Amir, Voynov, Andrey, Fruchter, Shlomi, Cohen-Or, Daniel
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Hertz, Amir
Voynov, Andrey
Fruchter, Shlomi
Cohen-Or, Daniel
description Large-scale Text-to-Image (T2I) models have rapidly gained prominence across creative fields, generating visually compelling outputs from textual prompts. However, controlling these models to ensure consistent style remains challenging, with existing methods necessitating fine-tuning and manual intervention to disentangle content and style. In this paper, we introduce StyleAligned, a novel technique designed to establish style alignment among a series of generated images. By employing minimal `attention sharing' during the diffusion process, our method maintains style consistency across images within T2I models. This approach allows for the creation of style-consistent images using a reference style through a straightforward inversion operation. Our method's evaluation across diverse styles and text prompts demonstrates high-quality synthesis and fidelity, underscoring its efficacy in achieving consistent style across various inputs.
doi_str_mv 10.48550/arxiv.2312.02133
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2312_02133</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2312_02133</sourcerecordid><originalsourceid>FETCH-LOGICAL-a673-c3660c4e7cee4d81b910c8c2b1830f17ee65276b1627345ea181c66397c21ca93</originalsourceid><addsrcrecordid>eNotzsGKwjAUBdBsZiHqB7gyP9BOXl6bpMsijiMILuq-vMZnDbR1iEXGvx91XF24Fy5HiAWoNHN5rj4p_oZbqhF0qjQgTgRW471jWXahHfgotz21LDc8cKQxXAZ5CySrM8XHVo4jD89yJj5O1F15_s6pOHytD6vvZLffbFflLiFjMfFojPIZW8-cHR00BSjvvG7AoTqBZTa5tqYBoy1mORM48MZgYb0GTwVOxfL_9qWuf2LoKd7rp75-6fEPnMw9zA</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Style Aligned Image Generation via Shared Attention</title><source>arXiv.org</source><creator>Hertz, Amir ; Voynov, Andrey ; Fruchter, Shlomi ; Cohen-Or, Daniel</creator><creatorcontrib>Hertz, Amir ; Voynov, Andrey ; Fruchter, Shlomi ; Cohen-Or, Daniel</creatorcontrib><description>Large-scale Text-to-Image (T2I) models have rapidly gained prominence across creative fields, generating visually compelling outputs from textual prompts. However, controlling these models to ensure consistent style remains challenging, with existing methods necessitating fine-tuning and manual intervention to disentangle content and style. In this paper, we introduce StyleAligned, a novel technique designed to establish style alignment among a series of generated images. By employing minimal `attention sharing' during the diffusion process, our method maintains style consistency across images within T2I models. This approach allows for the creation of style-consistent images using a reference style through a straightforward inversion operation. Our method's evaluation across diverse styles and text prompts demonstrates high-quality synthesis and fidelity, underscoring its efficacy in achieving consistent style across various inputs.</description><identifier>DOI: 10.48550/arxiv.2312.02133</identifier><language>eng</language><subject>Computer Science - Computer Vision and Pattern Recognition ; Computer Science - Graphics ; Computer Science - Learning</subject><creationdate>2023-12</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2312.02133$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2312.02133$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Hertz, Amir</creatorcontrib><creatorcontrib>Voynov, Andrey</creatorcontrib><creatorcontrib>Fruchter, Shlomi</creatorcontrib><creatorcontrib>Cohen-Or, Daniel</creatorcontrib><title>Style Aligned Image Generation via Shared Attention</title><description>Large-scale Text-to-Image (T2I) models have rapidly gained prominence across creative fields, generating visually compelling outputs from textual prompts. However, controlling these models to ensure consistent style remains challenging, with existing methods necessitating fine-tuning and manual intervention to disentangle content and style. In this paper, we introduce StyleAligned, a novel technique designed to establish style alignment among a series of generated images. By employing minimal `attention sharing' during the diffusion process, our method maintains style consistency across images within T2I models. This approach allows for the creation of style-consistent images using a reference style through a straightforward inversion operation. Our method's evaluation across diverse styles and text prompts demonstrates high-quality synthesis and fidelity, underscoring its efficacy in achieving consistent style across various inputs.</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><subject>Computer Science - Graphics</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotzsGKwjAUBdBsZiHqB7gyP9BOXl6bpMsijiMILuq-vMZnDbR1iEXGvx91XF24Fy5HiAWoNHN5rj4p_oZbqhF0qjQgTgRW471jWXahHfgotz21LDc8cKQxXAZ5CySrM8XHVo4jD89yJj5O1F15_s6pOHytD6vvZLffbFflLiFjMfFojPIZW8-cHR00BSjvvG7AoTqBZTa5tqYBoy1mORM48MZgYb0GTwVOxfL_9qWuf2LoKd7rp75-6fEPnMw9zA</recordid><startdate>20231204</startdate><enddate>20231204</enddate><creator>Hertz, Amir</creator><creator>Voynov, Andrey</creator><creator>Fruchter, Shlomi</creator><creator>Cohen-Or, Daniel</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20231204</creationdate><title>Style Aligned Image Generation via Shared Attention</title><author>Hertz, Amir ; Voynov, Andrey ; Fruchter, Shlomi ; Cohen-Or, Daniel</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a673-c3660c4e7cee4d81b910c8c2b1830f17ee65276b1627345ea181c66397c21ca93</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><topic>Computer Science - Graphics</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Hertz, Amir</creatorcontrib><creatorcontrib>Voynov, Andrey</creatorcontrib><creatorcontrib>Fruchter, Shlomi</creatorcontrib><creatorcontrib>Cohen-Or, Daniel</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Hertz, Amir</au><au>Voynov, Andrey</au><au>Fruchter, Shlomi</au><au>Cohen-Or, Daniel</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Style Aligned Image Generation via Shared Attention</atitle><date>2023-12-04</date><risdate>2023</risdate><abstract>Large-scale Text-to-Image (T2I) models have rapidly gained prominence across creative fields, generating visually compelling outputs from textual prompts. However, controlling these models to ensure consistent style remains challenging, with existing methods necessitating fine-tuning and manual intervention to disentangle content and style. In this paper, we introduce StyleAligned, a novel technique designed to establish style alignment among a series of generated images. By employing minimal `attention sharing' during the diffusion process, our method maintains style consistency across images within T2I models. This approach allows for the creation of style-consistent images using a reference style through a straightforward inversion operation. Our method's evaluation across diverse styles and text prompts demonstrates high-quality synthesis and fidelity, underscoring its efficacy in achieving consistent style across various inputs.</abstract><doi>10.48550/arxiv.2312.02133</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2312.02133
ispartof
issn
language eng
recordid cdi_arxiv_primary_2312_02133
source arXiv.org
subjects Computer Science - Computer Vision and Pattern Recognition
Computer Science - Graphics
Computer Science - Learning
title Style Aligned Image Generation via Shared Attention
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-25T18%3A10%3A45IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Style%20Aligned%20Image%20Generation%20via%20Shared%20Attention&rft.au=Hertz,%20Amir&rft.date=2023-12-04&rft_id=info:doi/10.48550/arxiv.2312.02133&rft_dat=%3Carxiv_GOX%3E2312_02133%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true