EdgeFont: Enhancing style and content representations in few-shot font generation with multi-scale edge self-supervision
Font design is a meticulous and resource-intensive endeavor, especially for the intricate Chinese font. The few-shot font generation (FFG), i.e., employing a few reference characters to create diverse characters with the same style, has garnered significant interest recently. Existing models are pre...
Gespeichert in:
Veröffentlicht in: | Expert systems with applications 2025-03, Vol.262, p.125547, Article 125547 |
---|---|
Hauptverfasser: | , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Font design is a meticulous and resource-intensive endeavor, especially for the intricate Chinese font. The few-shot font generation (FFG), i.e., employing a few reference characters to create diverse characters with the same style, has garnered significant interest recently. Existing models are predominantly based on the aggregation of style and content representations through learning either global or local style representations with neural networks. Yet, existing models commonly lack effective information guidance during the training or necessitate costly character information acquisition, limiting their performance and applicability in more intricate scenarios. To address this issue, this paper proposes a novel self-supervised few-shot font generation model called EdgeFont by introducing a Self-Supervised Multi-Scale edge Information based Self-Supervised (MSEE) module motivated by the observation that the multi-scale edge can simultaneously capture global and local style information. The introduced self-supervised module can not only provide nice supervision for the learning of style and content but also have a low cost of edge extraction. The experimental results on various datasets show that the proposed model outperforms existing models in terms of PSNR, SSIM, MSE, LPIPS, and FID. In the most challenging task for few-shot generation, Unseen fonts and Seen character, the proposed model achieves improvements of 0.95, 0.055, 0.063, 0.085, and 51.73 in PSNR, SSIM, MSE, and LPIPS, respectively, compared to FUNIT. Specifically, after integrating the MESS module into CG-GAN, the FID improves by 4.53, which fully demonstrates the strong scalability of MESS.
•A novel few-shot font generation model is proposed by integrating edge information.•Edge information is used by a self-supervised way without manual edge labeling.•The self-supervised module can enhance the style and content representations.•Effectiveness of proposed model is demonstrated by extensive experiments.•Good generalization of proposed model for cross-lingual font generation is shown. |
---|---|
ISSN: | 0957-4174 |
DOI: | 10.1016/j.eswa.2024.125547 |