yeungchenwa / FontDiffuser

[AAAI2024] FontDiffuser: One-Shot Font Generation via Denoising Diffusion with Multi-Scale Content Aggregation and Style Contrastive Learning
https://yeungchenwa.github.io/fontdiffuser-homepage/
231 stars 21 forks source link

Regarding the problem that the style transformation effect of English characters is relatively poor. #36

Closed XieJinYan1127 closed 5 months ago

XieJinYan1127 commented 5 months ago

Thank you for your work, I was very excited when I saw this project on [spaces], and I tried it several times in English and Chinese characters. The result of the experiment was that the stylistic variation of Chinese characters was very successful, but the English was very poor. I'd like to confirm to you that this is in line with expectations?

yeungchenwa commented 5 months ago

Hi@XieJinYan1127, our released model is trained solely by Chinese dataset, which may lead to poor performance on large style variations of English characters. Theoretically, when FontDiffuser is trained using English dataset, it would solve the problem you mentioned.

mwsunshine commented 1 week ago

hi @yeungchenwa I also want to ask some questions about the training of the English letter.

  1. If I add English dataset, I need to train both the phase 1 and phase 2 process, right?
  2. how many fonts do you think is efficient?
  3. do I need to train from scratch or refine with your model would be enough?