TextFace: Text-to-Style Mapping Based Face Generation and Manipulation

Document Type

Article

Publication Date

1-1-2023

Abstract

As a subtopic of text-to-image synthesis, text-to-face generation has great potential in face-related applications. In this paper, we propose a generic text-to-face framework, namely, TextFace, to achieve diverse and high-quality face image generation from text descriptions. We introduce text-to-style mapping, a novel method where the text description can be directly encoded into the latent space of a pretrained StyleGAN. Guided by our text-image similarity matching and face captioning-based text alignment, the textual latent code can be fed into the generator of a well-trained StyleGAN to produce diverse face images with high resolution (1024×1024). Furthermore, our model inherently supports semantic face editing using text descriptions. Finally, experimental results quantitatively and qualitatively demonstrate the superior performance of our model.

Publication Title

IEEE Transactions on Multimedia

First Page Number

3409

Last Page Number

3419

DOI

10.1109/TMM.2022.3160360

This document is currently not available here.

Share

COinS