We use cookies to improve your experience with our site.
Wen-Wen Fu, Wen-Juan Gong, Chen-Yang Yu, Wei Wang, Jordi Gonzàlez. Facial Expression Generation from Text with FaceCLIP[J]. Journal of Computer Science and Technology. DOI: 10.1007/s11390-024-3661-z
Citation: Wen-Wen Fu, Wen-Juan Gong, Chen-Yang Yu, Wei Wang, Jordi Gonzàlez. Facial Expression Generation from Text with FaceCLIP[J]. Journal of Computer Science and Technology. DOI: 10.1007/s11390-024-3661-z

Facial Expression Generation from Text with FaceCLIP

  • Facial expression generation from pure textual descriptions is widely applied in human-computer interaction, computer-aided design, assisted education, etc. However, this task is challenging due to intricate facial structure and complex mapping between texts and images. Existing methods face limitations in generating high-resolution images or capturing diverse facial expressions. In this study, we propose a novel generation approach, named FaceCLIP, to tackle these problems. The proposed method utilizes a CLIP-based multi-stage generative adversarial model to produce vivid facial expressions with high-resolutions. With strong semantic priors from multi-modal textual and visual cues, the proposed model effectively disentangles facial attributes, enabling attribute editing and semantic reasoning. To facilitate text-to-expression generation, we build a new dataset called the FET dataset, which contains facial expression images and corresponding textual descriptions. Experiments on the dataset demonstrate improved image quality and semantic consistency compared with the state-of-the-art methods. The dataset and models are available: https://github.com/ourpubliccodes/FaceCLIP.
  • loading

Catalog

    /

    DownLoad:  Full-Size Img  PowerPoint
    Return
    Return