ApprGAN: Appearance-Based Generative Adversarial Network for Facial Expression Synthesis

Yao Peng, Hujun Yin

    Research output: Contribution to journalArticlepeer-review

    357 Downloads (Pure)

    Abstract

    Facial expression synthesis has drawn increasing attention in computer vision, graphics and animation. Recently generative adversarial nets (GANs) have become a new perspective for face synthesis and have had remarkable success in generating photorealistic images and image-to-image translation. In this paper, we present an appearance-based facial expression synthesis framework, ApprGAN, by combining shape and texture and introducing cycle-consistency and identity mapping into the adversarial learning. Specifically, given an input face image, a pair of shape and texture generators are trained for synthetic shape deformation and expression detail generation, respectively. Extensive experiments on expression synthesis and cross-database synthesis were conducted, together with comparisons with the existing methods. Results of expression synthesis and quantitative verification on various databases show the effectiveness of ApprGAN in synthesising photorealistic and identity-preserving expressions and its marked improvement over the existing methods.
    Original languageEnglish
    JournalIET Image Processing
    Early online date29 Mar 2019
    DOIs
    Publication statusPublished - 2019

    Keywords

    • Generative adversarial nets (GANs)
    • Face recognition
    • Expression analysis

    Fingerprint

    Dive into the research topics of 'ApprGAN: Appearance-Based Generative Adversarial Network for Facial Expression Synthesis'. Together they form a unique fingerprint.

    Cite this