Mask-aware photorealistic facial attribute manipulation

Ruoqi Sun, Chen Huang, Hengliang Zhu, Lizhuang Ma

Research output: Contribution to journalArticlepeer-review

18 Scopus citations

Abstract

The technique of facial attribute manipulation has found increasing application, but it remains challenging to restrict editing of attributes so that a face’s unique details are preserved. In this paper, we introduce our method, which we call a mask-adversarial autoencoder (M-AAE). It combines a variational autoencoder (VAE) and a generative adversarial network (GAN) for photorealistic image generation. We use partial dilated layers to modify a few pixels in the feature maps of an encoder, changing the attribute strength continuously without hindering global information. Our training objectives for the VAE and GAN are reinforced by supervision of face recognition loss and cycle consistency loss, to faithfully preserve facial details. Moreover, we generate facial masks to enforce background consistency, which allows our training to focus on the foreground face rather than the background. Experimental results demonstrate that our method can generate high-quality images with varying attributes, and outperforms existing methods in detail preservation.

Original languageEnglish
Pages (from-to)363-374
Number of pages12
JournalComputational Visual Media
Volume7
Issue number3
DOIs
StatePublished - Sep 2021
Externally publishedYes

Keywords

  • face attribute manipulation
  • generative adversarial network (GAN)
  • partial dilated layers
  • photorealism
  • variational autoencoder (VAE)

Fingerprint

Dive into the research topics of 'Mask-aware photorealistic facial attribute manipulation'. Together they form a unique fingerprint.

Cite this