Semantics-guided generative diffusion model with a 3DMM model condition for face swapping

Xiyao Liu, Yang Liu, Yuhao Zheng, Ting Yang, Jian Zhang*, Victoria Wang, Hui Fang*

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

9 Downloads (Pure)


Face swapping is a technique that replaces a face in a target media with another face of a different identity from a source face image. Currently, research on the effective utilisation of prior knowledge and semantic guidance for photo-realistic face swapping remains limited, despite the impressive synthesis quality achieved by recent generative models. In this paper, we propose a novel conditional Denoising Diffusion Probabilistic Model (DDPM) enforced by a two-level face prior guidance. Specifically, it includes (i) an image-level condition generated by a 3D Morphable Model (3DMM), and (ii) a high-semantic level guidance driven by information extracted from several pre-trained attribute classifiers, for high-quality face image synthesis. Although swapped face image from 3DMM does not achieve photo-realistic quality on its own, it provides a strong image-level prior, in parallel with high-level face semantics, to guide the DDPM for high fidelity image generation. The experimental results demonstrate that our method outperforms state-of-the-art face swapping methods on benchmark datasets in terms of its synthesis quality, and capability to preserve the target face attributes and swap the source face identity.

Original languageEnglish
Article numbere14949
Number of pages12
JournalComputer Graphics Forum
Issue number7
Publication statusPublished - 30 Oct 2023


  • CCS Concepts
  • computing methodologies
  • computer graphics
  • image manipulation
  • computational photography

Cite this