TCDiff: Triple Condition Diffusion Model with 3D Constraints for Stylizing Synthetic Faces
A robust face recognition model must be trained using datasets that include a large number of subjects and numerous samples per subject under varying conditions (such as pose, expression, age, noise, and occlusion). Due to ethical and privacy concerns, large-scale real face datasets have been discon...
Saved in:
Main Authors | , , , , |
---|---|
Format | Journal Article |
Language | English |
Published |
05.09.2024
|
Subjects | |
Online Access | Get full text |
DOI | 10.48550/arxiv.2409.03600 |
Cover
Loading…
Summary: | A robust face recognition model must be trained using datasets that include a
large number of subjects and numerous samples per subject under varying
conditions (such as pose, expression, age, noise, and occlusion). Due to
ethical and privacy concerns, large-scale real face datasets have been
discontinued, such as MS1MV3, and synthetic face generators have been proposed,
utilizing GANs and Diffusion Models, such as SYNFace, SFace, DigiFace-1M,
IDiff-Face, DCFace, and GANDiffFace, aiming to supply this demand. Some of
these methods can produce high-fidelity realistic faces, but with low
intra-class variance, while others generate high-variance faces with low
identity consistency. In this paper, we propose a Triple Condition Diffusion
Model (TCDiff) to improve face style transfer from real to synthetic faces
through 2D and 3D facial constraints, enhancing face identity consistency while
keeping the necessary high intra-class variance. Face recognition experiments
using 1k, 2k, and 5k classes of our new dataset for training outperform
state-of-the-art synthetic datasets in real face benchmarks such as LFW,
CFP-FP, AgeDB, and BUPT. Our source code is available at:
https://github.com/BOVIFOCR/tcdiff. |
---|---|
DOI: | 10.48550/arxiv.2409.03600 |