학술논문

Self-Supervised Spontaneous Latent-Based Facial Expression Sequence Generation
Document Type
Periodical
Source
IEEE Open Journal of Signal Processing IEEE Open J. Signal Process. Signal Processing, IEEE Open Journal of. 4:304-312 2023
Subject
Signal Processing and Analysis
Faces
Mathematical models
Markov processes
Gold
Training
Task analysis
Image sequences
Affective computing
artificial neural networks
self-supervised learning
Language
ISSN
2644-1322
Abstract
In this article, we investigate the spontaneity issue in facial expression sequence generation. Current leading methods in the field are commonly reliant on manually adjusted conditional variables to direct the model to generate a specific class of expression. We propose a neural network-based method which uses Gaussian noise to model spontaneity in the generation process, removing the need for manual control of conditional generation variables. Our model takes two sequential images as input, with additive noise, and produces the next image in the sequence. We trained two types of models: single-expression, and mixed-expression. With single-expression, unique facial movements of certain emotion class can be generated; with mixed expressions, fully spontaneous expression sequence generation can be achieved. We compared our method to current leading generation methods on a variety of publicly available datasets. Initial qualitative results show our method produces visually more realistic expressions and facial action unit (AU) trajectories; initial quantitative results using image quality metrics (SSIM and NIQE) show the quality of our generated images is higher. Our approach and results are novel in the field of facial expression generation, with potential wider applications to other sequence generation tasks.