학술논문

Dynamic Facial Models for Video-Based Dimensional Affect Estimation
Document Type
Conference
Source
2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW) ICCVW Computer Vision Workshop (ICCVW), 2019 IEEE/CVF International Conference on. :1608-1617 Oct, 2019
Subject
Computing and Processing
Face
Heuristic algorithms
Shape
Task analysis
Image sequences
Estimation
Standards
Dimensional affect estimation
facial dynamic modelling
Deep learning
Language
ISSN
2473-9944
Abstract
Dimensional affect estimation from a face video is a challenging task, mainly due to the large number of possible facial displays made up of a set of behaviour primitives including facial muscle actions. The displays vary not only in composition but also in temporal evolution, with each display composed of behaviour primitives with varying in their short and long-term characteristics. Most existing work models affect relies on complex hierarchical recurrent models unable to capture short-term dynamics well. In this paper, we propose to encode these short-term facial shape and appearance dynamics in an image, where only the semantic meaningful information is encoded into the dynamic face images. We also propose binary dynamic facial masks to remove 'stable pixels' from the dynamic images. This process allows filtering of non-dynamic information, i.e. only pixels that have changed in the sequence are retained. Then, the final proposed Dynamic Facial Model (DFM) encodes both filtered facial appearance and shape dynamics of a image sequence preceding to the given frame into a three-channel raster image. A CNN-RNN architecture is tasked with modelling primarily the long-term changes. Experiments show that our dynamic face images achieved superior performance over the standard RGB face images on dimensional affect prediction task.