mysee1989 / TCAE

Self-supervised Representation Learning from Videos for Facial Action Unit Detection
162 stars 27 forks source link

Self-supervised Representation Learning from Videos for Facial Action Unit Detection, CVPR 2019 (oral)

We propose a Twin-Cycle Autoencoder (TCAE) that self-supervisedly learns two embeddings to encode the movements of Facial Actions and Head Motions.

Given a source and target facial images, TCAE is tasked to change the AUs or head poses of the source frame to those of the target frame by predicting the AU-related and pose-related movements, respectively.

The generated AU-changed and pose-changed faces are shown as below:

Please refer to the original paper and supplementary file for more examples.

Prerequisites

Training yourself

The learned AU embedding from TCAE can be used for both AU detection and facial image retrieval.

Pretrained model released !

Frequently asked questions ...

If you use this code in your paper, please cite the following:

@InProceedings{Li_2019_CVPR,
author = {Li, Yong and Zeng, Jiabei and Shan, Shiguang and Chen, Xilin},
title = {Self-Supervised Representation Learning From Videos for Facial Action Unit Detection},
booktitle = {The IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2019}
}