LetheSec / Fer2013-Facial-Emotion-Recognition-Pytorch

This method achieves SOTA single model accuracy of 73.70 % on FER2013 without using extra training data.
MIT License
84 stars 16 forks source link

Fer2013 - Facial Emotion Recognition

This work is the final project of the Computer Vision Course of USTC. However, I achieve the highest single-network classification accuracy on FER2013 based on ResNet18. To my best knowledge, this work achieves state-of-the-art single-network accuracy of 73.70 % on FER2013 without using extra training data, which exceeds the previous work [1] of 73.28%. (Chineses Post)

Method Private Test Data
[1] 73.28%
This work 73.70%

Official model checkpoint and training log can be found following:

Fer2013 Leaderboard: Here

环境

其他

First, you should download the official fer2013 dataset, and place it in the outmost folder with the following folder structure datasets/fer2013/fer2013.csv

To train your own model, run the following:

python train.py --name='your_version'

To evaluate the model, run the follwing

python evaluate.py --checkpoint='xxx/best_checkpoint.tar'

Result

Citation

If you are considering citing this work, please refer to the following:

@misc{yuan2021fer,
  title        = {Fer2013-Facial-Emotion-Recognition-Pytorch},
  author       = {Yuan, Xiaojian},
  year         = {2021},
  publisher    = {GitHub},
  journal      = {GitHub repository},
  howpublished = {\url{https://github.com/LetheSec/Fer2013-Facial-Emotion-Recognition-Pytorch}},
}

Reference

[1] Khaireddin, Yousif, and Zhuofa Chen. "Facial Emotion Recognition: State of the Art Performance on FER2013." arXiv preprint arXiv:2105.03588 (2021).