Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/donydchen/fmpn-fer
😁[VCIP'19 Oral] Official PyTorch Implementation of Facial Motion Prior Networks for Facial Expression Recognition
https://github.com/donydchen/fmpn-fer
expression facial-expression-recognition pytorch vcip-2019
Last synced: 7 days ago
JSON representation
😁[VCIP'19 Oral] Official PyTorch Implementation of Facial Motion Prior Networks for Facial Expression Recognition
- Host: GitHub
- URL: https://github.com/donydchen/fmpn-fer
- Owner: donydchen
- License: mit
- Created: 2019-09-09T06:44:52.000Z (about 5 years ago)
- Default Branch: master
- Last Pushed: 2020-05-11T07:46:38.000Z (over 4 years ago)
- Last Synced: 2023-10-20T19:25:06.859Z (about 1 year ago)
- Topics: expression, facial-expression-recognition, pytorch, vcip-2019
- Language: Python
- Homepage:
- Size: 1.93 MB
- Stars: 92
- Watchers: 2
- Forks: 23
- Open Issues: 4
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
# FMPN-FER
[![PWC](https://img.shields.io/endpoint.svg?style=flat-square&url=https://paperswithcode.com/badge/facial-motion-prior-networks-for-facial/facial-expression-recognition-on-mmi)](https://paperswithcode.com/sota/facial-expression-recognition-on-mmi?p=facial-motion-prior-networks-for-facial)
[![PWC](https://img.shields.io/endpoint.svg?style=flat-square&url=https://paperswithcode.com/badge/facial-motion-prior-networks-for-facial/facial-expression-recognition-on-ck)](https://paperswithcode.com/sota/facial-expression-recognition-on-ck?p=facial-motion-prior-networks-for-facial)
Official PyTorch Implementation of **Facial Motion Prior Networks for Facial Expression Recognition** by Yuedong Chen, Jianfeng Wang, Shikai Chen, Zhongchao Shi, and Jianfei Cai.
VCIP 2019, Oral, \[[arXiv](https://arxiv.org/abs/1902.08788)\]![FMPN Model Structure](images/proposed_model.png)
## Setup
### Prerequisites
* python >= 3.6
* pytorch >= 0.4.1
* visdom == 0.1.8.9### Training
The training phase mainly contains two stages, including training the generator for generating facial motion mask, and jointly training the whole recognition network. Running scripts examples are given as below.
#### Pretrain the Facial-Motion Mask Generator (FMG)
``` sh
python main.py --mode train --data_root datasets/CKPlus --train_csv train_ids_0.csv \
--print_losses_freq 4 --use_data_augment --visdom_env res_baseline_ckp_0 \
--niter 150 --niter_decay 150 --gpu_ids 0 \
--model res_baseline --solver resface --img_nc 1 \
/
```#### Jointly Train the Facial-Motion Prior Network
Load the saved weight for the FMG pertained in the previous stage, and ImageNet pretrained weight for the Classifer Net.
```sh
python main.py --mode train --data_root datasets/CKPlus --train_csv train_ids_0.csv \
--print_losses_freq 4 --use_data_augment --visdom_env res_cls_ckp_0 --niter 100 \
--niter_decay 100 --gpu_ids 0 --model res_cls --solver res_cls --lambda_resface 0.1 \
--batch_size 16 --backend_pretrain \
--load_model_dir ckpts/CKPlus/res_baseline/fold_0/190117_165651 --load_epoch 300 \
/
```### Testing
``` sh
python main.py --mode test --data_root datasets/CKPlus --test_csv test_ids_0.csv \
--gpu_ids 0 --model res_cls --solver res_cls --batch_size 4 \
--load_model_dir ckpts/CKPlus/res_cls/fold_0/190118_170050 --load_epoch 200 \
/
```## Datasets
* The pseudo ground truth facial motion masks for dataset CK+, MMI and AffectNet are presented in `/datasets`.
* Training and Testing datasets splits are located in `/datasets`, 10-fold cross validation is used in the project as a general approach.## Results Detail
10 folds accuracy for CK+, MMI and AffectNet.
| Fold ID | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | Avg |
|-----------|-------|-------|-------|-------|-------|-------|-------|-------|-------|-------|------|
| CK+ | 0.885 | 1.000 | 0.921 | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 | 0.98 |
| MMI | 0.778 | 0.731 | 0.667 | 0.744 | 0.857 | 0.855 | 0.958 | 0.889 | 1.000 | 0.795 | 0.83 |
| AffectNet | - | - | - | - | - | - | - | - | - | - | 0.62 |## Citation
If you use this code for your research, please cite our paper Facial Motion Prior Networks for Facial Expression Recognition:
```
@article{chen2019facial,
title={Facial Motion Prior Networks for Facial Expression Recognition},
author={Chen, Yuedong, Jianfeng Wang, Shikai Chen, Zhongchao Shi, and Jianfei Cai},
journal={IEEE Visual Communications and Image Processing (VCIP)},
year={2019},
organization={IEEE}
}
```## Acknowledgments
This work is mainly conducted when Yuedong is a research intern in Lenovo AI Lab, Beijing. For any questions regarding this project, feel free to create an issue or reach Yuedong by Email.