ADLD
IEEE Transactions on Affective Computing "Unconstrained Facial Action Unit Detection via Latent Feature Domain"
Install / Use
/learn @ZhiwenShao/ADLDREADME
ADLD
This repository implements the training and testing of ADLD for Unconstrained Facial Action Unit Detection via Latent Feature Domain. The repository offers the implementation of the paper in PyTorch
Getting Started
Installation
- This code was tested with PyTorch 0.4.0 and Python 2.7
- Clone this repo:
git clone https://github.com/ZhiwenShao/ADLD
cd ADLD
Datasets
Put BP4D and EmotioNet into the folder "dataset" following the paths shown in the list files of the folder "data/list"
Preprocessing
- Conduct similarity transformation for face images:
- We provide the landmarks annotated using OpenPose for EmotioNet here. Each line in the landmark annotation file corresponds to 49 facial landmark locations (x1,y1,x2,y2...). Put these annotation files into the folder "dataset"
- An example of processed image can be found in the folder "data/imgs/EmotioNet/optimization_set/N_0000000001/"
cd dataset python face_transform.py - Compute the weight of the loss of each AU in the BP4D training set:
- The AU annoatation files should be in the folder "data/list"
cd dataset python write_AU_weight.py
Training
- Train a model without using target-domain pseudo AU labels:
python train.py --mode='weak'
- Train a model using target-domain pseudo AU labels:
python train.py --mode='full'
Testing
- Test a model trained without using target-domain pseudo AU labels:
python test.py --mode='weak'
- Test a model trained using target-domain pseudo AU labels:
python test.py --mode='full'
Citation
If you use this code for your research, please cite our paper:
@article{shao2021unconstrained,
title={Unconstrained Facial Action Unit Detection via Latent Feature Domain},
author={Shao, Zhiwen and Cai, Jianfei and Cham, Tat-Jen and Lu, Xuequan and Ma, Lizhuang},
journal={IEEE Transactions on Affective Computing},
year={2021},
publisher={IEEE}
}
Should you have any questions, just contact with us through email zhiwen_shao@cumt.edu.cn
Related Skills
node-connect
353.3kDiagnose OpenClaw node connection and pairing failures for Android, iOS, and macOS companion apps
frontend-design
111.7kCreate distinctive, production-grade frontend interfaces with high design quality. Use this skill when the user asks to build web components, pages, or applications. Generates creative, polished code that avoids generic AI aesthetics.
openai-whisper-api
353.3kTranscribe audio via OpenAI Audio Transcriptions API (Whisper).
qqbot-media
353.3kQQBot 富媒体收发能力。使用 <qqmedia> 标签,系统根据文件扩展名自动识别类型(图片/语音/视频/文件)。
