PCGNet
The official implementation for paper: A Persona-Infused Cross-Task Graph Network for Multimodal Emotion Recognition with Emotion Shift Detection in Conversations, SIGIR 2024.
Install / Use
/learn @HITSZ-HLT/PCGNetREADME
A Persona-Infused Cross-Task Graph Network for Multimodal Emotion Recognition with Emotion Shift Detection in Conversations
The official implementation for paper: A Persona-Infused Cross-Task Graph Network for Multimodal Emotion Recognition with Emotion Shift Detection in Conversations, SIGIR 2024.
<img src="https://img.shields.io/badge/Venue-SIGIR--24-blue" alt="venue"/> <img src="https://img.shields.io/badge/Status-Accepted-success" alt="status"/> <img src="https://img.shields.io/badge/Issues-Welcome-red">
Requirements
- Python 3.10.13
- PyTorch 1.13.1
- torch_geometric 2.4.0
- torch-scatter 2.1.0
- torch-sparse 0.5.15
- CUDA 11.7
Preparation
- Download multimodal-features
- Save data/iemocap/iemocap_features_roberta.pkl, data/iemocap/IEMOCAP_features.pkl in
data/iemocap/; Save meld_features_roberta.pkl, data/meld/MELD_features_raw1.pkl indata/meld/.
Training & Evaluation
- train PCGNet on IEMOCAP for ERC task
python code/run_train_erc.py --dataset IEMOCAP --data_dir data/iemocap/IEMOCAP_features.pkl \
--valid_rate 0.0 --modals avl --lr 0.0001 --batch-size 32 --l2 0.0001 --dropout 0.2 --gamma 0.5 --class_weight --reason_flag \
--mtl --use_clone --hidden_l 400 --hidden_a 400 --hidden_v 400 --persona_l_heads 8 --persona_a_heads 8 --persona_v_heads 8 \
--persona_l_layer 1 --persona_a_layer 1 --persona_v_layer 1 --interactive_layer 1 --interactive_heads 4 --dropout_forward 0.3 \
--dropout_persona_lstm_modeling 0.2 --dropout_interactive 0.2 --dropout_persona 0.2 --erc_windows 1 --shift_windows 1 \
--persona_transform --interactive_windows 1 --epochs 140 --seed 6500
- train PCGNet on MELD for ERC task
python code/run_train_erc.py --dataset MELD --data_dir ./data/meld/MELD_features_raw1.pkl \
--valid_rate 0.0 --modals avl --lr 0.0001 --batch-size 32 --l2 0.0001 \
--mtl --use_clone --hidden_l 200 --hidden_a 200 --hidden_v 200 --persona_transform\
--persona_l_heads 4 --persona_a_heads 4 --persona_v_heads 4 \
--persona_l_layer 1 --persona_a_layer 1 --persona_v_layer 1 \
--interactive_layer 1 --interactive_heads 4 \
--dropout_forward 0 --dropout_persona_lstm_modeling 0.2 --dropout_interactive 0.2 --dropout_persona 0.2 \
--erc_windows 1 --shift_windows 1 --interactive_windows 1 --epochs 30 --seed 11407
- evaluation PCGNet on IEMOCAP for ERC task
python code/inference.py --dataset IEMOCAP --data_dir data/iemocap/IEMOCAP_features.pkl \
--valid_rate 0.0 --modals avl --lr 0.0001 --batch-size 32 --l2 0.0001 --dropout 0.2 --gamma 0.5 --class_weight --reason_flag \
--mtl --use_clone --hidden_l 400 --hidden_a 400 --hidden_v 400 --persona_l_heads 8 --persona_a_heads 8 --persona_v_heads 8 \
--persona_l_layer 1 --persona_a_layer 1 --persona_v_layer 1 --interactive_layer 1 --interactive_heads 4 --dropout_forward 0.3 \
--dropout_persona_lstm_modeling 0.2 --dropout_interactive 0.2 --dropout_persona 0.2 --erc_windows 1 --shift_windows 1 \
--persona_transform --interactive_windows 1 --seed 6500 --ckpt checkpoints/IEMOCAP_ckpt.pkl
- evaluation PCGNet on MELD for ERC task
python code/inference.py --dataset MELD --data_dir ./data/meld/MELD_features_raw1.pkl \
--valid_rate 0.0 --modals avl --lr 0.0001 --batch-size 32 --l2 0.0001 \
--mtl --use_clone --hidden_l 200 --hidden_a 200 --hidden_v 200 --persona_transform\
--persona_l_heads 4 --persona_a_heads 4 --persona_v_heads 4 \
--persona_l_layer 1 --persona_a_layer 1 --persona_v_layer 1 \
--interactive_layer 1 --interactive_heads 4 \
--dropout_forward 0 --dropout_persona_lstm_modeling 0.2 --dropout_interactive 0.2 --dropout_persona 0.2 \
--erc_windows 1 --shift_windows 1 --interactive_windows 1 --seed 11407 --ckpt checkpoints/MELD_ckpt.pkl
Citation
If you find our work useful for your research, please kindly cite our paper as follows:
@inproceedings{tu2024persona,
title = {A Persona-Infused Cross-Task Graph Network for Multimodal Emotion Recognition with Emotion Shift Detection in Conversations},
author = {Tu, Geng and Xiong, Feng and Liang, Bin and Xu, Ruifeng},
booktitle={Proceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval},
pages = {2266–2270},
year = {2024}
}
Acknowledgements
Special thanks to the following authors for their contributions through open-source implementations.
Related Skills
node-connect
354.3kDiagnose OpenClaw node connection and pairing failures for Android, iOS, and macOS companion apps
frontend-design
112.3kCreate distinctive, production-grade frontend interfaces with high design quality. Use this skill when the user asks to build web components, pages, or applications. Generates creative, polished code that avoids generic AI aesthetics.
openai-whisper-api
354.3kTranscribe audio via OpenAI Audio Transcriptions API (Whisper).
qqbot-media
354.3kQQBot 富媒体收发能力。使用 <qqmedia> 标签,系统根据文件扩展名自动识别类型(图片/语音/视频/文件)。
