MDNet is a state-of-the-art deep learning model designed for emotion recognition using multimodal physiological signals, with a focus on cross-subject generalizability. This work extends existing methodologies by introducing a dual disentangling mechanism that efficiently captures both modality-invariant and modality-specific features along with subject-shared and subject-private features, enhancing performance in cross-subject settings.
This repository contains the implementation of the model described in the paper titled "Multi-level Disentangling Network for Cross-Subject Emotion Recognition based on Multimodal Physiological Signals" by Ziyu Jia, Fengming Zhao, Yuzhe Guo, Hairong Chen, and Tianzi Jiang.
The paper has been accepted for publication at IJCAI 2024 [paper].
Code for the model in the paper Multi-level Disentangling Network for Cross-Subject Emotion Recognition based on Multimodal Physiological Signals [github].
The architecture of MDNet includes two primary modules:
- Modality-level disentangling module: Separates modality-invariant and modality-specific features from multimodal physiological signals.
- Subject-level disentangling module: Separates subject-shared and subject-private features to handle individual differences across subjects effectively.
- CUDA 11.4
- Python 3.8
- Pytorch 2.3.0
This section outlines the organization of the repository and the purpose of its directories and files to help you navigate and utilize this project effectively.
MDNet/
│
├── functions.py
# Contains loss functions used by MDNet.
│
├── modality_Encoder.py
# Modality encoder that extract initial features. EEG_Encoder can only be used to process EEG signals. EOG_Encoder can be adapted to process EOG, ECG, EMG signals.
│
├── models.py
# Contains the implementation of the MDNet model, detailing the architecture and its functionalities.
│
├── LICENSE
# The license file, delineating the terms under which the project can be used.
│
└── README.md
# Provides a comprehensive overview of the project, including its purpose, structure, and instructions for setup and usage.
If you find this useful, please cite our work as follows:
@inproceedings{DBLP:conf/ijcai/JiaZGCJ24,
author = {Ziyu Jia and
Fengming Zhao and
Yuzhe Guo and
Hairong Chen and
Tianzi Jiang},
title = {Multi-level Disentangling Network for Cross-Subject Emotion Recognition
Based on Multimodal Physiological Signals},
booktitle = {Proceedings of the Thirty-Third International Joint Conference on
Artificial Intelligence, {IJCAI} 2024, Jeju, South Korea, August 3-9,
2024},
pages = {3069--3077},
publisher = {ijcai.org},
year = {2024},
url = {https://www.ijcai.org/proceedings/2024/340},
timestamp = {Fri, 18 Oct 2024 20:53:37 +0200},
biburl = {https://dblp.org/rec/conf/ijcai/JiaZGCJ24.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
- For academic and non-commercial use only
- Apache License 2.0