Skip to content
forked from xmh1011/AT-DGNN

Attention-Based Temporal Learner With Dynamical Graph Neural Network for EEG Emotion Recognition

License

Notifications You must be signed in to change notification settings

chen51628/AT-DGNN

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

21 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

AT-DGNN

PWC

Attention-Based Temporal Learner With Dynamical Graph Neural Network for EEG Emotion Recognition.

Introduction

The MEEG dataset, capturing emotional responses to various musical stimuli across different valence and arousal levels, enables an in-depth analysis of brainwave patterns within musical contexts. We introduce the Attention-based Temporal Learner with Dynamic Graph Neural Network (AT-DGNN), a novel framework for EEG emotion recognition. By integrating an attention mechanism with a dynamic graph neural network (DGNN), the AT-DGNN model captures complex local and global EEG dynamics, demonstrating superior performance with accuracy of 83.74% in arousal and 86.01% in valence, outperforming current state-of-the-art (SOTA) methods.

Paper

MEEG and AT-DGNN: Improving EEG Emotion Recognition with Music Introducing and Graph-based Learning

Network

AT-DGNN

The AT-DGNN model comprises two core modules: a feature extraction module (a) and a dynamic graph neural network learning module (b). The feature extraction module consists of a temporal learner, a multi-head attention mechanism, and a temporal convolution module. These components effectively leverage local features of EEG signals through a sliding window technique, thereby enhancing the model's capacity to dynamically extract complex temporal patterns in EEG signals. In the graph-based learning module, the model initially employs local filtering layers to segment and filter features from specific brain regions. Subsequently, the architecture employs three layers of stacked dynamic graph convolutions to capture complex interactions among different brain regions. This structure enhances the AT-DGNN's capacity for integrating temporal features effectively.

Run

The source code is totally compatible with DEAP dataset and MEEG dataset. You can refer to the run to run the code.

Dataset

If you are interested in the dataset, you can refer to the dataset to download the dataset.

Visualization

The visualization of the EEG signals can be found in the visualization.

Citation

If you find our work useful, please consider citing our paper:

@misc{xiao2024meegatdgnnimprovingeeg,
      title={MEEG and AT-DGNN: Improving EEG Emotion Recognition with Music Introducing and Graph-based Learning}, 
      author={Minghao Xiao and Zhengxi Zhu and Bin Jiang and Meixia Qu and Wenyu Wang},
      year={2024},
      eprint={2407.05550},
      archivePrefix={arXiv},
      primaryClass={cs.HC},
      url={https://arxiv.org/abs/2407.05550}, 
}

Acknowledgement

The music for introducing the emotional state of the participants in the MEEG dataset is provided by Rui Zhang Prof., Shandong University, Department of Music.

Some of the source code is originally from LGGNet. We appreciate the authors for their contribution.

Reference

The other models compared with AT-DGNN and unitized in the source code are listed in the reference.

About

Attention-Based Temporal Learner With Dynamical Graph Neural Network for EEG Emotion Recognition

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%