EMPT: a sparsity Transformer for EEG-based motor imagery recognition

Liu, Ming and Liu, Yanbing and Shi, Weiyou and Lou, Yitai and Sun, Yuan and Meng, Qi and Wang, Dezheng and Xu, Fangzhou and Zhang, Yang and Zhang, Lei and Leng, Jiancai (2024) EMPT: a sparsity Transformer for EEG-based motor imagery recognition. Frontiers in Neuroscience, 18. ISSN 1662-453X

[thumbnail of fnins-18-1366294.pdf] Text
fnins-18-1366294.pdf - Published Version

Download (3MB)

Abstract

Introduction: Transformer network is widely emphasized and studied relying on its excellent performance. The self-attention mechanism finds a good solution for feature coding among multiple channels of electroencephalography (EEG) signals. However, using the self-attention mechanism to construct models on EEG data suffers from the problem of the large amount of data required and the complexity of the algorithm.

Methods: We propose a Transformer neural network combined with the addition of Mixture of Experts (MoE) layer and ProbSparse Self-attention mechanism for decoding the time-frequency-spatial domain features from motor imagery (MI) EEG of spinal cord injury patients. The model is named as EEG MoE-Prob-Transformer (EMPT). The common spatial pattern and the modified s-transform method are employed for achieving the time-frequency-spatial features, which are used as feature embeddings to input the improved transformer neural network for feature reconstruction, and then rely on the expert model in the MoE layer for sparsity mapping, and finally output the results through the fully connected layer.

Results: EMPT achieves an accuracy of 95.24% on the MI EEG dataset for patients with spinal cord injury. EMPT has also achieved excellent results in comparative experiments with other state-of-the-art methods.

Discussion: The MoE layer and ProbSparse Self-attention inside the EMPT are subjected to visualisation experiments. The experiments prove that sparsity can be introduced to the Transformer neural network by introducing MoE and kullback-leibler divergence attention pooling mechanism, thereby enhancing its applicability on EEG datasets. A novel deep learning approach is presented for decoding EEG data based on MI.

Item Type: Article
Subjects: Academic Digital Library > Medical Science
Depositing User: Unnamed user with email info@academicdigitallibrary.org
Date Deposited: 18 Apr 2024 12:23
Last Modified: 18 Apr 2024 12:23
URI: http://publications.article4sub.com/id/eprint/3274

Actions (login required)

View Item
View Item