Full text is available at the source.
Attention-based convolutional neural network with multi-modal temporal information fusion for motor imagery EEG decoding
Using an attention-based neural network combining time-based data to decode movement-related brain signals
AI simplified
Abstract
A 4-class average accuracy of 85.03% was achieved on the BCIC-IV-2a dataset using a novel deep learning network.
- The proposed network combines convolutional neural networks with a self-attention mechanism to enhance EEG decoding.
- Multi-modal temporal information is extracted from both average and variance perspectives to better capture neural dynamics.
- Global dependencies are captured through a shared self-attention module designed for the extracted feature dimensions.
- A convolutional encoder explores the relationship between average and variance features to create more discriminative outputs.
- A new data augmentation method, signal segmentation and recombination, is introduced to improve the network's generalization capability.
AI simplified