面向运动想象解码的双尺度时空特征融合网络
Dual-Scale Spatiotemporal Feature Fusion Network for Motor Imagery Decoding
摘要: 卷积神经网络(Convolutional neural networks, CNNs)已在运动想象(Motor Imagery, MI)脑机接口(Brain-Computer Interface, BCI)领域取得了广泛应用和良好效果。然而,基于单一尺度卷积结构的模型难以从脑电信号中充分挖掘多样化信息;而现有的多尺度CNN虽然能够提取不同尺度的特征,但通常仅通过简单拼接的方式进行融合,难以实现多尺度信息之间的深度协同建模。为了解决上述问题,相关研究提出了双尺度时空特征融合网络(Dual-Scale Spatiotemporal Feature Fusion Network, DSSFFN)。该方法通过双尺度卷积分支以提取脑电信号不同尺度的特征,并通过Transformer模块对来自不同尺度的特征进行融合,从而提升特征的判别性。本文采用BCI竞赛IV的2a数据集进行实验,同时与多个前沿的运动想象算法进行比较。结果显示,DSSFFN在数据集上平均准确率为72.65%,高于所有的对比模型,展现出了DSSFFN模型在运动想象脑电识别任务上的优异性能。此外,研究还通过消融实验分析了双尺度时空卷积分支与Transformer模块对整体性能的贡献,验证了各个关键模块的有效性。同时比较了不同融合模型对模型影响,进一步验证了Transformer模块融合双尺度的有效性。
Abstract: Convolutional neural networks (CNNs) have been widely applied in motor imagery (MI) brain-computer interface (BCI) systems and have achieved promising performance. However, models based on single-scale convolutional structures are often insufficient to fully exploit the diverse and rich information embedded in EEG signals. Although existing multi-scale CNNs can extract features at different scales, they usually fuse multi-scale features through simple concatenation, which makes it difficult to achieve deep and effective collaboration among features from different scales. To address these issues, related research proposes a Dual-Scale Spatiotemporal Feature Fusion Network (DSSFFN). This method employs dual-scale convolutional branches to extract EEG features at different scales, and further integrates the features from different branches using a Transformer module, thereby enhancing the discriminative capability of the feature representations. In this study, experiments are conducted on the BCI Competition IV-2a dataset, and the proposed method is compared with several state-of-the-art MI recognition algorithms. The results show that DSSFFN achieves an average classification accuracy of 72.65% on this dataset, outperforming all competing methods, thus demonstrating its superior performance in motor imagery EEG classification tasks. Furthermore, ablation studies are performed to analyze the contributions of the dual-scale spatiotemporal convolution branches and the Transformer fusion module, which verifies the effectiveness of each key component in the proposed network. In addition, the impacts of different fusion models on the proposed model’s performance were compared, which further verifies the effectiveness of the Transformer module in fusing dual-scale features.
文章引用:陈志城. 面向运动想象解码的双尺度时空特征融合网络[J]. 计算机科学与应用, 2026, 16(2): 514-521. https://doi.org/10.12677/csa.2026.162080

参考文献

[1] Wei, F., Xu, X., Jia, T., Zhang, D. and Wu, X. (2023) A Multi-Source Transfer Joint Matching Method for Inter-Subject Motor Imagery Decoding. IEEE Transactions on Neural Systems and Rehabilitation Engineering, 31, 1258-1267. [Google Scholar] [CrossRef] [PubMed]
[2] Tang, X., Yang, C., Sun, X., Zou, M. and Wang, H. (2023) Motor Imagery EEG Decoding Based on Multi-Scale Hybrid Networks and Feature Enhancement. IEEE Transactions on Neural Systems and Rehabilitation Engineering, 31, 1208-1218. [Google Scholar] [CrossRef] [PubMed]
[3] Müller-Gerking, J., Pfurtscheller, G. and Flyvbjerg, H. (1999) Designing Optimal Spatial Filters for Single-Trial EEG Classification in a Movement Task. Clinical Neurophysiology, 110, 787-798. [Google Scholar] [CrossRef] [PubMed]
[4] Ang, K.K., Chin, Z.Y., Zhang, H., et al. (2008) Filter Bank Common Spatial Pattern (FBCSP) in Brain-Computer Interface. 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence), Hong Kong, 1-8 June 2008, 2390-2397. [Google Scholar] [CrossRef
[5] Schirrmeister, R.T., Springenberg, J.T., Fiederer, L.D.J., Glasstetter, M., Eggensperger, K., Tangermann, M., et al. (2017) Deep Learning with Convolutional Neural Networks for EEG Decoding and Visualization. Human Brain Mapping, 38, 5391-5420. [Google Scholar] [CrossRef] [PubMed]
[6] Lawhern, V.J., Solon, A.J., Waytowich, N.R., Gordon, S.M., Hung, C.P. and Lance, B.J. (2018) EEGNet: A Compact Convolutional Neural Network for EEG-Based Brain-Computer Interfaces. Journal of Neural Engineering, 15, Article 056013. [Google Scholar] [CrossRef] [PubMed]
[7] Ingolfsson, T.M., Hersche, M., Wang, X., Kobayashi, N., Cavigelli, L. and Benini, L. (2020) EEG-TCNet: An Accurate Temporal Convolutional Network for Embedded Motor-Imagery Brain-Machine Interfaces. 2020 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Toronto, 11-14 October 2020, 2958-2965. [Google Scholar] [CrossRef
[8] Mane, R., Chew, E., Chua, K., et al. (2021) FBCNet: A Multi-View Convolutional Neural Network for Brain-Computer Interface.
[9] Altaheri, H., Muhammad, G. and Alsulaiman, M. (2022) Physics-Informed Attention Temporal Convolutional Network for EEG-Based Motor Imagery Classification. IEEE Transactions on Industrial Informatics, 19, 2249-2258. [Google Scholar] [CrossRef
[10] Altaheri, H., Muhammad, G. and Alsulaiman, M. (2023) Dynamic Convolution with Multilevel Attention for EEG-Based Motor Imagery Decoding. IEEE Internet of Things Journal, 10, 18579-18588. [Google Scholar] [CrossRef
[11] Zhao, W., Jiang, X., Zhang, B., Xiao, S. and Weng, S. (2024) CTNet: A Convolutional Transformer Network for EEG-Based Motor Imagery Classification. Scientific Reports, 14, Article No. 20237. [Google Scholar] [CrossRef] [PubMed]
[12] Dai, G., Zhou, J., Huang, J. and Wang, N. (2020) HS-CNN: A CNN with Hybrid Convolution Scale for EEG Motor Imagery Classification. Journal of Neural Engineering, 17, Article 016025. [Google Scholar] [CrossRef] [PubMed]
[13] Chakladar, D.D., Kumar, P., Roy, P.P., Dogra, D.P., Scheme, E. and Chang, V. (2021) A Multimodal-Siamese Neural Network (mSNN) for Person Verification Using Signatures and EEG. Information Fusion, 71, 17-27. [Google Scholar] [CrossRef
[14] Salami, A., Andreu-Perez, J. and Gillmeister, H. (2022) EEG-ITNet: An Explainable Inception Temporal Convolutional Network for Motor Imagery Classification. IEEE Access, 10, 36672-36685. [Google Scholar] [CrossRef
[15] Santamaría-Vázquez, E., Martínez-Cagigal, V., Vaquerizo-Villar, F. and Hornero, R. (2020) EEG-Inception: A Novel Deep Convolutional Neural Network for Assistive Erp-Based Brain-Computer Interfaces. IEEE Transactions on Neural Systems and Rehabilitation Engineering, 28, 2773-2782. [Google Scholar] [CrossRef] [PubMed]
[16] Zhang, C., Kim, Y. and Eskandarian, A. (2021) EEG-Inception: An Accurate and Robust End-to-End Neural Network for EEG-Based Motor Imagery Classification. Journal of Neural Engineering, 18, Article 046014. [Google Scholar] [CrossRef] [PubMed]
[17] Hu, J., Shen, L. and Sun, G. (2018) Squeeze-and-Excitation Networks. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, 18-23 June 2018, 7132-7141. [Google Scholar] [CrossRef