基于SE-TCN-Transformer融合网络的下肢外骨骼机器人运动意图识别研究
Research on Motion Intention Recognition for Lower-Limb Exoskeleton Robots Based on SE-TCN-Transformer Fusion Network
摘要: 针对下肢外骨骼机器人人机交互系统中对运动意图识别的高精度与实时性要求,本文提出了一种基于SE-TCN-Transformer的新型融合网络架构。现有的卷积神经网络(CNN)和长短期记忆网络(LSTM)在处理长时序多模态传感器数据时,往往面临局部特征提取不足、全局依赖捕捉困难以及难以有效抑制传感器噪声等挑战。为解决上述问题,本研究设计了“局部特征增强–全局语义建模”的双阶段特征提取策略。首先,引入嵌入Squeeze-and-Excitation (SE)通道注意力机制的时间卷积网络(Temporal Convolutional Network, TCN)作为前端,通过因果空洞卷积提取高频局部运动特征,并利用SE模块自适应重标定通道权重,有效筛选关键传感器信息并抑制冗余噪声。其次,采用Transformer编码器对特征序列进行全局建模,利用多头自注意力机制精准捕捉步态相位间的长距离时序依赖。在公开数据集HuGaDB上的实验结果表明,该模型对8种复杂日常步态的平均识别准确率达到96.73%,F1分数达到95.43%。此外,基于Shapley Additive exPlanations (SHAP)的可解释性分析揭示,足部垂直加速度与大腿角速度是步态识别中作用较大,表明模型提取的关键特征分布与人体下肢生物力学的先验知识具有较高的一致性。特别是在坐下、起立等具有瞬态突变特征的转换动作识别中,该模型表现出显著优势,优于传统的SVM、LSTM及CNN-LSTM混合模型。该研究不仅验证了混合架构在多模态步态分析中的有效性,也为外骨骼机器人的实时精准控制提供了有力的技术支撑。
Abstract: To achieve high-precision and real-time motion intention recognition for lower-limb exoskeletons, this paper proposes a novel hybrid architecture: the SE-TCN-Transformer. Addressing the limitations of conventional CNNs and LSTMs in processing long-sequence multimodal data, we introduce a “local enhancement-global modeling” strategy. First, a Temporal Convolutional Network (TCN) integrated with Squeeze-and-Excitation (SE) attention extracts high-frequency local features while adaptively suppressing sensor noise. Second, a Transformer encoder captures long-range temporal dependencies via multi-head self-attention. Experiments on the HuGaDB dataset demonstrate an average accuracy of 96.73% and an F1-score of 95.43% across eight gaits. SHAP-based interpretability analysis identifies foot vertical acceleration and thigh angular velocity as the most critical features, demonstrating a higher consistency with biomechanical priors. The proposed model significantly outperforms traditional SVM, LSTM, and CNN-LSTM baselines, particularly in recognizing transient actions like “Sit” and “Stand”, thereby providing robust technical support for real-time exoskeleton control.
文章引用:高婧祯, 蓝艺亮, 张国成, 刘欣怡, 颜建军. 基于SE-TCN-Transformer融合网络的下肢外骨骼机器人运动意图识别研究[J]. 人工智能与机器人研究, 2026, 15(2): 660-671. https://doi.org/10.12677/airr.2026.152063

参考文献

[1] Belal, M., Alsheikh, N., Aljarah, A. and Hussain, I. (2024) Deep Learning Approaches for Enhanced Lower-Limb Exoskeleton Control: A Review. IEEE Access, 12, 143883-143907. [Google Scholar] [CrossRef
[2] 李畅, 郑义, 王优强, 等. 下肢康复外骨骼机器人研究现状及应用进展[J]. 机械传动, 2025, 49(11): 1-16.
[3] Lau, H., Tong, K. and Zhu, H. (2008) Support Vector Machine for Classification of Walking Conditions Using Miniature Kinematic Sensors. Medical & Biological Engineering & Computing, 46, 563-573. [Google Scholar] [CrossRef] [PubMed]
[4] Casale, P., Pujol, O. and Radeva, P. (2011) Human Activity Recognition from Accelerometer Data Using a Wearable Device. In: Vitrià, J., Sanches, J.M. and Hernández, M., Eds., Pattern Recognition and Image Analysis, Springer, 289-296. [Google Scholar] [CrossRef
[5] Figo, D., Diniz, P.C., Ferreira, D.R. and Cardoso, J.M.P. (2010) Preprocessing Techniques for Context Recognition from Accelerometer Data. Personal and Ubiquitous Computing, 14, 645-662. [Google Scholar] [CrossRef
[6] Jiang, W. and Yin, Z. (2015) Human Activity Recognition Using Wearable Sensors by Deep Convolutional Neural Networks. Proceedings of the 23rd ACM International Conference on Multimedia, Brisbane, 26-30 October 2015, 1307-1310. [Google Scholar] [CrossRef
[7] Ordóñez, F. and Roggen, D. (2016) Deep Convolutional and LSTM Recurrent Neural Networks for Multimodal Wearable Activity Recognition. Sensors, 16, Article 115. [Google Scholar] [CrossRef] [PubMed]
[8] Jeon, S., Choi, J. and Cheong, J. (2020) Gait Phase Estimation from Lower Limb Orientation Using Deep Learning. IEEE Access, 8, 217207-217216.
[9] Fang, H. and Liang, Y. (2019) Human Activity Recognition Based on Feature Selection and CNN-LSTM. 2019 IEEE International Conference on Computer and Communications (ICCC), Chengdu, 6-9 December 2019, 1997-2001.
[10] Bai, S., Kolter, J.Z. and Koltun, V. (2018) An Empirical Evaluation of Generic Convolutional and Recurrent Networks for Sequence Modeling. arXiv: 1803.01271
[11] Vaswani, A., Shazeer, N., Parmar, N., et al. (2017) Attention Is All You Need. Advances in Neural Information Processing Systems (NeurIPS) 2017, Long Beach, 4-9 December 2017, 5998-6008.
[12] Song, C., Huang, Y., Huang, Y., Jia, N. and Wang, L. (2019) GaitNet: An End-To-End Network for Gait Based Human Identification. Pattern Recognition, 96, Article ID: 106988. [Google Scholar] [CrossRef
[13] Hu, J., Shen, L. and Sun, G. (2018) Squeeze-And-Excitation Networks. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, 18-23 June 2018, 7132-7141. [Google Scholar] [CrossRef
[14] Chereshnev, R. and Kertész-Farkas, A. (2017) HuGaDB: Human Gait Database for Activity Recognition from Wearable Inertial Sensor Networks. In: van der Aalst, W., et al., Eds., Analysis of Images, Social Networks and Texts, Springer, 131-141. [Google Scholar] [CrossRef