基于多域特征融合与通道注意力的异常心音检测
Anomaly Heart Sound Detection Based on Multi-Domain Feature Fusion and Channel Attention
摘要: 心音信号蕴含丰富的生理与病理信息,是辅助心血管疾病筛查与早期诊断的重要生物特征。传统心音分析方法通常依赖单一特征或手工规则,难以充分刻画时频结构的复杂变化,导致在多噪声环境和不同个体差异下识别准确率受限。为此,本文提出一种基于多域特征融合与通道注意力机制的心音异常检测卷积网络(TTFI-CNN)。该网络综合利用MFCC (Mel频率倒谱系数)、STFT (短时傅里叶变换)与DWT (离散小波)三类互补特征,构建“时域–频域–时频域”的多特征表示,并设计轻量化SE (Squeeze-and-Excitation)通道注意力模块增强特征选择能力。此外,模型还结合EarlyStopping与交叉验证设计,以提升泛化能力。基于2016 PhysioNet/CinC心音挑战赛数据集进行验证,所提出的模型在异常心音检测任务上取得97.8%的准确率。其中,消融实验证明:去除DWT分支会导致准确率下降至96.4%,去除SE注意力则下降至97.09%,充分说明三分支融合与通道注意力在提升分类性能方面的有效性。实验结果表明,本方法能够稳健地捕获多尺度时频特征,对心音异常检测具有良好的应用潜力。
Abstract: Heart sound signals contain rich physiological and pathological information, serving as crucial biomarkers for assisting in the screening and early diagnosis of cardiovascular diseases. Traditional heart sound analysis methods typically rely on single features or manually designed rules, making it difficult to fully characterize the complex variations in time-frequency structures. This often results in limited recognition accuracy in environments with high noise levels or significant individual differences. To address this, this paper proposes a time-frequency feature fusion and channel attention-based convolutional neural network (TTFI-CNN) for abnormal heart sound detection. The network comprehensively utilizes three complementary features—Mel-Frequency Cepstral Coefficients (MFCC), Short-Time Fourier Transform (STFT), and Discrete Wavelet Transform (DWT)—to construct a multi-feature representation spanning the “time-domain, frequency-domain, and time-frequency domain”. Additionally, a lightweight Squeeze-and-Excitation (SE) channel attention module is designed to enhance feature selection capabilities. Furthermore, the model incorporates EarlyStopping and cross-validation strategies to improve generalization. Validated on the 2016 PhysioNet/CinC heart sound challenge dataset, the proposed model achieves an accuracy of 97.8% in the task of abnormal heart sound detection. Ablation studies demonstrate that removing the DWT branch reduces accuracy to 96.4%, while removing the SE attention module lowers it to 97.09%, fully confirming the effectiveness of the three-branch fusion and channel attention in enhancing classification performance. The experimental results indicate that this method can robustly capture multi-scale time-frequency features, demonstrating promising potential for application in abnormal heart sound detection.
文章引用:和含雷, 李子阳, 郑钟月, 朱加飞, 张丽艳. 基于多域特征融合与通道注意力的异常心音检测[J]. 图像与信号处理, 2026, 15(2): 155-164. https://doi.org/10.12677/jisp.2026.152013

参考文献

[1] 国家心血管病中心, 中国心血管健康与疾病报告编写组. 中国心血管健康与疾病报告2024概要[J]. 中国循环杂志, 2025, 40(6): 521-559.
[2] Zeinali, Y. and Niaki, S.T.A. (2022) Heart Sound Classification Using Signal Processing and Machine Learning Algorithms. Machine Learning with Applications, 7, Article ID: 100206. [Google Scholar] [CrossRef
[3] Krishnan, P.T., Balasubramanian, P. and Umapathy, S. (2020) Automated Heart Sound Classification System from Unsegmented Phonocardiogram (PCG) Using Deep Neural Network. Physical and Engineering Sciences in Medicine, 43, 505-515. [Google Scholar] [CrossRef] [PubMed]
[4] Nguyen, M.T., Lin, W.W. and Huang, J.H. (2022) Heart Sound Classification Using Deep Learning Techniques Based on Log-Mel Spectrogram. Circuits, Systems, and Signal Processing, 42, 344-360. [Google Scholar] [CrossRef
[5] Liu, C., Springer, D., Li, Q., Moody, B., Juan, R.A., Chorro, F.J., et al. (2016) An Open Access Database for the Evaluation of Heart Sound Algorithms. Physiological Measurement, 37, 2181-2213. [Google Scholar] [CrossRef] [PubMed]
[6] Potes, C., Parvaneh, S., Rahman, A. and Conroy, B. (2016). Ensemble of Feature: Based and Deep Learning: Based Classifiers for Detection of Abnormal Heart Sounds. 2016 Computing in Cardiology Conference (CinC), Vancouver, 11-14 September 2016, 621-624.[CrossRef
[7] Yaseen, Son, G. and Kwon, S. (2018) Classification of Heart Sound Signal Using Multiple Features. Applied Sciences, 8, Article No. 2344. [Google Scholar] [CrossRef
[8] Li, F., Liu, C., Wang, X. and Liu, C. (2017) Classification of Heart Sounds Using Convolutional Neural Networks Based on Time-Frequency Representations. Computers in Biology and Medicine, 84, 67-76.
[9] Mubarak, Q., Akram, M.U., Shaukat, A., Hussain, F., Khawaja, S.G. and Butt, W.H. (2018) Analysis of PCG Signals Using Quality Assessment and Homomorphic Filters for Localization and Classification of Heart Sounds. Computer Methods and Programs in Biomedicine, 164, 143-157. [Google Scholar] [CrossRef] [PubMed]
[10] Li, F., Tang, H., Shang, S., et al. (2020) Classification of Heart Sounds Using Convolutional Neural Network. Applied Sciences, 10, Article No. 3956.
[11] Zhang, W., Han, J. and Deng, S. (2017) Heart Sound Classification Based on Scaled Spectrogram and Partial Least Squares Regression. Biomedical Signal Processing and Control, 32, 20-28. [Google Scholar] [CrossRef
[12] Ren, J., Liu, C., Zhang, H. and Wang, X. (2021) Heart Sound Classification Using Deep Convolutional Neural Networks and Mel-Frequency Spectral Features. Biomedical Signal Processing and Control, 68, Article ID: 102720.
[13] Kumar, A., Sharma, L.D. and Sunkaria, R.K. (2022) Attention-Based Deep Learning Model for Heart Sound Classification. Biomedical Signal Processing and Control, 71, Article ID: 103173.