*Result*: [A study on heart sound classification algorithm based on improved Mel-frequency cepstrum coefficient feature extraction and deep Transformer].
Original Publication: Chengdu : Sichuan Sheng sheng wu yi xue gong cheng xue hui : Hua xi yi ke da xue : Chengdu ke ji da xue
Sheng Wu Yi Xue Gong Cheng Xue Za Zhi. 2024 Oct 25;41(5):977-985. (PMID: 39462666)
Comput Methods Programs Biomed. 2021 Mar;200:105940. (PMID: 33494031)
Comput Biol Med. 2018 Sep 1;100:132-143. (PMID: 29990646)
IEEE/ACM Trans Comput Biol Bioinform. 2024 Jul-Aug;21(4):936-947. (PMID: 37027654)
Biomed Res Int. 2020 Jan 10;2020:5846191. (PMID: 32420352)
J Med Syst. 2019 May 6;43(6):168. (PMID: 31056720)
Annu Int Conf IEEE Eng Med Biol Soc. 2022 Jul;2022:4469-4472. (PMID: 36085633)
IEEE J Biomed Health Inform. 2024 Dec;28(12):7151-7163. (PMID: 39028592)
Physiol Meas. 2016 Dec;37(12):2181-2213. (PMID: 27869105)
Sci Rep. 2024 Nov 25;14(1):29132. (PMID: 39587136)
Sheng Wu Yi Xue Gong Cheng Xue Za Zhi. 2021 Oct 25;38(5):969-978. (PMID: 34713665)
Biomed Tech (Berl). 2023 Feb 15;68(3):285-295. (PMID: 36780471)
Comput Methods Programs Biomed. 2024 May;248:108122. (PMID: 38507960)
Neural Netw. 2020 Oct;130:22-32. (PMID: 32589588)
Front Physiol. 2022 Dec 22;13:1084420. (PMID: 36620204)
Local Abstract: [Publisher, Chinese] 心音在心血管疾病早期检测中至关重要,现有的研究成果集中于传统的信号分割、特征提取和使用浅层分类器的阶段。为解决特征提取不充分难以充分捕捉心音的动态和非线性特征、对复杂心音信号识别能力不足、易受数据不平衡影响、识别性能差的问题,本文提出了采用改进梅尔频率倒谱系数(MFCC)提取特征,并基于卷积神经网路(CNN)和深层变换器(Transformer)的新型心音分类方法。在预处理阶段,本研究使用巴特沃斯滤波器去噪,无需分割心音周期,直接对连续心音信号进行改进的MFCC特征提取,以捕捉其动态特征;然后,将动态特征输入到CNN中学习,再采用全局平均池化(GAP)降低模型权重并防止过拟合;最终,利用深层Transformer模块进一步提取和融合特征,完成心音分类。为应对数据不平衡问题,本文模型采用焦点损失(focal loss)作为损失函数。通过在两个公开数据集上进行的实验验证结果表明,本文所提方法在二分类和多分类任务中均较为有效且表现出一定优势。本文旨在实现对连续心音信号进行高效分类,为后续心音分类疾病研究提供参考方法,有助于可穿戴设备和家庭监测系统的开发。.
*Further Information*
*Heart sounds are critical for early detection of cardiovascular diseases, yet existing studies mostly focus on traditional signal segmentation, feature extraction, and shallow classifiers, which often fail to sufficiently capture the dynamic and nonlinear characteristics of heart sounds, limit recognition of complex heart sound patterns, and are sensitive to data imbalance, resulting in poor classification performance. To address these limitations, this study proposes a novel heart sound classification method that integrates improved Mel-frequency cepstral coefficients (MFCC) for feature extraction with a convolutional neural network (CNN) and a deep Transformer model. In the preprocessing stage, a Butterworth filter is applied for denoising, and continuous heart sound signals are directly processed without segmenting the cardiac cycles, allowing the improved MFCC features to better capture dynamic characteristics. These features are then fed into a CNN for feature learning, followed by global average pooling (GAP) to reduce model complexity and mitigate overfitting. Lastly, a deep Transformer module is employed to further extract and fuse features, completing the heart sound classification. To handle data imbalance, the model uses focal loss as the objective function. Experiments on two public datasets demonstrate that the proposed method performs effectively in both binary and multi-class classification tasks. This approach enables efficient classification of continuous heart sound signals, provides a reference methodology for future heart sound research for disease classification, and supports the development of wearable devices and home monitoring systems.*