A novel feature-level fusion scheme with multimodal attention CNN for heart sound classification

被引:3
作者
Ranipa, Kalpeshkumar [1 ]
Zhu, Wei -Ping [1 ]
Swamy, M. N. S. [1 ]
机构
[1] Concordia Univ, Dept Elect & Comp Engn, Montreal, PQ, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Heart sound classification; Feature fusion; Neural network; Deep learning; DEEP; ENSEMBLE;
D O I
10.1016/j.cmpb.2024.108122
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Background and objective: Most of the existing machine learning -based heart sound classification methods achieve limited accuracy. Since they primarily depend on single domain feature information and tend to focus equally on each part of the signal rather than employing a selective attention mechanism. In addition, they fail to exploit convolutional neural network (CNN) - based features with an effective fusion strategy. Methods: In order to overcome these limitations, a novel multimodal attention convolutional neural network (MACNN) with a feature -level fusion strategy, in which Mel-cepstral domain as well as general frequency domain features are incorporated to increase the diversity of the features, is proposed in this paper. In the proposed method, DilationAttenNet is first utilized to construct attention -based CNN feature extractors and then these feature extractors are jointly optimized in MACNN at the feature -level. The attention mechanism aims to suppress irrelevant information and focus on crucial diverse features extracted from the CNN. Results: Extensive experiments are carried out to study the efficacy of the feature level fusion in comparison to that with early fusion. The results show that the proposed MACNN method significantly outperforms the state-of-the-art approaches in terms of accuracy and score for the two publicly available Github and Physionet datasets. Conclusion: The findings of our experiments demonstrated the high performance for heart sound classification based on the proposed MACNN, and hence have potential clinical usefulness in the identification of heart diseases. This technique can assist cardiologists and researchers in the design and development of heart sound classification methods.
引用
收藏
页数:12
相关论文
共 37 条
[1]   Detection of pathological heart sounds [J].
Abdollahpur, Mostafa ;
Ghaffari, Ali ;
Ghiasi, Shadi ;
Mollakazemi, M. Javad .
PHYSIOLOGICAL MEASUREMENT, 2017, 38 (08) :1616-1630
[2]   Classification of heart sounds using fractional fourier transform based mel-frequency spectral coefficients and traditional classifiers [J].
Abduh, Zaid ;
Nehary, Ebrahim Ameen ;
Wahed, Manal Abdel ;
Kadah, Yasser M. .
BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2020, 57
[3]   Convolutional and recurrent neural networks for the detection of valvular heart diseases in phonocardiogram recordings [J].
Alkhodari, Mohanad ;
Fraiwan, Luay .
COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE, 2021, 200
[4]   Heart Sound Multiclass Analysis Based on Raw Data and Convolutional Neural Network [J].
Avanzato, Roberta ;
Beritelli, Francesco .
IEEE SENSORS LETTERS, 2020, 4 (12)
[5]   Phonocardiogram Signal Processing for Automatic Diagnosis of Congenital Heart Disorders through Fusion of Temporal and Cepstral Features [J].
Aziz, Sumair ;
Khan, Muhammad Umar ;
Alhaisoni, Majed ;
Akram, Tallha ;
Altaf, Muhammad .
SENSORS, 2020, 20 (13) :1-20
[6]   Automatic diagnosis of multiple cardiac diseases from PCG signals using convolutional neural network [J].
Baghel, Neeraj ;
Dutta, Malay Kishore ;
Burget, Radim .
COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE, 2020, 197
[7]   Analysis of heart sound anomalies using ensemble learning [J].
Baydoun, Mohammed ;
Safatly, Lise ;
Ghaziri, Hassan ;
El Hajj, Ali .
BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2020, 62
[8]  
Chen LC, 2017, Arxiv, DOI [arXiv:1706.05587, DOI 10.48550/ARXIV.1706.05587]
[9]   COMPARISON OF PARAMETRIC REPRESENTATIONS FOR MONOSYLLABIC WORD RECOGNITION IN CONTINUOUSLY SPOKEN SENTENCES [J].
DAVIS, SB ;
MERMELSTEIN, P .
IEEE TRANSACTIONS ON ACOUSTICS SPEECH AND SIGNAL PROCESSING, 1980, 28 (04) :357-366
[10]   Towards the classification of heart sounds based on convolutional deep neural network [J].
Demir, Fatih ;
Sengur, Abdulkadir ;
Bajaj, Varun ;
Polat, Kemal .
HEALTH INFORMATION SCIENCE AND SYSTEMS, 2019, 7 (01)