Abstract
During each cardiac cycle of heart, vibrations creates sound and murmur. When these sound and murmur wave is represented graphically then it is called phonocardiogram (PCG). Digital stethoscope is used to record the audio wave signals generated due to heart vibration. Audio waves recorded through digital stethoscope can be used to fetch information like tone, quality, intensity, frequency, heart rate etc. Based on the heart condition, this information will be different for different people and can be used to predict the status of heart at early stage in non-invasive manner. In this research work, by using deep learning models, authors have classified PCG signals into 5 classes namely extra systole, extra heart sound, artifacts, normal heartbeat and murmur. Initially spectrograms in the form of images are extracted from PCG sound and feed into Regularized Convolutional Neural Network. From the simulation environment designed in python, it has found that proposed model has shown the average accuracy of 94% while doing the classification of PCG sound in five classes.
Similar content being viewed by others
References
Banerjee, M., & Majhi, S. (2020, October). Multi-class heart sounds classification using 2D-convolutional neural network. In 2020 5th International conference on computing, communication and security (ICCCS) (pp. 1–6), IEEE.
CVD data. Retrieved April 7, 2021, from https://www.who.int/cardiovascular_diseases/about_cvd/en/
Cable, C. (1997). The aucultation assistant. Retrieved April 3, 2014, from www.med.ucla.edu/wilkes/intro.html.
Deng, M., Meng, T., Cao, J., Wang, S., Zhang, J., & Fan, H. (2020). Heart sound classification based on improved MFCC features and convolutional recurrent neural networks. Neural Networks, 130, 22–32.
Deperlioglu, O., Kose, U., Gupta, D., Khanna, A., & Sangaiah, A. K. (2020). Diagnosis of heart diseases by a secure internet of health things system based on autoencoder deep neural network. Computer Communications, 162, 31–50.
Dwivedi, A. K., Imtiaz, S. A., & Rodriguez-Villegas, E. (2018). Algorithms for automatic analysis and classification of heart sounds—a systematic review. IEEE Access, 7, 8316–8345.
Huang, Y., Hou, H., Wang, Y., Zhang, Y., & Fan, M. (2020). A long sequence speech perceptual hashing authentication algorithm based on constant Q transform and tensor decomposition. IEEE Access, 8, 34140–34152.
Jain, A., Tiwari, S., & Sapra, V. (2019). Two-phase heart disease diagnosis system using deep learning. International Journal of Control and Automation, 12(5), 558–573.
Janssens, A. C. J., & Martens, F. K. (2020). Reflection on modern methods: Revisiting the area under the ROC curve. International Journal of Epidemiology, 49(4), 1397–1403.
Kannan, R., & Vasanthi, V. (2019). Machine learning algorithms with ROC curve for predicting and diagnosing the heart disease. In Soft computing and medical bioinformatics (pp. 63–72). Springer, Singapore.
Kumar, A., Shaikh, A. M., Li, Y., Bilal, H., & Yin, B. (2020). Pruning filters with L1-norm and capped L1-norm for CNN compression. Applied Intelligence, 51, 1152–1160.
Lichtenberg, R. (n.d.). Heart sounds. Retrieved May 1, 2014, from http://www.loyolauniversity.adam.com.
Mannor, S. (2011). The PASCAL classifying heart sounds challenge. http://www.peterjbentley.com/heartchallenge/
Mateo, C., & Talavera, J. A. (2020). Bridging the gap between the short-time Fourier transform (STFT), wavelets, the constant-Q transform and multi-resolution STFT. Signal, Image and Video Processing, 14, 1535–1543.
Özcan, Z., & Kayıkçıoğlu, T. (2021). Evaluating MFCC-based speaker identification systems with data envelopment analysis. Expert Systems with Applications, 168, 114448.
Patidar, S., Pachori, R. B., & Garg, N. (2015). Automatic diagnosis of septal defects based on tunable-Q wavelet transform of cardiac sound signals. Expert Systems with Applications, 42(7), 3315–3326.
PhysioNet/Computing in Cardiology Challenge. (2016). Classification of normal/abnormal heart sound recordings. Retrieved May 31, 2018, from https://www.physionet.org/challenge/2016/.
Rani, P., Kumar, R., Jain, A., & Chawla, S. K. (2021). A hybrid approach for feature selection based on genetic algorithm and recursive feature elimination. International Journal of Information System Modeling and Design (IJISMD), 12(2), 17–38. https://doi.org/10.4018/IJISMD.2021040102
Shuvo, S. B., Ali, S. N., & Swapnil, S. I. (2020). CardioXNet: A novel lightweight CRNN framework for classifying cardiovascular diseases from phonocardiogram recordings. https://doi.org/10.1109/ACCESS.2021.3063129
Son, G. Y., & Kwon, S. (2018). Classification of heart sound signal using multiple features. Applied Sciences, 8(12), 2344.
Stillman. (2007). ALDMD clinical cardiology tools from Hennepin county medical center. Retrieved April 8, 2014, from http://www.aldmd.com.
Tiwari, S. (2020). A comparative study of deep learning models with handcraft features and non-handcraft features for automatic plant species identification. International Journal of Agricultural and Environmental Information Systems (IJAEIS), 11(2), 44–57.
Tiwari, S. (2021a). Dermatoscopy using multi-layer perceptron, convolution neural network, and capsule network to differentiate malignant melanoma from benign nevus. International Journal of Healthcare Information Systems and Informatics (IJHISI), 16(3), 58–73.
Tiwari, S. (2021b). An ensemble deep neural network model for onion-routed traffic detection to boost cloud security. International Journal of Grid and High Performance Computing (IJGHPC), 13(1), 1–17.
Tiwari, S., & Jain, A. (2021). Convolutional capsule network for COVID-19 detection using radiography images. International Journal of Imaging Systems and Technology, 31, 525–539.
University of Michigan Heart Sound and Murmur Library. Retrieved March 16, 2014, from http://www.med.umich.edu.
Upretee, P., & Yüksel, M. E. (2019, April). Accurate classification of heart sounds for disease diagnosis by a single time-varying spectral feature: Preliminary results. In 2019 Scientific meeting on electrical-electronics & biomedical engineering and computer science (EBBT) (pp. 1–4). IEEE.
Wang, M., Wang, R., Zhang, X. L., & Rahardja, S. (2019, November). Hybrid constant-Q transform based CNN ensemble for acoustic scene classification. In 2019 Asia-Pacific Signal and Information Processing Association annual summit and conference (APSIPA ASC) (pp. 1511–1516). IEEE.
Wilson, J. M. (2009). Heart sound pod cast series. Retrieved April 6, 2014, from www.texasheartinstitute.org.
Zhai, Y., Deng, W., Xu, Y., Ke, Q., Gan, J., Sun, B., & Piuri, V. (2019). Robust SAR automatic target recognition based on transferred MS-CNN with L2-regularization. Computational Intelligence and Neuroscience. https://doi.org/10.1155/2019/9140167
Zheng, F., Zhang, G., & Song, Z. (2001). Comparison of different implementations of MFCC. Journal of Computer Science and Technology, 16(6), 582–589.
Author information
Authors and Affiliations
Contributions
All authors have equally contributed.
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Yadav, K., Tiwari, S., Jain, A. et al. Deep learning based cardiovascular disease diagnosis system from heartbeat sound. Int J Speech Technol (2021). https://doi.org/10.1007/s10772-021-09890-4
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s10772-021-09890-4