Next Article in Journal
Transient Response of Miniature Piezoresistive Pressure Sensor Dedicated to Blast Wave Monitoring
Next Article in Special Issue
Measuring the Effectiveness of a Multicomponent Program to Manage Academic Stress through a Resilience to Stress Index
Previous Article in Journal
Three-Dimensional Encoding Approach for Wearable Tactile Communication Devices
Previous Article in Special Issue
Optimized Acoustic Phantom Design for Characterizing Body Sound Sensors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Classification of Heart Sounds Using Chaogram Transform and Deep Convolutional Neural Network Transfer Learning

by
Ali Harimi
1,
Yahya Majd
2,
Abdorreza Alavi Gharahbagh
3,
Vahid Hajihashemi
3,
Zeynab Esmaileyan
1,
José J. M. Machado
4 and
João Manuel R. S. Tavares
4,*
1
Department of Electrical Engineering, Shahrood Branch, Islamic Azad University, Shahrood 43189-36199, Iran
2
School of Surveying and Built Environment, Toowoomba Campus, University of Southern Queensland (USQ), Darling Heights, QLD 4350, Australia
3
Faculdade de Engenharia, Universidade do Porto, Rua Dr. Roberto Frias, s/n, 4200-465 Porto, Portugal
4
Departamento de Engenharia Mecânica, Faculdade de Engenharia, Universidade do Porto, Rua Dr. Roberto Frias, s/n, 4200-465 Porto, Portugal
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(24), 9569; https://doi.org/10.3390/s22249569
Submission received: 12 November 2022 / Revised: 4 December 2022 / Accepted: 4 December 2022 / Published: 7 December 2022

Abstract

:
Heart sounds convey important information regarding potential heart diseases. Currently, heart sound classification attracts many researchers from the fields of telemedicine, digital signal processing, and machine learning—among others—mainly to identify cardiac pathology as quickly as possible. This article proposes chaogram as a new transform to convert heart sound signals to colour images. In the proposed approach, the output image is, therefore, the projection of the reconstructed phase space representation of the phonocardiogram (PCG) signal on three coordinate planes. This has two major benefits: (1) it makes possible to apply deep convolutional neural networks to heart sounds and (2) it is also possible to employ a transfer learning scheme by converting a heart sound signal to an image. The performance of the proposed approach was verified on the PhysioNet dataset. Due to the imbalanced data on this dataset, it is common to assess the results quality using the average of sensitivity and specificity, which is known as score, instead of accuracy. In this study, the best results were achieved using the I n c e p t i o n V 3 model, which achieved a score of 88.06%.

1. Introduction

Recent studies show that cardiovascular diseases ( C V D s ), with approximately 18 million deaths annually, have become one of the leading causes of mortality. Thus, the early detection of heart disease is an important issue due to its crucial role in saving people’s lives. This puts studies in related fields as a canonical problem involving various fields of research and applications.
Various C V D s and cardiac pathologies, such as coronary heart disease, congenital heart disease, and peripheral arterial disease, can be diagnosed through cardiac auscultation. Additionally, most heart diseases related to heart valves are detectable through the phonocardiography (PCG) signal. Therefore, the automatic detection of cardiac abnormalities through PCG signals has attracted the interest of many researchers. It can be considered a multidisciplinary research topic involving telemedicine; digital signals, i.e., audio processing; and machine learning. Hence, this research field aims to analyse the PCG signal, i.e., the heart sound, to detect any abnormality in heart functioning. This is known as PCG signal classification, mainly into normal or abnormal signals.
Each period of the PCG signal comprises two major states: systole and diastole. Systole, with a typical time duration of 300 ms, consists of a 70–150 ms S1 state, wherein the mitral and tricuspid valves are closely followed by a silent segment. Diastole, with a typical duration of 500 ms, includes the S2 state related to the closure of aortic and pulmonary valves and high-frequency S3 and S4 noises that are usually not heard. Figure 1 illustrates a typical PCG signal and its four states: S1, S2, S3, and S4.
Figure 1 shows how different heart symptoms affect the temporal dynamics of PCG signals. Accordingly, while heart sound segmentation algorithms aim to segment the four major states of the PCG signal, heart sound classification algorithms tend to develop machine learning-based algorithms to classify normal heart sounds versus abnormal heart sounds. It is common to detect abnormal heart function through electrocardiogram ( E C G ) signal analysis. However, acquiring PCG signals is more straightforward, cheaper, and less invasive than E C G signals. Additionally, analysing PCG signals allows experts to check the patient and help to make a quick initial decision about the patient’s condition.
There are three main tasks related to the processing and analysis of PCG signals: signal denoising, segmentation, and classification. Denoising aims to provide a clean PCG signal using, for example, noise cancellation techniques. It is therefore common to use denoising as a preprocessing stage. Segmentation detects the four major states of the PCG signal throughout each cardiac cycle, and, finally, PCG classification categorises the heart functioning under analysis into normal and abnormal.
According to the classic point of view of the PCG or similar audio classification problems, the current problem can be addressed based on two major steps: finding specifications, i.e., feature extraction, and their categorisation, i.e., classification. Thus, first, in the preprocessing step, a noise cancellation method is usually applied to remove noise [1,2], and then, the temporal–spectral information is gathered as numerical data. The extraction of discriminative features is one of the main challenges in this area. Mell Frequency Spectrum Coefficients [3,4,5,6], Hilbert–Huang transform [7], S transform [8], Wavelet transform [9,10,11,12], and multi-domain features [13] are well-known feature extractors that have been proposed for PCG classification.
Usually, the parameters of the used classification model are adjusted to classify normal versus abnormal signals during its training phase. Then, in the test phase, the trained model is employed to predict whether an input signal is normal or abnormal. In contrast to the classic methods, deep learning algorithms have partially solved the feature extraction issue by implementing networks that can automatically derive features from the raw input signal throughout their pipeline. Hence, these kinds of algorithms learn to extract discriminative features through training; their only disadvantage is the requirement of large-scale training datasets and proper hardware. A comprehensive survey on employing deep neuronal networks ( D N N s ) for PCG signal classification is given, for example, in [14].
Many related works use the raw signal as the network’s input [15]. On the other hand, others perform a feature extraction task before inputting the signal to the network. For example, in [5], low-level features were extracted from the input PCG signal, and then a D N N was applied to derive high-level features and learn their relationship with the output classes. Furthermore, various D N N models and architectures have been proposed for this task. In [5], a long-short term memory ( L S T M ) D N N s was employed. In [5], recurrent neural networks ( R N N s ) were used, while [16] suggested autoencoders for the classification of heart sounds. On the other hand, although some researchers use one-dimensional (1D) convolutional neural networks [17,18], others suggest deep convolutional neural networks ( D C N N s ). While D C N N s are the most favourable networks for image processing, some researchers try to apply them to other signal processing problems. To this end, converting the original one-dimensional signal into corresponding two-dimensional (2D) tensors is recommended before applying the D C N N to benefit from the advantages of 2D convolutions [19]. In [20], frequency features were extracted using a CNN, and temporal features were extracted using a recurrent neural network. Then, the fusion of these features was inputted into a lightweight neural network for heart sound classification. In [21,22,23], the authors suggested time-frequency representation of the original signal such as a spectrogram image as a proper input for 2D DCNNs. For example, in [24], a signal-to-cycle conversion preprocessing step was employed to build the spectrogram based on such signal cycles instead of fixed-length time frames. Then, a hybrid classifier composed of AlexNet and SVM was employed, which achieved promising results even using just 2–3 s of data. Regarding the success of DCNNs in various applications, the motivation for this study is the effective application of DCNNs for classifying PCG signals. Therefore, one should adopt the PCG signal as a proper input for DCNNs.
Despite the success of DCNNs in various applications, these models comprise a large number of trainable parameters which demand a high amount of memory and processing capacity. Some studies suggest the use of network pruning and weight quantization to address this issue [25,26], which allows for the implementation of DCNNs on simple processors such as microcontrollers. However, in this study, the main focus was more on the classification performance and not on the computational complexity, which will be addressed in the near future.
Based on our knowledge, previous studies mostly convert the original one-dimensional signal into a 2D spectrogram image using techniques based on a fusion step. Therefore, the main contribution of this study is the representation of the original one-dimensional PCG signal as a 2D chaogram image and input this image into a pretrained DCNN model. Chaogram was first introduced in [27] for speech emotion recognition. It uses a reconstructed phase space, a well-known nonlinear dynamic processing tool useful in analysing chaotic systems, to convert the input PCG signal into a 2D image. Therefore, the chaogram transform can reflect the chaotic behaviour of a non-linear dynamic system on the output R G B image. Moreover, it allows for using 2D D C N N s to classify PCG signals.
The remainder of this article is organised as follows. The details of the proposed approach are given in Section 2. To assess the effectiveness and efficiency of the proposed approach, a group of experiments was designed and performed, and the experimental setup and results are presented in Section 3. Finally, the conclusion is provided in Section 4.

2. Proposed Approach

The proposed approach for computer-aided heart auscultation comprises a chaogram transformation applied to the original PCG signal and a DNN model that learns features and classifies them. Figure 2 depicts the proposed approach, which is described in detail as follows: After describing the two-folded preprocessing scheme in Section 2.1, Section 2.2 describes the transformation of the original one-dimensional PCG signal into the corresponding 2D chaogram image; finally, the pretrained convolutional neural network (CNN) model that was used is described in Section 2.3.

2.1. Preprocessing

The heart sound is affected by other sounds of the human body. Therefore, it is necessary to employ preprocessing and noise cancellation techniques. Whitaker et al. used a two-stage noise cancellation technique. The authors used a third-order Butterworth band-pass filter with bandwidths of 15 to 800 Hz. Then, the spectral subtraction denoising scheme was employed [28]. The authors used adaptive filters based on noise power outside the expected range of the heart sound spectrum, which performed well in removing background noise presented in biological signals such as speech and EEG. Finally, the clean PCG signal was obtained by subtracting the weighted version of predicted noise power from the PCG spectrum [28].

2.2. Transformation of One-Dimensional PCG Signal into a Chaogram Image

To build a chaogram image, reconstructed phase space (RPS) is a beneficial tool for analysing a system’s nonlinear and chaotic behaviour [29]. A phase space comprises the collection of all possible states of a system. However, determining all the possible states of real systems is usually impossible. Instead, the output signal of the system under study can be used to reconstruct its phase space. This signal inherits the main characteristics of its system. Here, the PCG signal is considered to be the output of the heart, i.e., the main system; then, the RPS can be built by defining the vectors:
S n ¯ = [ S n , S n + τ , S n + 2 τ , , S n + ( d 1 ) τ ] ,
where S n , with n = 1 , 2 , 3 N , is the nth sample of the PCG signal; d and τ denote the embedding dimension and time delay, respectively; and S n ¯ is a linear vector that specifies one of the possible states of the system as a point in R P S , which is determined by the consecutive collection of such vectors as:
S = [ S 1 ¯ , S 2 ¯ , , S n ¯ ] T ,
where S denotes the R P S , and T stands for the transpose operation. One of the central issues in R P S is determining the optimal values for τ and d based on mutual information and the false nearest neighbour algorithm, respectively [29]. The mutual information between two X and Y signals can be calculated as:
I ( X , Y ) = y Y x X P ( X , Y ) ( x , y ) log P ( X , Y ) ( x , y ) P ( X ) ( x ) P ( Y ) ( y ) ,
where p ( x ) stands for the probability density function of x, and p ( x , y ) denotes the joint probability density function of x and y variables. Here, for all the samples in the PhysioNet dataset, the mutual information of the PCG signals and their corresponding delayed versions were calculated for the time lags of 0 (zero) to 50, as shown in Figure 3. As seen from the graphs in Figure 3, the first minimum of mutual information offered by different samples is a digit between 10 and 20. According to Figure 3, the first minimum of mutual information in the averaged curve is placed on the time lag of 18. Therefore, it was considered that τ = 18 . Although, based on the experiments performed in this study, the optimum dimensionality for embedding dimension for PCG signals is 4, we considered that d = 3 in order to build the used chaogram images.
The R P S of a PCG signal is considered to model the nonlinear dynamics of the heart as the system that generates the PCG signal. Observations also confirm that the patterns formed in the R P S of PCG are strongly correlated with the heart’s functioning condition. Figure 4 shows six PCG signals (three normal and three abnormal) and their representation in RPS (with τ = 18 and d = 3 ). In the original R P S , a curve goes from one point to another according to a sequence (rows 2 and 5 of Figure 4). By eliminating the links between the points, the R P S would be converted into a set of points in a 3D space (rows 3 and 6 of Figure 4). The shapes of these data points form cloud-shape patterns giving fundamental information about the corresponding system [30].
When the R P S of a PCG signal is calculated using Equation (2), the space is partitioned into parts in each direction. Consequently, the space is a 224 × 224 × 224 cell net, while each vector in Equation (1) is considered a point in this net. Then, the frequency of points inside each cell is calculated to form a 3D tensor, T. After that, the projections of this 3D tensor on X Y , X Z , and Y Z planes are calculated as three images, I x y , I x z , and I y z , as:
I x y ( i , j ) = K = 1 224 T ( i , j , k ) ,
I x z ( i , k ) = j = 1 224 T ( i , j , k ) ,
I y z ( j , k ) = i = 1 224 T ( i , j , k ) ,
where T denotes the 3D tensor. Consequently, three images with the size of 224 × 224 are obtained. Finally, these three images act as colour channels of an R G B image to build the chaogram image. An image enhancement stage is employed to emphasise the weak details in these images. Figure 5 shows the flow diagram of this process.
This study employs a D C N N to learn the relationship between the cloud shapes and patterns in the chaogram image and the heart functioning condition. Since the size of the PhysioNet dataset is relatively small, the training of a deep neural network with a considerable number of variables is impossible since the risk of overfitting is unavoidable. Therefore, pretrained networks are recommended based on a transfer learning ( T L ) scheme. Hence, the chaogram images had to be of the same size as the input data admitted by the employed model. Hence, the size of the chaogram images must be equal to the input size admitted by the used VGG DCNN model, which is a 224 × 224 × 3   R G B image, if it is intended to employ it as a pretrained model. Figure 6 shows the chaogram images of the 6 PCG samples shown in Figure 4.
As shown in Figure 6, while the patterns of the chaogram images shown in each row resemble each other, the images of the two rows differ. In other words, it seems that the chaogram images provide a high intraclass similarity and interclass difference.

2.3. DCNN Model

In this study, four pretrained deep learning models, A l e x N e t , V G G 16 , I n c e p t i o n V 3 , and R e s N e t 50 , were evaluated for the classification of PCG signals using chaogram images. The A l e x N e t D C N N model [31] won the ImageNet I L S V R C 2012 challenge. It comprises five convolutional and three fully connected layers. The V G G 16 D C N N model gained the second rank in the ImageNet I L S V R C 2014 challenge with a slight difference from the first model: G o o g L e N e t network [32]. V G G 16 includes 16 convolutional layers. G o o g L e N e t consists of nine Inception modules with trainable hyperparameters. I n c e p t i o n v 3 , the upgraded version of Inception, has 11 Inception modules [33]. On the other hand, R e s N e t 50 (residual network) consists of 50 layers [34], and it won the I m a g e N e t I L S V R C 2015 challenge.
As already mentioned, training the four deep learning networks from scratch on the relatively small sample size dataset, PhysioNet, increases the risk of overfitting. Therefore, to reduce this risk, the transfer learning technique—where only the two last layers of the network were allowed to fine-tune on the used dataset, while the earlier layers were kept unchanged with the weights pretrained on the large-scale ImageNet dataset—was employed. Transfer learning is a common technique used in deep learning, particularly in D C N N models, since it can speed up the learning rate and improve the model’s generalisation [35,36]. Additionally, data augmentation was used in this study to enrich the used training dataset by building new samples by applying various transformations and deformations to the original samples. Therefore, the rotation of the chaogram image (with angles of 5, 10, 15, 20, 25, and 30 degrees), scale (with factors of 1.05, 1.1, 1.15, and 1.2), width shift (by 5, 10, 15, 20, and 25 pixels), and height shift (by 5, 10, 15, 20, and 25 pixels) were applied to the original images to generate the final training data. As a result, 20 ( 6 × 4 × 5 × 5 ) new versions for each original sample were generated, and, in total, the number of the training samples increased 21 times: from 2868 to 60,228 ( 2868 × 21 ) samples. Additionally, to also reduce the risk of overfitting, the dropout technique was used, where each neuron is selected to be removed with a predefined probability. It makes the learning process more independent from the neurons and simplifies the network, thus preventing overfitting. A dropout layer with p = 0.5 was used here to implement this technique.
The f i n e t u n i n g process in a D C N N network is controlled with a set of hyperparameters, including, optimiser functions: A d a m , S G D , R M S p r o p , A d a d e l t a , A d a g r a d , A d a m a x , N a d a m , and F t r l ; learning rate: 1 × 10 5 to 1 × 10 1 according to steps of 1 × 10 5 ; batch size: 50 to 500 according to steps of 2; and epochs: 50 to 400 according to steps of 20. These parameters can be selected based on a trial-and-error method. However, optimising such hyperparameters can improve the classification model’s performance. In this study, the Bayesian optimisation algorithm [37] in the “ b a y e s o p t ” package in Python was used for the f i n e t u n i n g of the hyperparameters. The stopping condition for the optimisation algorithm was when there was no improvement in 20 consecutive epochs.

3. Experimental Setup and Results

3.1. Experimental Setup

The proposed approach was evaluated on the P h y s i o N e t dataset. A computational system with an A M D R Y Z E N 9 C o r e i 9 4900 H C P U , a N V I D I A G T X 1660 T i g r a p h i c s c a r d , and 16GB D D R 4 of R A M was used. All simulations were performed in the Anaconda environment using S p y d e r I D E , and the algorithms were coded in Python. The phase space reconstruction was implemented using the skedm library. In addition, pretrained D C N N models were implemented using the TensorFlow framework and Keras library [38].

3.2. Dataset

The PhysioNet/CinC [39] dataset was used to evaluate the proposed approach. The included samples are not all similar due to different recording environments, equipment, and time duration. The dataset consists of 2868 PCG signals, including 2249 normal and 619 abnormal samples that were resampled to 2000 Hz and saved in ".wav” format [39]. The samples are organized into five subfolders (a, b, c, d, and e), with significant differences and variations in each subfolder due to their origin and recording condition. This dataset was built aiming the classification of heart sounds, such as in this study and in many other related works that also used it [5,10,40].

3.3. Evaluation Metrics

To evaluate the performance of the classification process, the confusion matrix was computed with the abnormal cases as the positive class, i.e., the normal cases, and the sensitivity, specificity, and accuracy were calculated using:
S e n s i t i v i t y ( S e ) = T P T P + F N ,
S p e c i f i c i t y ( S p ) = T N T N + F P ,
A c c u r a c y = T P + T N T P + T N + F P + F N .
where T P , T N , F P , and F N are the confusion matrix components representing true positive, true negative, false positive, and false negative cases, respectively. Since the data in the PhysioNet dataset is imbalanced, the accuracy tends to be biased toward the majority class and cannot measure misclassification costs. Hence, it is not the most suitable measure to be used in the evaluation of the classification performance. Therefore, the average between sensitivity and specificity was chosen as an alternative official evaluation metric for the 2016 PhysioNet/Computing in Cardiology Challenge, which is calculated as:
S c o r e = S e + S p T P + T N + F P + F N .

3.4. Experimental Results

The experiments were performed using the hyperparameters selected by the Bayesian optimisation algorithm, as reported in Table 1. Table 2 presents the results of the experiments, which were performed based on the 5-fold cross-validation technique.
According to Table 2, all tested models performed well in classifying the used PCG signals using the built chaogram images. This indicates the ability of the proposed transformation to convert PCG signals to R G B images. In other words, it implicitly shows that the employed models can efficiently extract useful features from the built chaogram images and consequently obtain results with a high classification rate. Among the employed models, the I n c e p t i o n V 3 model obtained the highest score of for accuracy (88.06), sensitivity (84.49), and specificity (91.63). On the other hand, the lowest accuracy rate was obtained by A l e x N e t .
However, although it seems pretty logical, what was the cause of the difference between the results of the four models? Undoubtedly, one can think that the main reason can be the different architectures of the used models. The optimum depth of a network model depends on the size of the used dataset or the number of trainable parameters, while a few trainable parameters of a too-shallow network may not be enough for learning the general rules of a complicated classification problem. In a deep network with numerous trainable parameters, the models may learn examples instead of rules, which is called overfitting. Therefore, depending on the size of the used dataset and complexity of the problem, the appropriate network model should be carefully selected.
To compare the proposed method against state-of-the-art methods, it was implemented according to same conditions of the ones suggested in [41,42,43,44]. Table 3 compares the results obtained by the proposed method and the ones obtained by state-of-the-art methods. Based on the data presented in Table 3, one can realize that the proposed method shows better accuracy and recall than all other methods. The precision of the proposed method is slightly lower than the ones of the other methods, and the f1 score is only lower than the one of [44]. The best recall shows that the proposed method led to the lowest false positive rate, which is the most important parameter in medical care. In future works, adding a postprocessing or fusion step to the output is recommended to increase the precision and F1 score of the proposed method.

4. Conclusions

In this article, a new approach for the classification of PCG signals was proposed; its main contribution is the use of 2D chaogram images to detect abnormalities from heart sounds. From the experiments conducted, it is possible to use a vast range of image processing tools to analyse one-dimensional signals. For instance, this study employed two-dimensional D C N N s to analyse PCG signals from the corresponding built 2D chaogram images. Second, the chaogram images derived from the reconstructed phase space of the PCG signal under analysis inherited the main topological features of the cloud-shaped patterns in that space. Hence, they tend to carry important information about the nonlinear chaotic behaviour of the PCG signal. According to the observations made during this study, the built chaogram images are similar in each normal and abnormal class, while they differ from one class to another. In other words, this provides high intraclass similarity and interclass distance. Consequently, the convolutional masks in D C N N layers can extract meaningful information from those images. Third, converting PCG signals to chaogram images reduces the risk of overfitting due to the possibility of using transfer learning and pretrained D C N N s . This reduces the trainable parameters while effectively maintaining the classification model’s complexity. Therefore, complex problems with a few training samples can be solved with less risk of overfitting.
Furthermore, due to the imbalanced distribution of the data included in the used PhysioNet dataset, the accuracy cannot correctly reflect the classification performance. Instead, the score factor, which is determined as the average of sensitivity and specificity, was used to evaluate the classification performance. According to the experiments conducted, the use of different pretrained DCNNs did not greatly impact the achieved score factor. The best score was achieved using InceptionV3, and the lowest score was obtained by AlexNet: 88.06% and 86.88%, respectively. The small difference between these values confirms the suitability of the chaogram transform to represent the used heart sound data. On the other hand, the structure of the used model had a considerable impact on the final results of the proposed method. This is related to both operators employed in each layer of the used classifier model, the depth of its layers, and the corresponding number of trainable parameters. If the number of trainable parameters is too large, the model tends to be overfitted to the training data, which can be avoided if transfer learning is employed to pretrain the used model.
To build the corresponding chaogram image, the PCG signal is represented in the 3D reconstructed phase space. However, it is recommended to determine the optimum embedding dimension by using a false nearest algorithm, which is one of the main limitations of the current method. Hence, a potential future study could investigate a solution that can build the chaogram images based on the raw signal’s optimal dimensional reconstructed phase space as an effective step in generalizing this paper’s proposed method. For example, dimensional reduction approaches such as principal component analysis can be beneficial in this regard. Moreover, none of the employed DCNNs were optimized for classifying PCG signals. Thus, future works can endeavour to design a new dedicated neural network architecture that can increase classification accuracy.

Author Contributions

Conceptualization and supervision: J.M.R.S.T.; investigation, data collection, formal analysis, and writing—original draft preparation: A.H., Y.M., A.A.G. and V.H.; writing—review and editing: Z.E., J.J.M.M. and J.M.R.S.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, S.; Zhang, R.; Chang, S.; Liu, C.; Sha, X. A low-noise-level heart sound system based on novel thorax-integration head design and wavelet denoising algorithm. Micromachines 2019, 10, 885. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Xu, C.; Li, H.; Xin, P. Research on Heart Sound Denoising Method Based on CEEMDAN and Optimal Wavelet. In Proceedings of the 2022 2nd IEEE International Conference on Consumer Electronics and Computer Engineering (ICCECE), Guangzhou, China, 14–16 January 2022; pp. 629–632. [Google Scholar]
  3. Abduh, Z.; Nehary, E.A.; Wahed, M.A.; Kadah, Y.M. Classification of heart sounds using fractional fourier transform based mel-frequency spectral coefficients and traditional classifiers. Biomed. Signal Process. Control 2020, 57, 101788. [Google Scholar] [CrossRef]
  4. Chowdhury, T.H.; Poudel, K.N.; Hu, Y. Time-frequency analysis, denoising, compression, segmentation, and classification of PCG signals. IEEE Access 2020, 8, 160882–160890. [Google Scholar] [CrossRef]
  5. Deng, M.; Meng, T.; Cao, J.; Wang, S.; Zhang, J.; Fan, H. Heart sound classification based on improved MFCC features and convolutional recurrent neural networks. Neural Netw. 2020, 130, 22–32. [Google Scholar] [CrossRef]
  6. Hajihashemi, V.; Gharahbagh, A.A.; Cruz, P.M.; Ferreira, M.C.; Machado, J.J.; Tavares, J.M. Binaural Acoustic Scene Classification Using Wavelet Scattering, Parallel Ensemble Classifiers and Nonlinear Fusion. Sensors 2022, 22, 1535. [Google Scholar] [CrossRef]
  7. Arslan, Ö.; Karhan, M. Effect of Hilbert-Huang transform on classification of PCG signals using machine learning. J. King Saud Univ.-Comput. Inf. Sci. 2022. [Google Scholar] [CrossRef]
  8. Chen, P.; Zhang, Q. Classification of heart sounds using discrete time-frequency energy feature based on S transform and the wavelet threshold denoising. Biomed. Signal Process. Control 2020, 57, 101684. [Google Scholar] [CrossRef]
  9. Li, J.; Ke, L.; Du, Q. Classification of heart sounds based on the wavelet fractal and twin support vector machine. Entropy 2019, 21, 472. [Google Scholar] [CrossRef] [Green Version]
  10. Sawant, N.K.; Patidar, S.; Nesaragi, N.; Acharya, U.R. Automated detection of abnormal heart sound signals using Fano-factor constrained tunable quality wavelet transform. Biocybern. Biomed. Eng. 2021, 41, 111–126. [Google Scholar] [CrossRef]
  11. Zeng, W.; Yuan, J.; Yuan, C.; Wang, Q.; Liu, F.; Wang, Y. A new approach for the detection of abnormal heart sound signals using TQWT, VMD and neural networks. Artif. Intell. Rev. 2021, 54, 1613–1647. [Google Scholar] [CrossRef]
  12. Hajihashemi, V.; Alavigharahbagh, A.; Oliveira, H.S.; Cruz, P.M.; Tavares, J.M. Novel Time-Frequency Based Scheme for Detecting Sound Events from Sound Background in Audio Segments. In Iberoamerican Congress on Pattern Recognition; Springer: Berlin/Heidelberg, Germany, 2021; pp. 402–416. [Google Scholar] [CrossRef]
  13. Alshamma, O.; Awad, F.H.; Alzubaidi, L.; Fadhel, M.A.; Arkah, Z.M.; Farhan, L. Employment of multi-classifier and multi-domain features for PCG recognition. In Proceedings of the 2019 12th IEEE International Conference on Developments in eSystems Engineering (DeSE), Kazan, Russia, 7–10 October 2019; pp. 321–325. [Google Scholar]
  14. Chen, W.; Sun, Q.; Chen, X.; Xie, G.; Wu, H.; Xu, C. Deep learning methods for heart sounds classification: A systematic review. Entropy 2021, 23, 667. [Google Scholar] [CrossRef] [PubMed]
  15. Avanzato, R.; Beritelli, F. Heart sound multiclass analysis based on raw data and convolutional neural network. IEEE Sens. Lett. 2020, 4, 1–4. [Google Scholar] [CrossRef]
  16. Deperlioglu, O. Heart sound classification with signal instant energy and stacked autoencoder network. Biomed. Signal Process. Control 2021, 64, 102211. [Google Scholar] [CrossRef]
  17. Er, M.B. Heart sounds classification using convolutional neural network with 1D-local binary pattern and 1D-local ternary pattern features. Appl. Acoust. 2021, 180, 108152. [Google Scholar] [CrossRef]
  18. Xu, Y.; Xiao, B.; Bi, X.; Li, W.; Zhang, J.; Ma, X. Pay more attention with fewer parameters: A novel 1-D convolutional neural network for heart sounds classification. In Proceedings of the 2018 IEEE Computing in Cardiology Conference (CinC), Maastricht, The Netherlands, 23–26 September 2018; Volume 45, pp. 1–4. [Google Scholar]
  19. Bakhshi, A.; Harimi, A.; Chalup, S. CyTex: Transforming speech to textured images for speech emotion recognition. Speech Commun. 2022, 139, 62–75. [Google Scholar] [CrossRef]
  20. Li, S.; Li, F.; Tang, S.; Luo, F. Heart sounds classification based on feature fusion using lightweight neural networks. IEEE Trans. Instrum. Meas. 2021, 70, 1–9. [Google Scholar] [CrossRef]
  21. Khare, S.K.; Bajaj, V. Time–frequency representation and convolutional neural network-based emotion recognition. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 2901–2909. [Google Scholar] [CrossRef]
  22. Lopac, N.; Hržić, F.; Vuksanović, I.P.; Lerga, J. Detection of Non-Stationary GW Signals in High Noise From Cohen’s Class of Time–Frequency Representations Using Deep Learning. IEEE Access 2021, 10, 2408–2428. [Google Scholar] [CrossRef]
  23. Arias-Vergara, T.; Klumpp, P.; Vasquez-Correa, J.C.; Nöth, E.; Orozco-Arroyave, J.R.; Schuster, M. Multi-channel spectrograms for speech processing applications using deep learning methods. Pattern Anal. Appl. 2021, 24, 423–431. [Google Scholar] [CrossRef]
  24. Ismail, S.; Ismail, B.; Siddiqi, I.; Akram, U. PCG classification through spectrogram using transfer learning. Biomed. Signal Process. Control 2023, 79, 104075. [Google Scholar] [CrossRef]
  25. Huang, Q. Weight-Quantized SqueezeNet for Resource-Constrained Robot Vacuums for Indoor Obstacle Classification. AI 2022, 3, 180–193. [Google Scholar] [CrossRef]
  26. Novac, P.E.; Boukli Hacene, G.; Pegatoquet, A.; Miramond, B.; Gripon, V. Quantization and deployment of deep neural networks on microcontrollers. Sensors 2022, 21, 2984. [Google Scholar] [CrossRef] [PubMed]
  27. Falahzadeh, M.R.; Farokhi, F.; Harimi, A.; Sabbaghi-Nadooshan, R. Deep convolutional neural network and gray wolf optimization algorithm for speech emotion recognition. Circuits Syst. Signal Process. 2022, 1–44. [Google Scholar] [CrossRef]
  28. Whitaker, B.M.; Suresha, P.B.; Liu, C.; Clifford, G.D.; Anderson, D.V. Combining sparse coding and time-domain features for heart sound classification. Physiol. Meas. 2017, 38, 1701. [Google Scholar] [CrossRef] [PubMed]
  29. Shekofteh, Y.; Almasganj, F. Autoregressive modeling of speech trajectory transformed to the reconstructed phase space for ASR purposes. Digit. Signal Process. 2013, 23, 1923–1932. [Google Scholar] [CrossRef]
  30. Harimi, A.; Fakhr, H.S.; Bakhshi, A. Recognition of emotion using reconstructed phase space of speech. Malays. J. Comput. Sci. 2016, 29, 262–271. [Google Scholar] [CrossRef]
  31. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
  32. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  33. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  34. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  35. Zheng, J.; Lu, C.; Hao, C.; Chen, D.; Guo, D. Improving the generalization ability of deep neural networks for cross-domain visual recognition. IEEE Trans. Cogn. Dev. Syst. 2020, 13, 607–620. [Google Scholar] [CrossRef]
  36. Hao, C.; Chen, D. Software/Hardware Co-design for Multi-modal Multi-task Learning in Autonomous Systems. In Proceedings of the 2021 IEEE 3rd International Conference on Artificial Intelligence Circuits and Systems (AICAS), Washington, DC, USA, 6–9 June 2021; pp. 1–5. [Google Scholar] [CrossRef]
  37. Brochu, E.; Cora, V.M.; De Freitas, N. A tutorial on Bayesian optimization of expensive cost functions, with application to active user modeling and hierarchical reinforcement learning. arXiv 2010, arXiv:1012.2599. [Google Scholar]
  38. Chollet, F. Deep Learning with Python; Manning: New York, NY, USA, 2018; Volume 361. [Google Scholar]
  39. Liu, C.; Springer, D.; Li, Q.; Moody, B.; Juan, R.A.; Chorro, F.J.; Castells, F.; Roig, J.M.; Silva, I.; Johnson, A.E.; et al. An open access database for the evaluation of heart sound algorithms. Physiol. Meas. 2016, 37, 2181. [Google Scholar] [CrossRef] [PubMed]
  40. Milani, M.G.; Abas, P.E.; De Silva, L.C.; Nanayakkara, N.D. Abnormal heart sound classification using phonocardiography signals. Smart Health 2016, 21, 100194. [Google Scholar] [CrossRef]
  41. Zhong, W.; Liao, L.; Guo, X.; Wang, G. A deep learning approach for fetal QRS complex detection. Physiol. Meas. 2018, 39, 045004. [Google Scholar] [CrossRef] [PubMed]
  42. Lee, J.S.; Seo, M.; Kim, S.W.; Choi, M. Fetal QRS detection based on convolutional neural networks in noninvasive fetal electrocardiogram. In Proceedings of the 2018 4th International Conference on Frontiers of Signal Processing (ICFSP), Poitiers, France, 24–27 September 2018; pp. 75–78. [Google Scholar]
  43. Vo, K.; Le, T.; Rahmani, A.M.; Dutt, N.; Cao, H. An efficient and robust deep learning method with 1-D octave convolution to extract fetal electrocardiogram. Sensors 2020, 20, 3757. [Google Scholar] [CrossRef]
  44. Krupa, A.J.; Dhanalakshmi, S.; Lai, K.W.; Tan, Y.; Wu, X. An IoMT enabled deep learning framework for automatic detection of fetal QRS: A solution to remote prenatal care. J. King Saud Univ.-Comput. Inf. Sci. 2022, 34, 7200–7211. [Google Scholar]
Figure 1. Typical waveform of a phonocardiogram (PCG) signal and its components: S 1 ( T S 1 = 70–150 ms), S 2 ( T S 2 = 60–120 ms), S 3 ( T S 3 = 40–100 ms), and S 4 ( T S 4 = 40–80 ms).
Figure 1. Typical waveform of a phonocardiogram (PCG) signal and its components: S 1 ( T S 1 = 70–150 ms), S 2 ( T S 2 = 60–120 ms), S 3 ( T S 3 = 40–100 ms), and S 4 ( T S 4 = 40–80 ms).
Sensors 22 09569 g001
Figure 2. Flow diagram of the proposed approach for PCG signal classification: (1) building RPS from the original PCG signal; (2) projections of data represented in RPS are calculated on the XY, XZ, and YZ coordinate planes to form the red, green, and blue channels of the corresponding RGB image; (3) the chaogram is formed as a compatible input for a deep convolutional neural network; (4) a DCNN model pretrained on the large ImageNet Dataset is used for learning the extraction and classification of the chaogram’s high-level features; (5) the PCG signals are classified as normal or abnormal signals using the trained DCNN model.
Figure 2. Flow diagram of the proposed approach for PCG signal classification: (1) building RPS from the original PCG signal; (2) projections of data represented in RPS are calculated on the XY, XZ, and YZ coordinate planes to form the red, green, and blue channels of the corresponding RGB image; (3) the chaogram is formed as a compatible input for a deep convolutional neural network; (4) a DCNN model pretrained on the large ImageNet Dataset is used for learning the extraction and classification of the chaogram’s high-level features; (5) the PCG signals are classified as normal or abnormal signals using the trained DCNN model.
Sensors 22 09569 g002
Figure 3. Mutual information of all samples in the PhysioNet dataset (a) and their average mutual information (b).
Figure 3. Mutual information of all samples in the PhysioNet dataset (a) and their average mutual information (b).
Sensors 22 09569 g003
Figure 4. Six PCG signals and their representation in R P S with τ = 18 and d = 3 : (ac) are normal, and (df) are abnormal signals.
Figure 4. Six PCG signals and their representation in R P S with τ = 18 and d = 3 : (ac) are normal, and (df) are abnormal signals.
Sensors 22 09569 g004
Figure 5. Extracting the chaogram image from the reconstructed phase space of a PCG signal.
Figure 5. Extracting the chaogram image from the reconstructed phase space of a PCG signal.
Sensors 22 09569 g005
Figure 6. Chaograms of six PCG signals: on the top row, 3 normal cases, and, on the bottom row, 3 abnormal cases.
Figure 6. Chaograms of six PCG signals: on the top row, 3 normal cases, and, on the bottom row, 3 abnormal cases.
Sensors 22 09569 g006
Table 1. Results of the hyperparameter optimisation.
Table 1. Results of the hyperparameter optimisation.
HyperparameterAlexNetVGG16InceptionV3ResNet50
Optimizer functionAdamRMSpropAdamaxAdam
Learning Rate 10 × 10 5 10 × 10 3 10 × 10 5 10 × 10 4
Batch size986410286
Epoch260320280260
Table 2. Results regarding the classification of the PCG signals included in the PhysioNet dataset (best found values are in bold).
Table 2. Results regarding the classification of the PCG signals included in the PhysioNet dataset (best found values are in bold).
NetworkSensitivitySpecificityAccuracyScore
AlexNet82.5591.2189.6886.88
VGG1683.3691.4990.0587.43
InceptionV384.4991.6390.3688.06
ResNet5083.6891.3990.0287.54
Table 3. Comparison among the results obtained by the proposed method and those obtained by state-of-art methods.
Table 3. Comparison among the results obtained by the proposed method and those obtained by state-of-art methods.
AlgorithmAccuracyRecallPrecisionF1 Score
[41] 77.38---
[42] -86.6992.77-
[43] -90.3291.8281.5
[44] 91.3289.5294.3790.12
Proposed92.1990.9590.9583.41
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Harimi, A.; Majd, Y.; Gharahbagh, A.A.; Hajihashemi, V.; Esmaileyan, Z.; Machado, J.J.M.; Tavares, J.M.R.S. Classification of Heart Sounds Using Chaogram Transform and Deep Convolutional Neural Network Transfer Learning. Sensors 2022, 22, 9569. https://doi.org/10.3390/s22249569

AMA Style

Harimi A, Majd Y, Gharahbagh AA, Hajihashemi V, Esmaileyan Z, Machado JJM, Tavares JMRS. Classification of Heart Sounds Using Chaogram Transform and Deep Convolutional Neural Network Transfer Learning. Sensors. 2022; 22(24):9569. https://doi.org/10.3390/s22249569

Chicago/Turabian Style

Harimi, Ali, Yahya Majd, Abdorreza Alavi Gharahbagh, Vahid Hajihashemi, Zeynab Esmaileyan, José J. M. Machado, and João Manuel R. S. Tavares. 2022. "Classification of Heart Sounds Using Chaogram Transform and Deep Convolutional Neural Network Transfer Learning" Sensors 22, no. 24: 9569. https://doi.org/10.3390/s22249569

APA Style

Harimi, A., Majd, Y., Gharahbagh, A. A., Hajihashemi, V., Esmaileyan, Z., Machado, J. J. M., & Tavares, J. M. R. S. (2022). Classification of Heart Sounds Using Chaogram Transform and Deep Convolutional Neural Network Transfer Learning. Sensors, 22(24), 9569. https://doi.org/10.3390/s22249569

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop