Next Article in Journal
Deep Learning-Based Driver’s Hands on/off Prediction System Using In-Vehicle Data
Next Article in Special Issue
Camera- and Viewpoint-Agnostic Evaluation of Axial Postural Abnormalities in People with Parkinson’s Disease through Augmented Human Pose Estimation
Previous Article in Journal
Outlier Vehicle Trajectory Detection Using Deep Autoencoders in Santiago, Chile
Previous Article in Special Issue
Medical Image Classification Based on Semi-Supervised Generative Adversarial Network and Pseudo-Labelling
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Compressed Sensing Data with Performing Audio Signal Reconstruction for the Intelligent Classification of Chronic Respiratory Diseases

Department of Computing and Informatics, Bournemouth University, Bournemouth BH12 5BB, UK
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(3), 1439; https://doi.org/10.3390/s23031439
Submission received: 2 December 2022 / Revised: 23 January 2023 / Accepted: 25 January 2023 / Published: 28 January 2023

Abstract

:
Chronic obstructive pulmonary disease (COPD) concerns the serious decline of human lung functions. These have emerged as one of the most concerning health conditions over the last two decades, after cancer around the world. The early diagnosis of COPD, particularly of lung function degradation, together with monitoring the condition by physicians, and predicting the likelihood of exacerbation events in individual patients, remains an important challenge to overcome. The requirements for achieving scalable deployments of data-driven methods using artificial intelligence for meeting such a challenge in modern COPD healthcare have become of paramount and critical importance. In this study, we have established the experimental foundations for acquiring and indeed generating biomedical observation data, for good performance signal analysis and machine learning that will lead us to the intelligent diagnosis and monitoring of COPD conditions for individual patients. Further, we investigated on the multi-resolution analysis and compression of lung audio signals, while we performed their machine classification under two distinct experiments. These respectively refer to conditions involving (1) “Healthy” or “COPD” and (2) “Healthy”, “COPD”, or “Pneumonia” classes. Signal reconstruction with the extracted features for machine learning and testing was also performed for securing the integrity of the original audio recordings. These showed high levels of accuracy together with the performances of the selected machine learning-based classifiers using diverse metrics. Our study shows promising levels of accuracy in classifying Healthy and COPD and also Healthy, COPD, and Pneumonia conditions. Further work in this study will be imminently extended to new experiments using multi-modal sensing hardware and data fusion techniques for the development of the next generation diagnosis systems for COPD healthcare of the future.

1. Introduction

The World Health Organization (WHO) reported that chronic obstruction pulmonary disease (COPD) was the fifth leading cause of death in the world at the beginning of the century [1]. However, in 2018, ref. [2] reported that COPD was the third largest cause of mortality in the world, and now, ref. [3] expects COPD deaths to grow to the leading cause of death by 2030. COPD is a complex respiratory disease defined as a degenerative inflammatory condition that chronically limits airflow for many pulmonary disorders [4].
Patients with COPD have acute exacerbations that may lead to emergency hospitalization; however, they are more likely to be re-hospitalized after their initial discharge [5]. The cost of healthcare for COPD is substantial, and expectations are that the costs will grow even more as COPD prevalence increases [2]. In the U.K. alone, ref. [6] reported that the cost of COPD reached £1.9 billion a year to the National Health Service (NHS). Hence, the prevention, early detection, and management of COPD conditions is an essential strategy for health care services [7]. Therefore, there is a need for the advancement of new decision support systems, which enable clinicians in monitoring, intelligently detecting, and understanding COPD conditions, leading to early preventions of likely exacerbation events. These systems will also serve the rest of the clinical community into pursuing their specific care operations more efficiently, including the timely drug delivery for COPD patients in their homes. Currently, the amount of effort for monitoring patients with COPD, both in their homes and hospitals, requires large deployments of medical care staff, which has become unsustainable. It is, therefore, important to opt for other approaches to meet the care needs of COPD patients of the present time and future. With the advancement and affordability of wearable sensors, information, and communication technologies, over the last two decades, it has become possible to generate large observation and measure big data, which can be efficiently analyzed for critical health conditions and processes, understanding, and extraction of knowledge. For the case of chronic respiratory diseases, there is a potential of exploring sensors and measurements-based patients signal data nowadays for the real-time analysis and diagnosis of their health conditions and sub-conditions potentially. The confirmation of these conditions using machine learning and classification methods may lead us to understand the likelihood of critical exacerbation events with lung function failures, which may occur for COPD patients and others with similar respiratory conditions.
In particular, adventitious lung sounds may occur on top of healthy lung sounds due to damage or obstructions of the lungs and airways. When observed and measured, they are normally classified into two main categories: Continuous, around 250 ms and discontinuous, about 25 ms [8]. Continuous lung sounds, such as wheezing, are often heard in conditions, such as chronic obstruction pulmonary disease (COPD), and discontinuous sounds, such as crackles, are common in pneumonia [9]. Additionally, within the respiratory auscultations, the background noise of the heart, digestive system, and internal and external noise can be heard, causing a low signal-to-noise ratio. Plus, the sounds overlap in the time and frequency domains, while the breathing rhythm makes the signals non-stationary by nature, with changing statistics over time.
As a result, the reconstruction and classification of respiratory auscultation present challenges from non-stationary signals, transient signals of discontinuous crackle sounds, to noises that can all overlap in time and frequency space. Notwithstanding, auscultation recordings are generated from a single sensory point, while the sounds are from multiple locations of the three-dimensional lung organ. This indeed creates multiple challenges in separating these mixed sounds [10].
A common theme will be in using a low band or pass filter to separate heart sounds from lung sounds [9,10,11]. However, low pass filters can induce unwanted artifacts or aliasing [12], that is undesirable noise. Additionally, ref. [11] found that not cutting out the heart sounds had negligible effects on the results. Therefore, separating the heart sounds may not be an essential step here.
Researchers have used a range of transform methods for time-frequency analysis for adventitious lung sounds and lung conditions from short-time Fourier transforms (STFT) [11,12,13], empirical mode decomposition (EMD) [13,14] to wavelet transforms (W.T.) [10,11]. Noticing the STFT and EMD have challenges in extracting features of non-stationary, transient, and overlapping signals where [11] suggests that Fourier-based methods cannot detect transient signals. Plus, ref. [14] shows that EMD can detect crackles, however, cannot distinguish overlapping lung sound crackles; this is in line with [15], where EMD works more effectively on non-overlapping signals. In addition, according to [11], the technique of utilizing continuous wavelet transform and STFT was an improvement on STFT alone. Through multi-resolution analysis, W.T. can capture more delicate signal details [12]. In addition, multi-resolution analysis in structural health monitoring of mechanical equipment has shown the ability to identify impulse and transient signals within noise [16]. Therefore, the W.T. form of multi-resolution offers a range of capabilities that aids in extracting features from respiratory auscultation audio over STFT and EMD.
The STFT, EMD, and W.T. all have inverse transforms; however, there is little research on signal reconstruction of respiratory auscultations from important representative features. Although, ref. [13] utilized compressed sensing and signal reconstruction to transmit the respiratory auscultation audio from a sensor to a smartphone. An essential factor in rebuilding the signal can map the output feature back to the input and shows the features selected capture the most important information in the original audio signals. Therefore, signal reconstruction is an essential part of this research work before we utilize most of its dominant features for respiratory diseases classification using machine learning. This paper is, therefore, purposely set out and presented in the following way: The data used in the study, the data cleaning process, the data transformation and feature reduction methods, and the reconstruction results. We then proceed with a review and implementation of classification methods and major results, leading to summarizing our findings, a discussion, and a conclusion with recommended future work.

2. Materials and Methods

The data utilized in this study was the ICBHI Respiratory challenge database [17]. The dataset contains 920 audio recordings of 126 patients. The audio samples vary in the number of channels (Mono and Stereo), sampling rate (4000–44,100 Hz), and duration (30–90 s). There is accompanying information on patient diagnosis and demographics for each patient. For this study, we used the Healthy, COPD, and Pneumonia of diagnosis classes of auscultation. Table 1 shows the classes used, breaking down demographics per class.
As modelling requires the data samples to be of the same length and the audio samples varied in duration, a random seven-second section was selected, which could capture a breathing cycle, where a breathing cycle ranges from 12–18 revolutions per minute [18]. Because of the imbalances between classes, the Healthy and Pneumonia classes of audio sections had two data augmentation options, out of five, applied to ensure each sample was different from each other. The augmentation options are time-stretching [19,20], where audio is sped up or down; pitch-shifting [20,21], where the audio frequency is moved up or down; added noise [19], where extra noise is added; time-shifting [20], where time is rolled forward or backward; and no augmentation. Two out of five options gave permutations of up to 20 different options, allowing for each sample to be augmented differently. The process increased the Healthy class from 35 to 735 audio samples and the Pneumonia class from 39 to 740 audio samples.

2.1. Audio Cleaning and Normalization

The pre-processing cleaning stage reduces noise and places all samples into a normalized format. The process contains the following steps:
  • Thresholding;
  • Signal smoothing;
  • Detrending;
  • Audio loudness normalization;
  • Normalization.
When loading the audio, the audio samples are down sampled to 4000 Hz, bringing all samples into the same sample rate. Outliers in the audio amplitude, expected by stethoscope contact movement, were reduced by thresholding. By thresholding the signal amplitude above four standard deviations and reduced to the mean, crackles can appear within four standard deviations. With down-sampling and removing outliers, cleaning the audio with a smoothing filter will also remove some noise. The choice of filter is the Savol filter, a moving filter with a polynomial function that is well suited for noise reduction for lung sounds [22]. The audio samples are non-stationary and can display trending; therefore, detrending reduces the non-stationary [23] (p. 47). The works of [24] highlight that respiratory audio has two components: air turbulence and lung structural sounds, which compete with each other when listened to from different locations. Therefore, the EUB R128 normalization is used. Finally, the values are normalized to bring them into the same range.

2.2. Wavelet Transform

Wavelet transform (W.T.) is used for multi-resolution audio signals, breaking them down into different levels of frequency ranges, where the formula is shown in Equation (1). The mother wavelet (Ψ*) chosen is the Morlet wavelet because the distribution characteristics are similar to the transient crackle with a sudden peak.
w n ( s ) = n = 0 N 1 x n Ψ * [ ( n n ) δ t s ]  
The complex Morlet wavelet returns the real and imaginary components that this study will analyze. This analysis supports our objectives as W.T. is robust to noise, localizes audio characteristics [12], and has inverse transform [25]. The inverse transform allows for the reconstruction of the signal from the multi-resolution analysis to audio signals

2.3. Compressed Sensing

Compressed sensing underlines the sparse encoder dictionary learning. The main principles of compressed sensing are:
  • Incoherence;
  • Sparsity.
Incoherence is a property in that the samples are not connected by time or spatial domains, which expanses the time-frequency localization problem or uncertainty problem. In that, the samples are more spread out and sparse within the domain [26]. Whereas in compressed sensing matrices, the values in the rows do not correlate with those in the columns [27] (p. 90). Sparsity is a property where samples are spread out, where the low values nearing zero can be zeroed out altogether. This allows the data to have minimal or low non-zero elements. The sparsity constraint placed on compressed sensing enables the change from an over-complete solution to be relaxed and a unique solution to be found [28]. When compressed sensing comes to matrix forms, the matrix structure, which maps linearly when restricted to sparsity [29] naturally preserving the so-called restricted isometric property (RIP) [27] (pp. 90–96). The ability to sub-sample from subspace aids feature reduction, which is with less than the Nyquist sample rate, which allows for meeting the objective of signal reconstruction.

2.4. Dictionary Learning

Dictionary leaning incorporates compressed sensing with the factors of sparsity by relaxing the linear constraints and utilizing an error-bound element and incoherence factor between each atom (column) in the dictionary [28]. Additionally, dictionary learning uses algorithms, such as gradient descent or orthogonal matching pursuit (OMP) to aid in finding a sparse representation and reconstruction process by selecting highly correlated samples for the dictionary atom [30]. Dictionary learning is calculated by Equation (2).
[ u , v ] = a r g m i n 1 2 | | X U V | | L 2 + α × | | U | | L 1 ( U , V ) with   | | V _ k | | 2 < = 1   f o r   a l l   0 < = k < n _ c o m p o n e n t s  
Dictionary learning supports the decomposition of the multi-resolution analysis matrix into a reduced number of components; the multiplication of the components and the transform results in reconstructing an approximation of the multi-resolution analysis matrix.

2.5. Singular Value Decomposition

Singular value decomposition (SVD) is a method that factorizes real, or complex, matrices into three matrices. It is often used in signals processing in order to compress signal data to their most representative matrix form of features and make it more efficient to work with complex signals. Specifically, the method exposes many of the important and interesting representational features of signals from the original matrix. For an illustration and the special case of real matrices, SVD is performed as follows:
A = U V T  
where A is a (n × p) matrix to decompose [31]. U is a (n × n) orthogonal matrix, whose columns are known as the left-singular vectors; has the same dimensions (n × p) as A and has the so-called singular values in its diagonal. V T is an orthogonal (p × p) matrix, which is the transpose matrix of V , whose rows are known as the right singular vectors. Further, SVD computations involve the extraction of the eigenvalues and eigenvectors of AAT and ATA. Their eigenvectors make up the columns of V and U , respectively. The singular values are the diagonal elements of the matrix. They are usually arranged in descending orders. Additionally, they are the square roots of the eigenvalues of AAT or ATA, ref. [32]. In addition, we note that SVD supports signals of noise reduction, in this case, through matrix characteristics decomposition, which leads to the most interesting number of features representing the signal, while assuring the ability to recover the original matrix through SVD matrices operations.

2.6. Signal Reconstruction Metrics

In order to understand the accuracy of the signal reconstruction, comparing the pre-processed signal with the reconstructed signal will highlight the accuracy. Therefore, the mean square error (MSE) and the correlation coefficients can be used as metrics for signal similarity analyses.
The mean square error is a measure of the difference calculated by Equation (4) [30], where A is the original signal and B is the reconstructed signal. The MSE shows the average difference in the distance between two signals.
M S E = n ( A [ n ] B [ n ] ) 2 m
Another measure of signal similarity is the correlation coefficient between the two signals A and B [33], as calculated by Equation (5).
C o r r   C o e f =   ( A i A ¯   )   ( B i B ¯ )   ( A i A ¯ ) 2     ( B i B ¯ ) 2  
where A ¯ is the original signal mean, and B ¯ is the recovered signal mean. The correlation coefficient shows the linear dependence between the signals.

2.6.1. Summary of Extracted Features

The framework extracted features is where U contained 153 features, VT contained 90 features, and S contained 9 features. The number of features was the same for the real and imaginary components of the signals.

2.6.2. Signal Reconstruction Results

The results of signal reconstruction are shown in Table 2 below.

2.6.3. Summary of Signal Reconstruction

The results of the MSE show the reconstruction accuracy averages at 3 × 10−3, with the best result reaching 5 × 10−4, meaning that the distance between the pre-processed and reconstructed signals is very small. Likewise, the correlation coefficients have a mean score of 0.57, while the highest score reaches 0.92. Reconstruction results demonstrate that the reconstruction is an excellent approximation of the pre-processed original audio signal.

2.7. Classification

The study covered two different classifications, one of “Healthy” and “COPD” and the second of “Healthy”,” COPD”, or “Pneumonia”. Pneumonia was chosen as the adventitious sounds are mainly crackles, whereas COPD is mainly wheezing, which allows for discrimination between the two classes. As the complex Morlet wavelet gives the real and the imaginary components of the signal, each component is classified. The models for classification are: The Gaussian mixture model (GMM), decision tree classifier (DTC), support vector machine (SVM), and random forest classifier (RFC).
The GMM is a classification algorithm, which allows for overlapping borders of Gaussian distribution clusters that may support the overlapping frequencies of lung sounds [34]. DTC uses a divide-and-conquer strategy for classification that offers transparency and, therefore, allows for an objective analysis [35]. The SVM utilizes a boundary separation, or if data are highly dimensional, a separation of categories with a hyper-plane, which can be linear, polynomial, quadratic, or of higher orders [35]. The RFC is an ensemble approach, a powerful tool for data mining in which the combining of multiple trees for the outcome can be viewed as a bias-variance decomposition. Specifically, it aids the performance [35], which is supported by the random bagging of sampling with replacement from the training data and bootstrap of the features [36]. Additionally, random forests can give information on feature importance; therefore, it is an excellent option for classification. Grid search, which cycles through different parameters for the models to find the optimal parameters, is used to increase the model’s performance. The grid search parameters for the RFC number of estimators range from one hundred to six hundred with increments of fifty, and the depth range from ten to one hundred with increments of ten.

Classification Metrics

The performance of the models is evaluated by looking at the true positives (T.P.), true negatives (T.N.), false positives (F.P.), and false negatives (F.N.) [10]. We utilized the accuracy, F1 scores, receiver operator characteristic (ROC) curves, and area under curve (AUC) scores [36]. For the Healthy, COPD, and Pneumonia classifications, the ROC curves will be the one-versus-all classification, which compares one class to the other two classes. five-fold cross-validation is utilized, while the results are the averages across the five-fold and the cross-validation standard deviation to ensure that the model performance is robustly assessed. The level of coverage of the model’s performance is reported with confidence intervals of 95% [36].

3. Results

3.1. Healthy and COPD Classification Results

The results are set out with baseline results, model parameter optimization results, the ROC, and the area under the curve plots. The baseline results for the classification of healthy and COPD is shown in Table 3.
Taking the SVD and random forest further with parameter tuning, the results are shown in Table 4. Cross-validation scores and confidence intervals are reported.
ROC curves are used to display the discriminative ability of the classification models. The comparison of the different models are shown in Figure 1, and the comparison of the real and imaginary components using Random forest classifier ROC curve results are shown in Figure 2.
The ROC curve results for the classification of Healthy, COPD, and pneumonia are shown in Figure 3 and Figure 4 below.

3.2. Healthy, COPD, and Pneumonia Classification Results

The baseline results for the classification of healthy, COPD, and pneumonia are shown in Table 5.
The random forest and SVC classifiers were the best performing and taken forward for parameter tuning; the results are shown in Table 6.

3.3. Summary of Classification Findings

The random forest models produced the best performing models for the classification of Healthy versus COPD and Healthy versus COPD versus Pneumonia. The best features for the Healthy versus COPD classification were the SVD’s U and VT for the imaginary component of the auscultation’s audio, both having accuracies of 80% and the area under ROC curves showed that the SVD U elements were better at discriminating between healthy and COPD than the SVD VT elements with values of 0.87 and 0.77, respectively, with the random forest model. Similarly, for the classification of Healthy versus COPD versus Pneumonia, the best results were from the random forest classifier, highlighted in Figure 3. However, the best features were on the SVD’s S (Singular) values of both the real and imaginary components of the auscultation recordings, while achieving 70% and 68% accuracy, respectively. The random forest model’s ability to discriminate between classes on the SVD S elements was relatively close values, with the real components ranging between 0.82 to 0.86 (see Figure 3c) and the imaginary components ranging from 0.80 to 0.83 (see Figure 3f).

4. Discussion

There are some encouraging results in the classification of Healthy and COPD; the imaginary components of the signal and the orthogonal SVD elements are the best performers, which may relate to the harmonic resonance of wheezes often identified in COPD patients. The classification of the Healthy versus COPD achieved a good accuracy of 80%, with 95% confidence levels of 76–79% on the audio signals imaginary components on the SVD’s U and V.T. elements. For the Healthy versus COPD versus Pneumonia, an acceptable level of accuracy of 70% with a 95% confidence level of 66–70% on the audio signals real components on the SVD’s S (singular values), with good levels of discrimination between conditions. For the signal reconstruction, the best scores are MSE of 5.2 × 10−3 with a mean score of 3.0 × 10−2 and a correlation coefficient score of 0.92 with a mean score of 0.57. Indeed, this suggests a good level of signal recovery. When comparing the results in the Healthy versus COPD versus Pneumonia, we find that the best performance was from the real component of the signal with the SVD’s element, which relates to the signal’s strength, especially between the COPD and Pneumonia that had higher classification numbers in the confusion matrix.
In comparison, ref. [11], who also utilized W.T., achieved scores of 39.97–49.86% in classifying normal lung sounds, wheezes, and crackles on the ICBHI 2017 challenge database. Ref. [11]’s choice of adventitious sounds can be related to Healthy, COPD, and Pneumonia, respectively, in which this study demonstrated higher accuracies of classification. In addition, ref. [37] discusses the challenge of achieving above 50% accuracy in the ICBHI 2017 challenge database, where they aimed to classify normal, wheezes, crackles, and both wheezes and crackles. Ref. [37] suggested that there may be issues with the dataset as they found an audio of a patient diagnosed with respiratory disease, but the annotated notes for the specific audio recording had no adventitious sound noted. However, no adventitious sounds do not mean a lack of disease, as [38] noted. Ref. [39] utilized discrete wavelet transforms and deep learning for classifying the ICBHI 2017 challenge database into healthy and unhealthy, which achieved an F1 score of 81.64%, similar to the F1 scores of the best models of Healthy versus COPD of 83%. However, this study’s approach was more focused on COPD, whereas [39] unhealthy had a broader range of diseases. Ref. [40] achieved high accuracy of 92.30% by utilizing a 17-layered 2D-convolutional neural network (CNN) with features of MFCC and spectrograms to classify the ICHBI dataset auscultation recordings into their associated diseases.
The advantage of our proposed approach is the ability to achieve signal reconstruction and recovery to approximate the original signal with high credibility. Furthermore, the recovery of our signals to their high level of accuracy, together with the good levels of their correct classification rates on the health conditions using machine learning highlights, is a way forward for understanding human respiratory conditions. Our method is specifically feasible for respiratory auscultation classifications and supports the hypotheses on health conditions.
In addition, while other work has focused on statistical and neural network-based approaches, our results demonstrate a new method of utilizing compressed sensing for auscultation classifications. Nevertheless, further optimization of the extraction process needs to be deployed together with large volumes of experimental datasets to increase the accuracy of both signal recovery and machine classifications. In future work, experimenting with multi-modal data and dictionary learning for improving the diagnostic and prognosis of COPD conditions should be the focus.

5. Conclusions

The developed benchmark work in this study not only provides good levels of accuracy for signal reconstruction, but it also brings good performing machine classification of respiratory lung sounds. These are brought in good context of their associated chronic health conditions. Specifically, on the machine classification side, the random forest classifier is the performing algorithm with accuracies ranging from around 80% for classifying cases of “Healthy” and “COPD”. It reaches accuracies of approximately 70% for classifying cases, including “Healthy”, “COPD”, and “Pneumonia”. These were all obtained with confidence intervals showing the stability of the models. The ROC curves show the discrimination ability of the classifiers, although with limitations. Our work has also the potential of applications in other respiratory disease classifications and beyond. However, more work needs to be performed, since we need to improve the performance of our classifiers to higher levels first while validating them under much larger and diverse datasets. Our future work will specifically involve research investigations on obstructive pulmonary chronic respiratory diseases using larger datasets in order to scale our approaches in terms of their accuracies and performances. We will also aim to identify lung sounds that correspond to various sub-conditions of COPD, particularly those which may highly lead to patients’ exacerbation events. We will aim in the near future to automatically predict the likelihood of occurrence of such serious events, ahead of time and with good contexts, in order to accelerate medical responses to patients under critical respiratory conditions.

Author Contributions

Formal research guided analyses and investigation, T.A.; writing—original draft preparation, T.A.; writing—review and editing, B.A.-Z. and Z.S.; supervision, Z.S. and B.A.-Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research has currently no external funding.

Institutional Review Board Statement

Approved by Bournemouth University Ethics Committee. Reference number 40455.

Informed Consent Statement

Not applicable.

Data Availability Statement

The open access dataset is available from https://bhichallenge.med.auth.gr/ICBHI_2017_Challenge.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pauwels, R.A.; Rabe, K.F. Burden and Clinical Features of Chronic Obstructive Pulmonary Disease (COPD). Lancet 2004, 364, 613–620. [Google Scholar] [CrossRef] [PubMed]
  2. Viniol, C.; Vogelmeier, C.F. Exacerbations of COPD. Eur. Respir. Rev. 2018, 27, 170103. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. WHO. Chronic Obstructive Pulmonary Disease (COPD). 2021. Available online: https://www.who.int/news-room/fact-sheets/detail/chronic-obstructive-pulmonary-disease-(copd) (accessed on 10 October 2021).
  4. Rabe, K.F.; Hurst, J.R.; Suissa, S. Cardiovascular Disease and COPD: Dangerous Liaisons? Eur. Respir. Rev. 2018, 27, 180057. [Google Scholar] [CrossRef]
  5. Min, X.; Yu, B.; Wang, F. Predictive Modeling of the Hospital Readmission Risk from Patients’ Claims Data Using Machine Learning: A Case Study on COPD. Sci. Rep. 2019, 9, 2362. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. British Lung Foundation. The battle for breath—The economic burden of lung disease—British Lung Foundation. British Lung Foundation. 2021. Available online: https://www.blf.org.uk/policy/economic-burden (accessed on 18 November 2021).
  7. Perna, D.; Tagarelli, A. Deep Auscultation: Predicting Respiratory Anomalies and Diseases via Recurrent Neural Networks. In Proceedings of the 2019 IEEE 32nd International Symposium on Computer-Based Medical Systems (CBMS), Cordoba, Spain, 5–7 June 2019; pp. 50–55. [Google Scholar] [CrossRef] [Green Version]
  8. Sarkar, M.; Madabhavi, I.; Niranjan, N.; Dogra, M. Auscultation of the Respiratory System. Ann. Thorac. Med. 2015, 10, 158–168. [Google Scholar] [CrossRef]
  9. Grønnesby, M.; Solis, J.C.A.; Holsbø, E.; Melbye, H.; Bongo, L.A. Feature Extraction for Machine Learning Based Crackle De-tection in Lung Sounds from a Health Survey. arXiv 2017, arXiv:1706.00005. [Google Scholar]
  10. Khan, S.I.; Pachori, R.B. Automated Classification of Lung Sound Signals Based on Empirical Mode Decomposition. Expert Syst. Appl. 2021, 184, 115456. [Google Scholar] [CrossRef]
  11. Serbes, G.; Ulukaya, S.; Kahya, Y.P. Precision Medicine Powered by pHealth and Connected Health. IFMBE Proc. 2017, 66, 45–49. [Google Scholar] [CrossRef]
  12. Kandaswamy, A.; Kumar, C.S.; Ramanathan, R.P.; Jayaraman, S.; Malmurugan, N. Neural Classification of Lung Sounds Using Wavelet Coefficients. Comput. Biol. Med. 2004, 34, 523–537. [Google Scholar] [CrossRef]
  13. Oletic, D.; Bilas, V. Asthmatic Wheeze Detection from Compressively Sensed Respiratory Sound Spectra. IEEE J. Biomed. Health 2018, 22, 1406–1414. [Google Scholar] [CrossRef]
  14. Charleston-Villalobos, S.; González-Camarena, R.; Chi-Lem, G.; Aljama-Corrales, T. Crackle Sounds Analysis by Empirical Mode Decomposition. IEEE Eng. Med. Biol. 2007, 26, 40–47. [Google Scholar] [CrossRef]
  15. Stankovi, L.; Mandi, D.; Dakovi, M.; Brajovi, M. Time-Frequency Decomposition of Multivariate Multicomponent Signals. Signal Process. 2018, 142, 468–479. [Google Scholar] [CrossRef]
  16. Chen, X.; Du, Z.; Li, J.; Li, X.; Zhang, H. Compressed Sensing Based on Dictionary Learning for Extracting Impulse Components. Signal Process. 2014, 96, 94–109. [Google Scholar] [CrossRef]
  17. Rocha, B.M.; Filos, D.; Mendes, L.; Vogiatzis, I.; Perantoni, E.; Kaimakamis, E.; Natsiavas, P.; Oliveira, A.; Jácome, C.; Marques, A.; et al. Precision Medicine Powered by pHealth and Connected Health. In Proceedings of the ICBHI 2017, Thessaloniki, Greece, 18–21 November 2017; pp. 33–37. [Google Scholar] [CrossRef]
  18. Tariq, Z.; Shah, S.K.; Lee, Y. Lung Disease Classification using Deep Convolutional Neural Network. In Proceedings of the IEEE International Conference on Bioinformatics and Biomedicine (BIBM), San Diego, CA, USA, 18–21 November 2019; Available online: https://ieeexplore.ieee.org/document/8983071 (accessed on 29 September 2022).
  19. Ko, T.; Peddinti, V.; Povey, D.; Khudanpur, S. Audio Augmentation for Speech Recognition. Interspeech 2015, 2015, 3586–3589. [Google Scholar] [CrossRef]
  20. Salamon, J.; Bello, J.P. Deep Convolutional Neural Networks and Data Augmentation for Environmental Sound Classification. IEEE Signal. Proc. Lett. 2016, 24, 279–283. [Google Scholar] [CrossRef]
  21. Tariq, Z.; Shah, S.K.; Lee, Y. Multimodal Lung Disease Classification Using Deep Convolutional Neural Network. In Proceedings of the 2020 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Seoul, Republic of Korea, 1 September 2020; pp. 2530–2537. [Google Scholar] [CrossRef]
  22. Haider, N.S.; Periyasamy, R.; Joshi, D.; Singh, B.K. Savitzky-Golay Filter for Denoising Lung Sound. Braz. Arch. Biol. Technol. 2018, 61, e18180203. [Google Scholar] [CrossRef]
  23. Cohen, M.X. Fundamentals of Time-Frequency Analyses in Matlab/Octave; Kindle Edition; Amazon.co.uk: London, UK, 2014; p. 47. [Google Scholar]
  24. Shiryaev, A.D.; Korenbaum, V.I. Frequency Characteristics of Air-Structural and Structural Sound Transmission in Human Lungs. Acoust Phys+ 2013, 59, 709–716. [Google Scholar] [CrossRef]
  25. Torrence, C.; Compo, G.P. A Practical Guide to Wavelet Analysis. Bull. Am. Meteorol. Soc. 1998, 79, 61–78. [Google Scholar] [CrossRef]
  26. Candes, E.J.; Wakin, M.B. An Introduction to Compressive Sampling. IEEE Signal. Proc. Mag. 2008, 25, 21–30. [Google Scholar] [CrossRef]
  27. Brunton, S.; Kutz, N. Data-Driven Science and Engineering Machine Learning, Dynamic Systems, And Control Systems, 1st ed.; Cambridge University Press: Cambridge, UK, 2019; p. 90. [Google Scholar]
  28. Tosic, I.; Frossard, P. Dictionary Learning. IEEE Signal. Proc. Mag. 2011, 28, 27–38. [Google Scholar] [CrossRef]
  29. Junge, M.; Lee, K. Generalized Notions of Sparsity and Restricted Isometry Property. Part I: A Unified Framework. Inf. Inference J. IMA 2019, 9, 157–193. [Google Scholar] [CrossRef]
  30. Gangannawar, S.A.; Siddmal, S.V. Compressed Sensing Reconstruction of an Audio Signal Using OMP—ProQuest. Int. J. Adv. Comput. Res. 2015, 5, 75–79. [Google Scholar]
  31. Zheng, Y.; Guo, X.; Jiang, H.; Zhou, B. An Innovative Multi-Level Singular Value Decomposition and Compressed Sensing Based Framework for Noise Removal from Heart Sounds. Biomed. Signal. Process. 2017, 38, 34–43. [Google Scholar] [CrossRef]
  32. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, UK, 2016. [Google Scholar]
  33. Sun, Z.; Wang, G.; Su, X.; Liang, X.; Liu, L. Similarity and Delay between Two Non-Narrow-Band Time Signals. arXiv 2020, arXiv:2005.02579. [Google Scholar]
  34. Scikit-Learn Developers. 2.1. Gaussian Mixture Models; [Online] Scikit-Learn. 2021. Available online: https://scikit-learn.org/stable/modules/mixture.html (accessed on 10 October 2021).
  35. Written, I.; Frank, E.; Hall, M.; Pal, C. Data Mining, Practical Machine Learning Tools and Techniques, 4th ed.; Kindle Edition; Elsevier: Montreal, QC, Canada, 2017. [Google Scholar]
  36. Bruce, P.; Bruce, A.; Gedeck, P. Practical Statistics for Data Scientists, 2nd ed.; Kindle Edition; O’Reilly: Farnham, UK, 2020. [Google Scholar]
  37. Chambres, G.; Hanna, P.; Desainte-Catherine, M. Automatic Detection of Patient with Respiratory Diseases Using Lung Sound Analysis. In Proceedings of the 2018 International Conference on Content-Based Multimedia Indexing (CBMI), La Rochelle, France, 4–6 September 2018; pp. 1–6. [Google Scholar] [CrossRef]
  38. Bohadana, A.; Izbicki, G.; Kraman, S.S. Fundamentals of Lung Auscultation. N. Engl. J. Med. 2014, 370, 2052–2053. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  39. Tiwari, U.; Bhosale, S.; Chakraborty, R.; Kopparapu, S.K. Deeplung Auscultation Using Acoustic Biomarkers for Abnormal Respiratory Sound Event Detection. In Proceedings of the ICASSP 2021—2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 1305–1309. [Google Scholar] [CrossRef]
  40. Hazra, R.; Majhi, S. Detecting Respiratory Diseases from Recorded Lung Sounds by 2D CNN. In Proceedings of the 2020 5th International Conference on Computing, Communication and Security (ICCCS), Patna, India, 14–16 October 2020; pp. 1–6. [Google Scholar] [CrossRef]
Figure 1. ROC curves of classification models of each SVD element and real and imaginary components: (a) ROC curves of real components of SVD U element; (b) ROC curves of real component of SVD VT element; (c) ROC curves of real component of SVD S element; (d) ROC curves of imaginary component of SVD U element; (e) ROC curves of imaginary component of SVD VT element; (f) ROC curves of Imaginary component of SVD S element.
Figure 1. ROC curves of classification models of each SVD element and real and imaginary components: (a) ROC curves of real components of SVD U element; (b) ROC curves of real component of SVD VT element; (c) ROC curves of real component of SVD S element; (d) ROC curves of imaginary component of SVD U element; (e) ROC curves of imaginary component of SVD VT element; (f) ROC curves of Imaginary component of SVD S element.
Sensors 23 01439 g001
Figure 2. ROC curve of classification of healthy and COPD classifications of each signal component and SVD elements: (a) Chart of the ROC curve of the RFC for real components on the classification of Healthy and COPD in the first panel; (b) chart of the ROC curve of the RFC for imaginary components on the classification of Healthy and COPD.
Figure 2. ROC curve of classification of healthy and COPD classifications of each signal component and SVD elements: (a) Chart of the ROC curve of the RFC for real components on the classification of Healthy and COPD in the first panel; (b) chart of the ROC curve of the RFC for imaginary components on the classification of Healthy and COPD.
Sensors 23 01439 g002
Figure 3. DTC and RFC classification of Healthy, COPD, and Pneumonia for each SVD element and real and imaginary components of the signals: (a) ROC curves of real components of SVD U element; (b) ROC curves of real component of SVD VT element; (c) ROC curves of real component of SVD S element; (d) ROC curves of imaginary component of SVD U element; (e) ROC curves of imaginary component of SVD VT element; (f) ROC curves of imaginary component of SVD S element.
Figure 3. DTC and RFC classification of Healthy, COPD, and Pneumonia for each SVD element and real and imaginary components of the signals: (a) ROC curves of real components of SVD U element; (b) ROC curves of real component of SVD VT element; (c) ROC curves of real component of SVD S element; (d) ROC curves of imaginary component of SVD U element; (e) ROC curves of imaginary component of SVD VT element; (f) ROC curves of imaginary component of SVD S element.
Sensors 23 01439 g003
Figure 4. RFC classification of Healthy, COPD, and Pneumonia one-vs-rest ROC curves for each SVD element and real and imaginary components of the signals: (a) ROC curves of real components of SVD U element; (b) ROC curves of real component of SVD VT element; (c) ROC curves of real component of SVD S element; (d) ROC curves of imaginary component of SVD U element; (e) ROC curves of imaginary component of SVD VT element; (f) ROC curves of imaginary component of SVD S element.
Figure 4. RFC classification of Healthy, COPD, and Pneumonia one-vs-rest ROC curves for each SVD element and real and imaginary components of the signals: (a) ROC curves of real components of SVD U element; (b) ROC curves of real component of SVD VT element; (c) ROC curves of real component of SVD S element; (d) ROC curves of imaginary component of SVD U element; (e) ROC curves of imaginary component of SVD VT element; (f) ROC curves of imaginary component of SVD S element.
Sensors 23 01439 g004
Table 1. ICBHI 2017 challenge database selected class breakdown.
Table 1. ICBHI 2017 challenge database selected class breakdown.
ConditionsNumber of
Recordings
Biological Sex
(Count)
Age Range
(Years)
MaleFemaleMinMax
COPD7935122664593
Healthy3515200.2516
Pneumonia37307481
Table 2. Result summary of signal reconstruction from extracted features.
Table 2. Result summary of signal reconstruction from extracted features.
StatsMSECorrelation Coefficient
count22682268
mean0.0306680.576079
std0.0121370.150377
min0.0051880.014053
0.250.0225980.488954
0.50.0291510.582712
0.750.0362620.682031
max0.1427990.924803
Table 3. Healthy vs COPD classification baseline results. All the baseline results have been achieved with the following parameter settings: Random forest (RFC): d = 500, e = 280 (in these, d stands for depth, and e stands for the number of estimators); GMM: components = 2, covariance = full; SVC: gamma = auto, C = 3000.
Table 3. Healthy vs COPD classification baseline results. All the baseline results have been achieved with the following parameter settings: Random forest (RFC): d = 500, e = 280 (in these, d stands for depth, and e stands for the number of estimators); GMM: components = 2, covariance = full; SVC: gamma = auto, C = 3000.
Classification DetailsClassification ModelF1-ScoreAccuracy
SVD U, RealRFC, d = 500, e = 28078.580
GMM, components = 233.544
DTC69.570
SVC, C = 300068.569
SVD Vt, RealRFC, d = 500, e = 2807172
GMM, components = 23547
DTC5959
SVC, C = 300053.554
SVD S, RealRFC, d = 500, e = 2807171
GMM, components = 235.538
DTC6060
SVC, C = 30003554
SVD U, ImagRFC, d = 500, e = 28078.579
GMM, components = 23753
DTC6969
SVC, C = 30007070
SVD Vt, ImagRFC, d = 500, e = 2807172
GMM, components = 247.548
DTC5959
SVC53.554
SVD S ImagRFC, d = 500, e = 2807172
GMM, components = 23554
DTC6364
SVC, C = 30003554
Table 4. Healthy vs. COPD classification of parameter tuning results. In these, d stands for depth, and e stands for the number of estimators.
Table 4. Healthy vs. COPD classification of parameter tuning results. In these, d stands for depth, and e stands for the number of estimators.
Classification DetailsClassification ModelMacro F1-ScoreAccuracyCV ScoreCV StdCI 95%
SVD U, RealRFC, d = 25, e = 39078.57976573–78
SVC, C = 2265.868.569
SVD Vt, RealRFC, d = 20, e = 40072.57368565–70
SVC, C = 17,911.653.554
SVD S, RealRFC, d = 25, e = 390727273670–75
SVC, C = 1251.93554
SVD U, ImagRFC, d = 30, e = 39079.58076574–79
SVC, C = 80,190.17070
SVD Vt, ImagRFC, d = 20, e = 40079.58076574–79
SVC, C = 58,523.67070
SVD S ImagRFC, d = 30, e = 400717273570–74
SVC, C = 2764.83554
Table 5. Healthy vs COPD vs Pneumonia baseline classification results. All the baseline results have been achieved with the following parameter settings: Random forest (RFC): d = 500, e = 280 (in these, d stands for depth, and e stands for the number of estimators); GMM: components = 2, covariance = full; SVC: gamma = auto, C = 3000.
Table 5. Healthy vs COPD vs Pneumonia baseline classification results. All the baseline results have been achieved with the following parameter settings: Random forest (RFC): d = 500, e = 280 (in these, d stands for depth, and e stands for the number of estimators); GMM: components = 2, covariance = full; SVC: gamma = auto, C = 3000.
DetailsClassification ModelMacro F1-ScoreAccuracy
SVD U, RealRFC, d = 500, e = 28059.751
GMM, components = 230.337
DTC50.760
SVC, C = 30004546
SVD Vt, RealRFC, d = 500, e = 28059.360
GMM, components = 23132
DTC4646
SVC, C = 300044.745
SVD S, RealRFC, d = 500, e = 28069.770
GMM, components = 22240
DTC55.356
SVC, C = 30001939
SVD U, ImagRFC, d = 500, e = 28060.361
GMM, components = 24850
DTC5252
SVC, C = 300046.347
SVD Vt, ImagRFC, d = 500, e = 28062.362
GMM, components = 232.732
DTC4950
SVC, C = 300044.345
SVD S ImagRFC, d = 500, e = 28067.367
GMM, components = 220.739
DTC58.759
SVC, C = 30001939
Table 6. Healthy vs COPD vs Pneumonia classification of parameter tuning results. In these, d stands for depth, and e stands for the number of estimators.
Table 6. Healthy vs COPD vs Pneumonia classification of parameter tuning results. In these, d stands for depth, and e stands for the number of estimators.
Classification DetailsClassification ModelMacro F1-ScoreAccuracyCV Score CV StdCI 95%
SVD U, RealRFC, d = 20, e = 30058.75958356–59
SVC, C = 1143.943.745
SVD Vt, RealRFC, d = 40, e = 50060.36159457–61
SVC, C = 1839.846.747
SVD S, RealRFC, d = 20, e = 4007070684.566–70
SVC, C = 1536.921
SVD U, ImagRFC, d = 30, e = 50060.361584.956–61
SVC, C = 1536.94647
SVD Vt, ImagRFC, d = 20, e = 40062.363593.857–61
SVC, C = 1536.94950
SVD S ImagRFC, d = 20, e = 30067.768684.265–70
SVC, C = 1536.919.739
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Albiges, T.; Sabeur, Z.; Arbab-Zavar, B. Compressed Sensing Data with Performing Audio Signal Reconstruction for the Intelligent Classification of Chronic Respiratory Diseases. Sensors 2023, 23, 1439. https://doi.org/10.3390/s23031439

AMA Style

Albiges T, Sabeur Z, Arbab-Zavar B. Compressed Sensing Data with Performing Audio Signal Reconstruction for the Intelligent Classification of Chronic Respiratory Diseases. Sensors. 2023; 23(3):1439. https://doi.org/10.3390/s23031439

Chicago/Turabian Style

Albiges, Timothy, Zoheir Sabeur, and Banafshe Arbab-Zavar. 2023. "Compressed Sensing Data with Performing Audio Signal Reconstruction for the Intelligent Classification of Chronic Respiratory Diseases" Sensors 23, no. 3: 1439. https://doi.org/10.3390/s23031439

APA Style

Albiges, T., Sabeur, Z., & Arbab-Zavar, B. (2023). Compressed Sensing Data with Performing Audio Signal Reconstruction for the Intelligent Classification of Chronic Respiratory Diseases. Sensors, 23(3), 1439. https://doi.org/10.3390/s23031439

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop