Next Article in Journal
The Measurement of Statistical Evidence as the Basis for Statistical Reasoning
Previous Article in Journal
Pandemic Influenza Preparedness in Africa
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Proceeding Paper

Detection of Arrhythmic Cardiac Signals from ECG Recordings Using the Entropy–Complexity Plane †

by
Pablo Martinez Coq
1,*,
Walter Legnani
1 and
Ricardo Armentano
2
1
Signals and Images Processing Centre (CPSI), Facultad Regional Buenos Aires, Universidad Tecnológica Nacional, Medrano 951, Buenos Aires C1179AAQ, Argentina
2
Bioengineering Development and Research Laboratory (GIBIO), Facultad Regional Buenos Aires, Universidad Tecnológica Nacional, Medrano 951, Buenos Aires C1179AAQ, Argentina
*
Author to whom correspondence should be addressed.
Presented at the 5th International Electronic Conference on Entropy and Its Applications, 18–30 November 2019; Available online: https://ecea-5.sciforum.net/.
Proceedings 2020, 46(1), 8; https://doi.org/10.3390/ecea-5-06693
Published: 18 November 2019

Abstract

:
The aim of this work was to analyze in the Entropy–Complexity plane (HxC) time series coming from ECG, with the objective to discriminate recordings from two different groups of patients: normal sinus rhythm and cardiac arrhythmias. The HxC plane used in this study was constituted by Shannon’s Entropy as one of its axes, and the other was composed using statistical complexity. To compute the entropy, the probability distribution function (PDF) of the observed data was obtained using the methodology proposed by Bandt and Pompe (2002). The database used in the present study was the ECG recordings obtained from PhysioNet, 47 long-term signals of patients with diagnosed cardiac arrhythmias and 18 long-term signals from normal sinus rhythm patients were processed. Average values of statistical complexity and normalized Shannon entropy were calculated and analyzed in the HxC plane for each time series. The average values of complexity of ECG for patients with diagnosed arrhythmias were bigger than normal sinus rhythm group. On the other hand, the Shannon entropy average values for arrhythmias patients were lower than the normal sinus rhythm group. This characteristic made it possible to discriminate the position of both signals’ groups in the HxC plane. The results were analyzed through a multivariate statistical test hypothesis. The methodology proposed has a remarkable conceptual simplicity, and shows a promising efficiency in the detection of cardiovascular pathologies.

1. Introduction

As its published by World Health Organization (WHO) of the 56.9 million deaths worldwide in 2016, light more than a half (54%) were due to the top 10 causes. Ischemic heart disease and stroke are the world’s biggest killers, accounting for a combined 15.2 million deaths in 2016. These diseases have remained the leading causes of death globally in the last 15 years [1].
The American Heart Association explains that some arrhythmias are so brief (for example, a temporary pause or premature beat) that the overall heart rate or rhythm is not affected at clinical level. However, if arrhythmias last longer, they may cause the heart rate to be too slow or too fast or the heart rhythm to be erratic—so the heart pumps less effectively [2].
The most important clinical significance of arrhythmias is related to an association with sudden cardiac death (Goldstein et al.; Recommended General Bibliography p. xvii; 1994) [3]. It is also important to remember that arrhythmias (especially atrial fibrillation) may frequently lead to embolisms, including cerebral emboli, often with severe consequences. Furthermore, it must be remembered that sometimes fast arrhythmias may trigger or worsen a heart failure.
For the analysis of certain time series coming from biomedical signals, such as an electrocardiogram (ECG) [4,5,6], there is a need to characterize the degree of organization or complexity of it. For this reason, the theory of information and the analysis of dynamics of the systems under study provide a way to obtain relevant results.
In order to analyze these time series, the first step is the computation of the PDF. For this, Bandt and Pompe [7] introduced a simple and robust symbolic method that takes into account the time causality connected with the dynamics of the system. Without any kind of assumption of the dynamics originating the signal, the permutation method proposed by Bandt and Pompe for the computation of the probability distribution function allows them to compute the entropy. Adding to the last, the complexity, computed from the Jensen–Shannon information divergence, constitutes the system’s behavior localization in an Entropy–Complexity plane [8,9,10,11,12,13,14].
As is mentioned in Olivares et al. [15], the concept of entropy has many interpretations arising from a wide diversity of scientific and technological fields. Among these associations is disorder, with the volume of state space, and with a lack of information too. There are various definitions according to ways of computing this important magnitude to study the dynamics of the systems, and one of the most frequent that could be considered of foundational definition is the denominated Shannon entropy [16], which can be interpreted as a measure of uncertainty. Shannon entropy can be considered as one of the most representative examples of information quantifiers.
Let an associated continuous Probability Distribution Function (PDF) be noted by ρ(x) with x ∈ Ω ⊂ R, a continuous variable measured, and ρ ( x ) d x = 1 ; then its associated Shannon Entropy S[ρ] is defined by [17] as:
S [ ρ ] = ρ ( x ) ln ( ρ ( x ) ) d x .
This concept allows for a global measure of the information contained in the time series; it has a low degree of sensitivity to strong changes in the distribution originating from a small-sized region of the set Ω.
For a time series X(t)≡{xt; t = 1, …, M}, a set of M measures of the observable X and the associated PDF, given by P = {pi; i = 1, …, N}, with i = 1 N p i = 1 , and N as the number of possible states of the system under study, the Shannon entropy [17] is defined in (2).
S [ ρ ] = i = 1 N p i   ln ( p i ) .
Equation (2) constitutes a function of the probability P = {pi; i = 1, …, N}, which is equal to zero when the outcomes of a certain experiment denoted by the index k associated with probabilities pk ≈ 1 will occur. Therefore, the known dynamics developed by the dynamical system under study are complete. If the knowledge of the system dynamics is minimal, all the states of the system can occur with equal probability; thus, this probability can be modeled by a uniform distribution Pe = {pi = 1/N;i = 1, …, N}.
It is useful to define the so-called normalized Shannon entropy, denoted as H[P], the expression of which is (3):
H [ P ] = S [ P ] S m a x   .  
Another information quantifier applied in this work was the Statistical Complexity Measure (SCM) which is a global informational quantifier. All the computations made in the present work were done with the definitions introduced in [18] and improved by [19]. For a discrete probability distribution function P = {pi; i = 1, …, N}, associated with a time series, this functional C[P] is given by (4).
C [ P ] = Q J [ P , P e ] . H [ P ] ,
where H denotes the amount of “disorder” given by the normalized Shannon entropy (3) and QJ is called “disequilibrium”, defined in terms of the Jensen–Shannon divergence, given by (5).
Q J [ P , P e ] = Q 0 J [ P , P e ] = Q 0 { S [ P + P e 2 ] S [ P ] 2 S [ P e ] 2 } ,
and Q0 denotes the normalization condition for the disequilibrium which corresponds to the inverse of the maximum possible value of Jensen–Shannon divergence, that is Q 0 = J [ P 0 , P e ] .
The C[P] quantifies the existence of correlational structures giving a measure of the complexity of a time series. In cases of, in perfect order or total randomness, a signal coming of a dynamical system, the value of the C[P] is identically null that means the signal possesses no structure. Between these two extreme instances, a large range of possible stages of physical structure may be realized using a dynamical system. These stages should be reflected in the features of the obtained PDF and quantified by a no-null C[P].
The global character of the SCM is due to the fact that its value does not change with different orderings computed by the PDF. Therefore, the C[P] quantifies the disorder, but also the degree of correlational structures. Given that the statistical complexity does not only quantify randomness but also the degree of correlation between structures; consequently, it is not a trivial function of entropy—in the sense that, for a given value of H, there is a range of possible values for C between a minimum value Cmin and a maximum value Cmax [20].
Once evaluated, the Shannon’s entropy and the statistical complexity, the results can be displayed in the HxC plane (global quantifier) with the objective to discriminate between the ECG coming from an arrhythmic diagnosed patient and that from a patient with normal sinus rhythms.
In general terms, these characterization methods have shown to be efficient in distinguishing between stochastic nature and deterministic chaos in different groups of time series, because it displays typical and specific features associated with its dynamics’ nature [21].

2. Results and Discussion

The groups of time series analyzed in this work were obtained from the PhysioNet [22] platform (managed by members of the Computational Physiology Laboratory of the M.I.T., Massachusetts Institute of Technology), which are available at https://physionet.org/.
As is shown in the Table 1, the registry of patients with normal sinus rhythms is made up of eighteen ECG recordings registered at the Beth Israel Deaconess Medical Center. On the other hand, the registry of patients with cardiac arrhythmias is made up of forty-seven ECG recordings, where twenty-five of them were taking Digoxin, a medication that is used to treat heart failure and certain types of irregular heartbeat, such as chronic atrial fibrillation. Likewise, an example of a normal sinus rhythm and a cardiac arrhythmia recording are represented in Figure 1, and the age distribution of each group of patients is shown in Figure 2.
Several computational codes were developed to evaluate the values of statistical complexity C and the normalized Shannon entropy H for each of the time series. Mean values of each pair of those calculations were represented in the HxC plane; where the statistical complexity was represented on Y-axis, and the normalized Shannon entropy was represented on the X-axis.
Table 2, Table 3, Table 4, Table 5, Table 6 and Table 7 are the summary of different comparisons of mean values, standard deviation, and the mean error of statistical complexity and the normalized Shannon entropy of the records under analysis.
Results arising from Table 3, Table 4, Table 5, Table 6 and Table 7 are represented in the HxC plane as it shown in Figure 3.

3. Conclusions

As is shown in Figure 4, the average values of statistical complexity obtained for the group of patients with arrhythmias were higher than those of the normal heart rate group. On the other hand, the average values of normalized Shannon Entropy for the group of patients with arrhythmias were smaller than those of the normal heart rate group. The combination of the above characteristics allowed us to discriminate the mean values of both groups of patients in the HxC plane. The difference between the two groups of interest was analyzed through multivariate statistical tests (Royston and Hotelling tests for samples of different sizes and with binormal distribution [23]). See Table 8 and Table 9.
According to values shown in Table 8, both groups of interest have a binormal distribution.
From the values obtained from previous multivariate statistical tests, the hypothesis that the mean values of both groups of interest in the HxC plane are different was accepted.
This shows that the signals from patients with normal ECGs have less statistical complexity in their waveform, measured from the divergence of the information of Jensen–Shannon and higher entropy values. This result confirms, as shown in the literature [24], that a normal ECG would be closer to the regions of the entropy complexity plane that usually brings together the dynamic systems called dissipative (by decreasing their n-dimensional volume in the phase space), and that they comprise the systems that develop deterministic chaos (characteristic of the control system of a healthy heart), while an ECG from a patient with arrhythmia is located in a region of the HxC plane that characterizes the so-called k-noise [21] and fractional Brownian motion. This last result indicates that cardiac arrhythmia could occur due to the loss of non-linear control (with the development of deterministic chaos behavior) in the heart, giving way to more disorderly (random) dynamics—typical of the pathology.
As seen in Table 7, referring to the use of Digoxin, the mean values for complexity and entropy are quite similar, which could be interpreted as showing that the information contained in the time series of the ECG does not allow the rejection of hypotheses of different means between the interest groups necessary to study the possible impact of the medication effects.
With the results obtained, an objective interpretation of cardiac pathology can be made from the informational measures (complexity and entropy) when they are used together, forming what has been agreed to be called the HxC plane—explaining part of the disease process in itself.
The representation in the HxC plane is a very useful method for the analysis of signals from the ECG of patients with this kind of pathology, providing an efficient way to make a statistical and deterministic analysis. Finally, the proposed methodology has a remarkable conceptual simplicity and shows a promising efficacy in the detection of cardiovascular pathologies.

Author Contributions

P.M.C. and W.L. conceived and designed the experiments; P.M.C. and W.L. performed the experiments; P.M.C. and W.L. analyzed the data; W.L. and R.A. contributed with his biomedical expertise; P.M.C., W.L. and R.A. wrote this paper.

Acknowledgments

This work was partially supported by the (CPSI) Signals and Images Processing Centre of the Facultad Regional Buenos Aires of the Universidad Tecnológica Nacional, City of Buenos Aires, Argentine. and Universidad Tecnológica Nacional Grant PID UTN 4729.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. WHO. The Top 10 Causes of Death. 2018. Available online: https://www.who.int/news-room/fact-sheets/detail/the-top-10-causes-of-death (accessed on 5 June 2019).
  2. American Heart Association. About Arrhythmia. 2016. Available online: https://www.heart.org/en (accessed on 3 May 2019).
  3. Bayés de Luna, A.; Baranchuk, A. Clinical Arrhythmology; John Wiley and Sons Ltd.: Hoboken, NJ, USA, 2017. [Google Scholar]
  4. Khan, G. Rapid ECG Interpretation; Humana Press Inc.: Totowa, NJ, USA, 2008. [Google Scholar]
  5. Bayés de Luna, A. Basic Electrocardiography: Normal and Abnormal ECG Patterns; Blackwell Publishing: Barcelona, Spain, 2007. [Google Scholar]
  6. Heart rate variability—Task Force of The European Society of Cardiology and The North American Society of Pacing and electrophysiology. Eur. Heart J. 1996, 17, 354–381.
  7. Bandt, C.; Pompe, B. Permutation entropy: A natural complexity measure for time series. Phys. Rev. Lett. 2002, 88, 174102. [Google Scholar] [CrossRef] [PubMed]
  8. Kolmogorov, A. A new metric invariant for transitive dynamical systems and automorphisms in Lebesgue spaces. Doklady Akademii Nauk SSSR 1959, 119, 861–864. [Google Scholar]
  9. Sinai, Y. On the concept of entropy for a dynamical system. Doklady Akademii Nauk SSSR 1959, 124, 768–771. [Google Scholar]
  10. Olivares, F.; Plastino, A.; Rosso, O. Ambiguities in the Bandt and Pompe’s methodology for local entropic quantifiers. Phys. A Stat. Mech. Its Appl. 2012, 391, 2518–2526. [Google Scholar] [CrossRef]
  11. Rosso, O.A.; Larrondo, H.A.; Martin, M.T.; Plastino, A.; Fuentes, M.A. Distinguishing noise from chaos. Phys. Rev. Lett. 2007, 99, 154102. [Google Scholar] [CrossRef] [PubMed]
  12. Rosso, O.; Olivares, F.; Plastino, A. Noise versus chaos in a causal Fisher- Shannon plane. Pap. Phys. 2015, 7, 070006. [Google Scholar] [CrossRef]
  13. Rosso, O.A.; Olivares, F.; Zunino, L.; De Micco, L.; Aquino, A.L.L.; Plastino, A.; Larrondo, H.A. Characterization of chaotic maps using the permutation Bandt–Pompe probability-distribution. Eur. Phys. J. 2013, 86, 116–128. [Google Scholar] [CrossRef]
  14. Amigó, J. Permutation Complexity in Dynamical Systems; Springer: Berlin/Heidelberg, Germany, 2010. [Google Scholar]
  15. Olivares, F.; Souza, L.; Legnani, W.; Rosso, O. Informational Time Causal Planes: A Tool for Chaotic Map Dynamic Visualization. Nonlinear Syst. 2019. [Google Scholar] [CrossRef]
  16. Shannon, C.E.; Weaver, W. The mathematical theory of communication. Bell Syst. Tech. J 1948, 27, 379–423. [Google Scholar] [CrossRef]
  17. Brissaud, J.B. The meaning of entropy. Entropy 2005, 7, 68–96. [Google Scholar] [CrossRef]
  18. López-Ruiz, R.; Mancini, H.L.; Calbet, X. A statistical measure of complexity. Phys. Lett. A 1995, 209, 321–326. [Google Scholar] [CrossRef]
  19. Lamberti, P.; Martín, M.; Plastino, A.; Rosso, O. Intensive entropic non-triviality measure. Phys. A Stat. Mech. Appl. 2004, 334, 119–131. [Google Scholar] [CrossRef]
  20. Martin, M.T.; Plastino, A.; Rosso, O.A. Generalized statistical complexity measures: Geometrical and analytical properties. Phys. A Stat. Mech. Appl. 2006, 369, 439–462. [Google Scholar] [CrossRef]
  21. Mateos, D. Doctoral thesis; Medidas de Complejidad y de Información como herramientas para el análisis de series temporales. Aplicaciones al estudio de señales de origen electrofisiológicas., Córdoba – Argentine, 2016, 10. Available online: http://hdl.handle.net/11086/2853 (accessed on 5 June 2019).
  22. PhysioNet, the Repository of Freely-Available Medical Research Data, Managed by the MIT Laboratory for Computational Physiology. Available online: https://physionet.org/about/database/ (accessed on 11 February 2019).
  23. Mardia, K.; Kent, J.; Bibby, J. Multivariate Analysis; Academic Press: Cambridge, MA, USA, 1995. [Google Scholar]
  24. Clinton Sprott, J. Chaos and Time-Series Analysis; Oxford University Press: Oxford, UK, 2003. [Google Scholar]
Figure 1. The orange trace represents record named from origin 16273 from MIT-BIH Normal sinus rhythm database, and light blue trace represents record 100 from MIT-BIH Arrhythmia database.
Figure 1. The orange trace represents record named from origin 16273 from MIT-BIH Normal sinus rhythm database, and light blue trace represents record 100 from MIT-BIH Arrhythmia database.
Proceedings 46 00008 g001
Figure 2. Age distribution of each group of patients: (a) Normal sinus rhythm recordings; (b) Cardiac arrhythmias recordings.
Figure 2. Age distribution of each group of patients: (a) Normal sinus rhythm recordings; (b) Cardiac arrhythmias recordings.
Proceedings 46 00008 g002
Figure 3. HxC plane. Mean values of statistical complexity and the normalized Shannon entropy of different comparisons of sub-groups of interest.
Figure 3. HxC plane. Mean values of statistical complexity and the normalized Shannon entropy of different comparisons of sub-groups of interest.
Proceedings 46 00008 g003
Figure 4. Detailed HxC for discriminating the mean values of both groups of interest.
Figure 4. Detailed HxC for discriminating the mean values of both groups of interest.
Proceedings 46 00008 g004
Table 1. Distribution of each group of interest.
Table 1. Distribution of each group of interest.
Normal Sinus Rhythm Cardiac Arrhythmias
Recordings1847 1
Males526
Females1321
Taking medicationnot available36
1 Signals denominated from origin 221 and 234 were discarded because there was no relevant information in it.
Table 2. Mean HxC values for the whole database under analysis.
Table 2. Mean HxC values for the whole database under analysis.
Normal Sinus RhythmCardiac Arrhythmias
Entropy—HComplexity—CEntropy—HComplexity—C
Mean0.895110.190600.819840.28982
S.D.0.030020.045620.049240.05739
Mean Error0.007280.011060.007260.00846
Table 3. Mean HxC values for records belonging to female patients of both groups of interest.
Table 3. Mean HxC values for records belonging to female patients of both groups of interest.
Normal Sinus RhythmCardiac Arrhythmias
Entropy—HComplexity—CEntropy—HComplexity—C
Mean0.907430.172340.824570.27700
S.D.0.017650.027250.046560.04939
Mean Error0.005090.007870.010160.01078
Table 4. Mean HxC values for records belonging to male patients of both groups of interest.
Table 4. Mean HxC values for records belonging to male patients of both groups of interest.
Normal Sinus RhythmCardiac Arrhythmias
Entropy—HComplexity—CEntropy—HComplexity—C
Mean0.865530.234410.815880.30059
S.D.0.034630.053700.051990.06228
Mean Error0.015490.024020.010400.01246
Table 5. Mean HxC values for records belonging to patients up to 40 years of both groups of interest.
Table 5. Mean HxC values for records belonging to patients up to 40 years of both groups of interest.
Normal Sinus RhythmCardiac Arrhythmias
Entropy—HComplexity—CEntropy—HComplexity—C
Mean0.902640.181400.829800.28171
S.D.0.028590.045380.022110.02109
Mean Error0.008250.013100.009030.00861
Table 6. Mean HxC values for records belonging to patients from 40 up to 60 years of both groups of interest.
Table 6. Mean HxC values for records belonging to patients from 40 up to 60 years of both groups of interest.
Normal Sinus RhythmCardiac Arrhythmias
Entropy—HComplexity—CEntropy—HComplexity—C
Mean0.877040.212690.820290.30795
S.D.0.027890.042400.033410.05456
Mean Error0.012470.018960.011810.01929
Table 7. Mean HxC values for all records taking and not taking medication. Only for cardiac arrhythmias.
Table 7. Mean HxC values for all records taking and not taking medication. Only for cardiac arrhythmias.
Taking MedicationsNot Taking Medications
Entropy—HComplexity—CEntropy—HComplexity—C
Mean0.813760.290370.827090.28917
S.D.0.056060.060200.039760.05533
Mean Error0.011210.012040.008680.01207
Table 8. Royston test for both groups of patients with a given significance of 0.050.
Table 8. Royston test for both groups of patients with a given significance of 0.050.
NormalCardiac Arrhythmias
Royston’s statistic1.827830.15397
Equivalent deg. of freedom0.5275070.92392
P-value0.0823300.662806
Table 9. Hotelling Test for binormal distributions of both groups of interest.
Table 9. Hotelling Test for binormal distributions of both groups of interest.
T2F(2,60)p
Normal vs. Arrhythmias40.14119.7410.0000002583
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Coq, P.M.; Legnani, W.; Armentano, R. Detection of Arrhythmic Cardiac Signals from ECG Recordings Using the Entropy–Complexity Plane. Proceedings 2020, 46, 8. https://doi.org/10.3390/ecea-5-06693

AMA Style

Coq PM, Legnani W, Armentano R. Detection of Arrhythmic Cardiac Signals from ECG Recordings Using the Entropy–Complexity Plane. Proceedings. 2020; 46(1):8. https://doi.org/10.3390/ecea-5-06693

Chicago/Turabian Style

Coq, Pablo Martinez, Walter Legnani, and Ricardo Armentano. 2020. "Detection of Arrhythmic Cardiac Signals from ECG Recordings Using the Entropy–Complexity Plane" Proceedings 46, no. 1: 8. https://doi.org/10.3390/ecea-5-06693

APA Style

Coq, P. M., Legnani, W., & Armentano, R. (2020). Detection of Arrhythmic Cardiac Signals from ECG Recordings Using the Entropy–Complexity Plane. Proceedings, 46(1), 8. https://doi.org/10.3390/ecea-5-06693

Article Metrics

Back to TopTop