Next Article in Journal
Economic Freedom: The Top, the Bottom, and the Reality. I. 1997–2007
Next Article in Special Issue
Permutation Entropy of Weakly Noise-Affected Signals
Previous Article in Journal
Reliable Fault Diagnosis of Bearings Using an Optimized Stacked Variational Denoising Auto-Encoder
Previous Article in Special Issue
Evaluating Temporal Correlations in Time Series Using Permutation Entropy, Ordinal Probabilities and Machine Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On Rényi Permutation Entropy

Institute of Mathematics, University of Lübeck, D-23562 Lübeck, Germany
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(1), 37; https://doi.org/10.3390/e24010037
Submission received: 6 December 2021 / Revised: 21 December 2021 / Accepted: 23 December 2021 / Published: 24 December 2021

Abstract

:
Among various modifications of the permutation entropy defined as the Shannon entropy of the ordinal pattern distribution underlying a system, a variant based on Rényi entropies was considered in a few papers. This paper discusses the relatively new concept of Rényi permutation entropies in dependence of non-negative real number q parameterizing the family of Rényi entropies and providing the Shannon entropy for q = 1 . Its relationship to Kolmogorov–Sinai entropy and, for q = 2 , to the recently introduced symbolic correlation integral are touched.

1. Introduction

1.1. Paper Background and Motivation

Since Bandt and Pompe [1] introduced the concept of permutation entropy (PE), it has been applied in different fields from biomedicine to econophysics (e.g., Zanin et al. [2], and Amigò et al. [3]) and developed into various directions. One relatively new variant of permutation entropy is based on Rényi entropies instead of the originally used Shannon entropy and is called Rényi permutation entropy (RPE). Roughly speaking, RPE quantifies the complexity of the distribution of ordinal patterns of some length n underlying a dynamical system, where ordinal patterns describe the up and down in the dynamics. As Rényi entropies depend on a parameter q [ 0 , ] , there are also different choices of RPE depending on q.
The central aim of the paper is to discuss the asymptotics of RPE for increasing pattern length. This is motivated by the striking fact that, under certain assumptions, asymptotic PE is equal to Kolmogorov–Sinai entropy, which was first observed by Bandt et al. [4]. This paper shows that the situation for q 1 is more complicated than that for q = 1 .
The paper is organized as follows. It first follows a short overview of first applications of the RPE. Section 2 provides the main definitions. The concepts of RPE are introduced in empirical and model-based settings. Moreover, RPE is discussed for some special q, including q = as a limit case. Section 3 is devoted to the asymptotics of RPE and PE. With Corollary 1, the section contains the main new result of the paper relating RPE to Kolmogorov–Sinai entropy for q [ 0 , 1 ] and measures with maximal entropy. Its proof and a class of discriminating examples for q > 1 (see Example A1) are given in Appendix A.

1.2. First Applications of RPE

To our best knowledge, the concept of RPE was first considered in the literature in 2015. In a study of monitoring the depth of anaesthesia by EEG, Liang et al. [5] systematically compared 12 entropy measures, with RPE among them. They reported that RPE had the best performance in distinguishing different anaesthesia states. Mammone et al. [6] discussed RPE in the context of absence epilepsy EEG. Their results suggested improved abilities in classifying ictal and interictal EEG by using RPE (with suitable parameters) instead of PE. Zunino et al. [7] introduced permutation min entropy, which is the limit of Rényi entropy for defining parameter q approaching to as a tool for finding temporal correlations in a time series.
Moreover, Rivero et al. [8] combined an enhanced Bayesian approach and RPE for predicting long-term time series. Following the results of Liang et al., Park et al. [9] used RPE for comparing anaesthetics given during a cesarean section with similar results as those for other entropy measures. Different variants of RPE, from weighting to multiscaling, have been applied to complex stock-market data (Zhou and Shang [10], and Chen et al. [11]). Some remarks on RPE can also be found in [12].

2. Rényi Entropies

2.1. General Entropy Concept

Given a finite index set I consisting of n elements, ( p i ) i I R n is called a stochastic vector if p i 0 for all i I and i I p i = 1 . The Rènyi entropy  RE ( ( p i ) i I , q ) of a stochastic vector ( p i ) i I for q [ 0 , ] is defined by
RE ( ( p i ) i I , q ) =   i I p i log p i for   q = 1 1 1 q log i I p i q else .
The Rényi entropy of a fixed stochastic vector monotonically decreases and is continuous with respect to q. It generalizes the Shannon entropy given in the standard case q = 1 . The larger that q is, the more the role of the largest entries in the stochastic vector is emphasized, and the smaller that q is, the more equal the role of all positive entries in the entropy formula is.
On the basis of the concept of Rényi entropies, we want to give precise definitions of RPE regarding both the empirical and the modelling viewpoint.

2.2. Empirical RPE

For n N , we denote the set of permutations of { 0 , 1 , , n 1 } by S n . A vector ( v 0 , v 1 , , v n 1 ) R n has ordinal pattern π = ( π 0 , π 1 , , π n 1 ) S n if
v π 0 v π 1 v π n 2 v π n 1
and
r l 1 > r l   i f   v r l 1 = v r l .
The latter requirement realises the uniqueness of ordinal patterns.
Definition 1.
The empirical Rényi permutation entropy for q [ 0 , ] and n N of a time series ( x t ) t = 0 N 1 is defined by
ePE ( x t ) t = 0 N 1 , q , n = RE ( ( p π ) π S n , q ) = π S n p π log p π for   q = 1 1 1 q log π S n p π q else
with
p π τ = # { t { 0 , 1 , , N n } ( x t , x t + 1 , , x t + n 1 ) has ordinal pattern π } N n + 1
being the relative frequency of ordinal patterns π in the time series, and 0 log 0 and 0 0 being defined by 0.

2.3. RPE

On the model side, we consider a measure-preserving dynamical system  ( Ω , A , μ , T ) , defined as a probability space ( Ω , A , μ ) being equipped with a A - A -measurable map T : Ω Ω which satisfies μ ( T 1 ( A ) ) = μ ( A ) for all A A . T and the system ( Ω , A , μ , T ) are called ergodic if T 1 ( A ) = A for A A implies μ ( A ) { 0 , 1 } .
Generally, the dynamics of T can be related to ordinal patterns via a real-valued random variable X on Ω by assigning ω Ω the ordinal pattern of ( X ( ω ) , X ( T ( ω ) ) , , X ( T n 1 ( ω ) ) . Here, X is interpreted as observable modelling of a measuring process. If X (or, more generally, a collection of random variables) has certain separation properties, the ordinal patterns obtained via X (or all random variables) contain much information on the given system. In the following, however, we usually assume that Ω is a subset of R , and the ordinal pattern Π ( x ) assigned to x Ω is that taken from ( x , T ( x ) , , T n ( x ) ) (this is equivalent to considering an X being the identity map).
For a permutation π = ( π 0 , π 1 , , π n 1 ) S n , we denote the sets of all points x with ordinal pattern Π ( x ) = π by P π . From these sets, we obtain the partition
O P ( n ) = { P π Ø π S n }
of Ω , with this being central to considering RPE on the model side. Analogously to empirical permutation entropy, we define Rényi permutation entropy for q [ 0 , ] and n N on the basis of O P ( n ) by
PE T , q , n = RE ( ( μ ( P π ) ) π S n , q ) = π S n μ ( P π ) log μ ( P π ) for   q = 1 1 1 q log π S n μ ( P π ) q else .

2.4. Estimation

Given an orbit
( x 0 , x 1 , , x N 2 , x N 1 ) = ( x , T ( x ) , , T N 2 ( x ) , T N 1 ( x ) )
of some x Ω , it is natural to estimate μ ( P π ) for π S n and PE T , q , n by p π and ePE ( x t ) t = 0 N 1 , q , n , respectively. In the case that T is ergodic, by Birkhoff’s ergodic theorem, the corresponding estimators are asymptotically consistent. This particularly means that
lim N ePE ( T t ( x ) ) t = 0 N 1 , q , n   = PE T , q , n
for μ -almost all x Ω .

2.5. RPE for Special Parameters q

In the following, we discuss the RPE for some special parameters q, and touch the general concept of Rényi entropies.
  • q = 0 : Rényi entropy for q = 0 is the well-known Hartley entropy, and the RPE of a measure preserving dynamical system is no more than the logarithm of the number of ordinal patterns appearing with positive probability. Rényi entropy for q = 0 is also called max entropy since it is maximal among Rényi entropies.
  • q = 1 : This case providing the standard (Shannon) permutation entropy has been discussed in various papers both from a theoretical and an application viewpoint. We particularly refer to the literature mentioned in several parts of this paper.
  • q = 2 : Rényi entropy for q = 2 , also called quadratic entropy or collision entropy, is used in different fields. It is obviously related to the Simpson index i = 1 n p i 2 given for a stochastic vector ( p i ) i = 1 n and used as a diversity measure in ecology (see [13]). Given a measure-preserving dynamical system ( Ω , A , μ , T ) , we look at the RPE for q = 2 .
By Fubini’s theorem, it holds that
PE T , 2 , n = log π S n μ ( P π ) 2 = log π S n 1 P π d μ 2 = log π S n 1 P π ( ω 1 ) 1 P π ( ω 2 ) d μ ( ω 1 ) d μ ( ω 2 ) = log π S n ( 1 P π ( ω 1 ) 1 P π ( ω 2 ) ) d μ ( ω 1 ) d μ ( ω 2 ) = log π S n ( 1 P π ( ω 1 ) 1 P π ( ω 2 ) ) d μ 2 ( ω 1 , ω 2 ) = log S n
with
S n = μ 2 { ( ω 1 , ω 2 ) Ω × Ω Π ( ω 1 ) = Π ( ω 2 ) } .
Here, 1 A stands for the indicator of a set A assigning a point the value 1 if it belongs to A and value 0 otherwise, and μ 2 denotes the product measure of μ with itself.
So, PE T , 2 , n is related to the probability that the ordinal patterns of length n of two independently (with respect to μ ) chosen points coincide.
A natural estimation of S n based on a finite orbit
( x 0 , x 1 , , x N 2 , x N 1 ) = ( x , T ( x ) , , T N 2 ( x ) , T N 1 ( x ) )
of some x Ω is given by
2 ( N + 1 n ) ( N n ) # { ( s , t ) 0 s < t N 1 , Π ( x s ) = Π ( x t ) }
providing the relative frequency of pairs in the orbit with coinciding (completely defined) ordinal patterns. This qualifies the RPE for q = 2 as a recurrence measure.
Quantity (1) was introduced by Caballero et al. [14] as the symbolic correlation integral in the context of a stochastic process and studied mainly in the i.i.d. case.
  • q = : It is well-known that the Rényi entropy of a stochastic vector ( p i ) i I for q converges to value
    RE ( ( p i ) i I , ) : = lim q RE ( ( p i ) i I , q ) = log max i I p i .
    This fact can be used to reconstruct a stochastic vector up to permuting its components from its Rényi entropies for an unbounded sequence ( q n ) n N (see Appendix A). Since RE ( ( p i ) i I , ) RE ( ( p i ) i I , q ) for all q [ 0 , ] , number RE ( ( p i ) i I , ) is called min entropy of ( p i ) i I . Applications of min entropy in the permutation entropy context can be found in Zunino et al. [7]. In the following, we further assume that q < .

3. Asymptotics of RPE and PE

As already mentioned, there is a strong relationship between Kolmogorov–Sinai entropy and PE. The result of Takens and Verbitskiy [15] that, for q > 1 , Kolmogorov–Sinai entropy can be expressed by a limit on the basis of Rényi entropies instead of Shannon entropies suggests the question whether PE can be similarly replaced by RPE in that relationship. This question addresses the asymptotics of RPE, and the general nature of RPE is thus in the centre of this section.

3.1. Kolmogorov–Sinai Entropy via Rényi Entropies

Definitions and statements of this subsection go back to Takens and Verbitsky [15]. Many considerations of this paper are related to partitions of Ω . We generally assume that, in a context where a σ -algebra on Ω is specified, such partitions are contained in it.
Let ( Ω , A , μ , T ) now be a measure-preserving dynamical system, and consider a finite partition P = { P i } i I of Ω . For n N and multi-indices i = ( i 0 , i 1 , , i n 1 ) I n define the sets
P ( i ) : = t = 0 n 1 T t ( P i t ) = P i 0 T 1 ( P i 1 ) T n + 1 ( P i n 1 )
forming the partition
P ( n ) = { P ( i ) Ø | i ( i 0 , i 1 , i n 1 ) I n }
of Ω . For q [ 0 , ] , the generalized entropy rate of T with respect to partition P is defined as
h ( T , P , q ) : = lim inf n 1 n H ( P ( n ) , q )
with
H ( Q , q ) = Q Q μ ( Q ) log μ ( Q ) for q = 1 1 1 q log Q Q μ ( Q ) q else
for a finite partition Q of Ω . Generalized Kolmogorov–Sinai entropy for q [ 0 , ] is defined as the supremum of generalized entropy rates taken over all finite partitions:
h ( T , q ) : = sup P   finite   partition   of   Ω h ( T , P , q ) .
(Standard) Kolmogorov–Sinai entropy is given by
h ( T ) : = h ( T , 1 ) .
In the case of q = 1 , the limit inferior in (3) can be replaced by a limit and that, for Ω being an interval and A being the Borel σ -algebra, Kolmogorov–Sinai entropy is already determined by finite interval partitions defined as finite partitions consisting of intervals (e.g., [16]):
h ( T ) = sup P   finite   interval   partition   of   Ω h ( T , P , 1 ) .
The following theorem of Takens and Verbitskiy [15] was originally proved for invertible systems; however, it also holds true for noninvertible systems (see Verbitskiy [17]). Assumption (i) of ergodicity can be relaxed (see Takens and Verbitskiy [18]); however, we do not go into the technical details.
Theorem 1.
Let ( Ω , A , μ ) be a standard-probability space and T : Ω Ω an aperiodic and ergodic measure-preserving function. Then,
h ( T , q ) = h ( T )
holds true for all q [ 1 , ] . Additionally, h ( T ) > 0 implies
h ( T , q ) =
for all q [ 0 , 1 ] .
Here, T is called aperiodic if the set of periodic points has measure zero with respect to μ . The property of a probability space to be standard is a relatively technical one; however, it is not very restrictive since it is principally satisfied for the most common probability spaces (e.g., Walters [16]).

3.2. Kolmogorov–Sinai Entropy and RPE

In order to discuss the relationship between RPE and Kolmogorov–Sinai entropy of a measure-preserving dynamical system ( Ω , A , T , μ ) , we define lower and upper Rényi permutation entropies PE ̲ ( T , q ) and PE ¯ ( T , q ) for q [ 0 , ] as
PE ̲ ( T , q ) = lim inf n 1 n H O P ( n ) , q
and
PE ¯ ( T , q ) = lim sup n 1 n H O P ( n ) , q ,
respectively. We write PE ̲ ( T ) and PE ¯ ( T ) in the case of q = 1 . Both PE ̲ ( T , q ) and PE ¯ ( T , q ) monotonically decreasee with respect to q by the definition of Rényi entropies.
q = 1 : The celebrated result of Bandt et al. [4] that, for piecewise continuous and monotone interval maps, the permutation and Kolmogorov-Sinai entropy coincide is the motivation for the following discussion. Here, we state the more general version of the result proved in Gutjahr and Keller [19], but afterwards return to the case of piecewise monotone interval maps. In the following, we call a subset of Ω R interval if it is the intersection of an interval of R with Ω or a one point set.
Theorem 2
([19]). Let ( Ω , B , μ , T ) be a measure-preserving dynamical system, with Ω R being compact, and B being the Borel σ-algebra on Ω. If there exists a finite partition M or a countable partition M with H ( M ) < of Ω into intervals, such that T is monotone on each of the intervals, then
PE ¯ ( T ) = h ( T ) .
Theorem 2 covers interval maps, since a noncompact Ω can be replaced by compactification without substantially changing the structure of the given system.
q > 1 : In light of the statement of Takens and Verbitskiy [15] mentioned above, it is a natural question whether also PE ¯ ( T , q ) = h ( T , q ) ( = h ( T ) ) . The general answer is no. Examples with PE ¯ ( T , q ) < h ( T , q ) covering all q > 1 are given by Example A1 in Appendix B.2.
q < 1 : We also look at case q < 1 in the class of maps considered by Bandt et al. in [4]. For this, let Ω be an interval, B the Borel σ -algebra on it, and M be a finite partition of Ω into intervals on each of which T is monotone and continuous. For such a map, it was shown in [4] that
lim n 1 n log # { P π O P ( n ) P π Ø } = lim n 1 n log # { M M ( n ) M Ø }
holds true. Using the fact that Rényi entropy monotonically decreases in q, this implies
PE ¯ ( T , q ) PE ¯ ( T , 0 ) = lim sup n 1 n log P π O P ( n ) μ ( P π ) 0 = lim sup n 1 n log # { P π O P ( n ) μ ( P π ) > 0 } lim n 1 n log # { P π O P ( n ) P π Ø } = lim n 1 n log # { M M ( n ) M Ø }
for all q [ 0 , ] . Let us summarize:
Proposition 1.
Let ( Ω , B , μ , T ) be a measure-preserving dynamical system, with Ω R being an interval, and B being the Borel σ-algebra on it. Suppose that M is a finite partition of Ω into intervals such that T is monotone and continuous on each of the intervals. Then
PE ¯ ( T , q ) lim n 1 n log # { M M ( n ) M Ø }
holds true for all q [ 0 , ] .
Quantity lim sup n 1 n log # { P π O P ( n ) P π Ø } could be considered to be a topological version of permutation entropy. This is justified by (4) for T, as defined above (4), and the following: If T is continuous on all of Ω , Misiurewicz and Szlenk showed that lim n 1 n log # { M M ( n ) M Ø } is equal to the topological entropy of T [20]. For the definition of topological entropy and the following, see, e.g., [16].
By variation principle, the topological entropy of a map T on a compact Hausdorff space is equal to the supremum of the Kolmogorov–Sinai entropy of all Borel measures for which T is measure-preserving (e.g., [16]). Often, topological entropy is assumed by the Kolmogorov–Sinai of such a measure. Generally, given a continuous map T on a metric space, a corresponding Borel measure being measure-preserving has maximal entropy if its Kolmogorov–Sinai entropy coincides with the topological entropy of T.
On the basis of the discussion above, we show the following statement (see Appendix B.1).
Corollary 1.
Let ( Ω , B , μ , T ) be a measure-preserving dynamical system, with Ω R being an interval, and B being the Borel σ-algebra on Ω. Suppose that T is continuous, and there exists a finite partition of Ω into intervals, such that T is monotone on each of those intervals, and μ is a measure of maximal entropy. Then
PE ̲ ( T , q ) = PE ¯ ( T , q ) = sup P   f i n i t e   i n t e r v a l   p a r t i t i o n   o f   Ω h ( T , P , q ) = h ( T )
holds true for all q [ 0 , 1 ] .

4. Conclusions

In this paper, we looked more closely at the recently introduced and used Rényi variant of permutation entropy, depending on a parameter q [ 0 , ] , which is called Rényi permutation entropy (RPE) here. Giving a summary of first applications of RPE, and discussing RPE for some special parameter q, we mainly focused on the asymptotics of RPE for ordinal pattern length going to . This was motivated by the fact that the usual permutation entropy (PE) often asymptotically coincides with Kolmogorov–Sinai entropy, and that, for q > 1 , Kolmogorov–Sinai entropy can be defined by Rényi entropies instead of Shannon entropies.
This paper showed that, for q > 1 , asymptotics of RPE can be different from that of PE, meaning that, for long ordinal patterns, the nature of RPE is also not the same as that of PE. One the other hand, it is interesting that, for continuous piecewise monotone interval maps with a measure of maximal entropy and q < 1 , asymptotics of RPE and PE are the same. Results indicate that the behaviour of general RPE is more specific than that of PE, although the asymptotics of PE is not completely understood. Further work for the better understanding of RPE for large pattern lengths is necessary.
The content of this paper is more or less purely mathematical, but in a certain sense, it justifies the application of RPE in dynamical systems and time series besides PE. Some of the applications mentioned at the beginning of the paper underline the benefit of using RPE. There is, however, the other interesting point that special q address special features; so, for example, q = 2 is related to recurrence. The symbolic correlation integral related to case q = 2 is a U-statistic, which is helpful in the statistical analysis of the corresponding entropy. Work on utilizing this fact for testing for asymmetry in temporal data is in progress.

Author Contributions

K.K. and T.G. designed and wrote the paper, and T.G. provided all statements and proofs given in Appendix B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Reconstruction of Stochastic Vectors from Rényi Entropies

Let ( p i ) i = 1 m be a stochastic vector, and let ( q n ) n N be an unbounded sequence of positive numbers, such that RE ( ( p i ) i = 1 m , q n ) is known for all n N . Assume that p 1 p 2 p m . The reconstruction we give is inductive.
If p 1 , p 2 , , p r 1 are known for some r { 1 , 2 , , m } and j = 1 r 1 p j < 1 , then p i 1 j = 1 r 1 p j i = r m is a stochastic vector and
RE ( ( p i ) i = 1 m , q n ) = 1 1 q n log i = 1 r 1 p i q n + i = r m p i q n
for all n N . This implies
i = r m p i q n = e ( 1 q n ) RE ( ( p i ) i = 1 m , q n ) i = 1 r 1 p i q n
for all n N . From this and (2), we obtain
log p r 1 j = 1 r 1 p j = inf n N RE p i 1 j = 1 r 1 p j i = r m , q n = inf n N 1 1 q n log i = r m p i 1 j = 1 r 1 p j q n = inf n N 1 1 q n log i = r m p i q n q n 1 q n log 1 j = 1 r 1 p j ,
hence
log p r = inf n N 1 1 q n log i = r m p i q n log 1 j = 1 r 1 p j
and
log p r = sup n N 1 q n 1 log e ( 1 q n ) RE ( ( p j ) j = 1 m , q n ) i = 1 r 1 p i q n log 1 j = 1 r 1 p j
Thus, p r is also known.

Appendix B. Proofs

Some of the following statements are only given on a level of generality required here. In some cases, for more general statements and further details, we refer to Gutjahr [21].

Appendix B.1. Proof of Corollary 1

First, if a finite partition P of Ω is finer than another one, Q , then H ( P , q ) H ( Q , q ) for all q [ 0 , ] . This is well-known and easily follows from the concavity and convexity of x x q for q < 1 and q > 1 , respectively, and from the convexity of x x log x for q = 1 on [ 0 , ] .
Given two finite partitions P , Q of Ω , let P Q = { P Q Ø P P , Q Q } be the largest common refinement of P and Q . Further, for a partition P of Ω and some Q A , let Δ ( P | Q ) : = { P P μ ( P Q ) > 0 } and # Δ ( P | Q ) the number of elements of Δ ( P | Q ) .
The proof of Corollary 1 is built on the two following lemmata.
Lemma A1.
Let ( Ω , A , μ ) be a probability space, and P , Q two finite partitions of Ω. Then
H ( P Q , q ) H ( Q , q ) + Q Q μ ( Q ) · log ( # Δ ( P | Q ) ) for q = 1 max Q Q log ( # Δ ( P | Q ) ) for q [ 0 , ] \ { 1 } .
Proof. 
Consider function f : [ 0 , ] R with f ( x ) = x q . Since f is concave for q < 1 and convex for q > 1 , Jensen’s inequality implies
P Δ ( P | Q ) f ( μ ( P Q ) ) # Δ ( P | Q ) f P Δ ( P | Q ) μ ( P Q ) # Δ ( P | Q ) if q < 1 f P Δ ( P | Q ) μ ( P Q ) # Δ ( P | Q ) if q > 1
for all Q Q . Using this and the fact that function 1 q 1 log monotonically increases for q < 1 and decreases for q > 1 provides
H ( P Q , q ) = 1 q 1 log Q Q P P μ ( P Q ) q = 1 q 1 log Q Q P Δ ( P | Q ) μ ( P Q ) q = 1 q 1 log Q Q # Δ ( P | Q ) P Δ ( P | Q ) f ( μ ( P Q ) ) # Δ ( P | Q ) 1 q 1 log Q Q # Δ ( P | Q ) · f P Δ ( P | Q ) μ ( P Q ) # Δ ( P | Q ) = 1 q 1 log Q Q # Δ ( P | Q ) · f μ ( Q ) # Δ ( P | Q ) = 1 q 1 log Q Q # Δ ( P | Q ) 1 q · μ ( Q ) q 1 q 1 log max Q Q # Δ ( P | Q ) 1 q · Q Q μ ( Q ) q = 1 q 1 log max Q Q # Δ ( P | Q ) 1 q + 1 q 1 log Q Q μ ( Q ) q = max Q Q log ( # Δ ( P | Q ) ) + H ( Q , q )
for all q 1 .
For q = 1 , consider function g : [ 0 , ] [ 0 , ] with g ( x ) = x log ( x ) , which is convex. Therefore, by Jensen’s inequality, it follows that
P Δ ( P | Q ) μ ( P Q ) log ( μ ( P Q ) ) = # Δ ( P | Q ) P Δ ( P | Q ) 1 # Δ ( P | Q ) · g ( μ ( P Q ) ) # Δ ( P | Q ) · g P Δ ( P | Q ) 1 # Δ ( P | Q ) · μ ( P Q ) = # Δ ( P | Q ) · g μ ( Q ) # Δ ( P | Q ) = # Δ ( P | Q ) · μ ( Q ) # Δ ( P | Q ) · log μ ( Q ) # Δ ( P | Q ) = μ ( Q ) · log ( μ ( Q ) ) log ( # Δ ( P | Q ) ) .
Using the above inequality provides
H ( P Q ) = Q Q P P μ ( P Q ) log ( μ ( P Q ) ) = Q Q P · ( P | Q ) μ ( P Q ) log ( μ ( P Q ) ) Q Q μ ( Q ) · log ( μ ( Q ) ) log ( # Δ ( P | Q ) ) = H ( Q ) + Q Q μ ( Q ) · log ( # Δ ( P | Q ) ) .
 □
The following lemma was proved in Gutjahr and Keller [19] (see Lemma A1). We present it here in a slightly weakened form since we do not need the full generality.
Lemma A2.
Let ( Ω , B , μ , T ) be a measure-preserving dynamical system, with Ω R being an interval and B being the Borel σ-algebra on Ω. Further, let P be a finite interval partition of Ω. Then for all P π O P ( n )
# Δ ( P ( n ) | P π )   n + # P 1 # P 1
holds true.
We can now finalise the proof of Corollary 1. Given the assumptions of the corollary, by Proposition 1 and the monotony of the Rényi entropy with respect to q, we have
PE ̲ ( T , q ) PE ¯ ( T , q ) lim n 1 n log # { M M ( n ) M Ø } = h ( T ) = sup P   finite   interval   partition   of   Ω h ( T , P ) sup P   finite   interval   partition   of   Ω h ( T , P , q )
for all q [ 0 , 1 ] . So, it remains to show that
PE ̲ ( T , q ) h ( T , P , q )
is true for all q [ 0 , 1 ] and for all finite interval partitions P of Ω .
Fixing some q [ 0 , 1 ] and some finite interval partition of Ω , Lemmas A1 and A2 yield
H ( P ( n ) , q ) H ( P ( n ) O P ( n ) , q ) H ( O P ( n ) , q ) + max π S n log ( # Δ ( P ( n ) | P π ) ) H ( O P ( n ) , q ) + max π S n log n + # P 1 # P 1 H ( O P ( n ) , q ) + max π S n log ( n + # P 1 ) # P 1 H ( O P ( n ) , q ) + ( # P 1 ) log ( n + # P 1 )
for all q [ 0 , 1 ] . Hence,
PE ̲ ( T , q ) = lim inf n 1 n H O P ( n ) , q lim inf n 1 n H ( P ( n ) , q ) ( # P 1 ) log ( n + # P 1 ) = lim inf n 1 n H P ( n ) , q   = h ( T , P , q ) .

Appendix B.2. Counter Examples

In order to give the examples after Theorem 2, we start with two lemmata.
Lemma A3.
Let ( Ω , B , μ , T ) be a measure-preserving dynamical system with Ω R and B being the Borel σ-algebra on Ω. Further, let P = { P i } i I be a finite partition of Ω. Then, for all q [ 1 , ] ,
PE ¯ ( T , q ) h ( T , P , q ) + lim sup n 1 n max i I n log ( # Δ ( O P ( n ) | P ( i ) ) )
hold true.
Proof. 
Let P = { P i } i I be a finite partition of Ω . Then, Lemma A1 provides
H ( O P ( n ) , q ) H ( O P ( n ) P ( n ) , q ) H ( P ( n ) , q ) + max i I n log ( # Δ ( O P ( n ) | P ( i ) ) )
for all q > 1 and n N . Dividing both sides by n and taking the limit superior n finishes the proof.   □
The following lemma was shown in [22] (see Lemma 2.2.).
Lemma A4.
Let T : Ω Ω be a map on Ω R . Assume that M = { M i } i I is a finite partition of Ω into sets on each of which T is monotone. Then, for all n N and multi-indices i = ( i 0 , i 1 , , i n 1 ) I n
# Δ ( O P ( n ) | M ( i ) ) 2 # { s { 0 , 1 , , n 2 } i s = i n 1 }
holds true.
As a consequence of the two lemmata, we obtain an upper bound of RPE that is important for the following.
Corollary A1.
Let ( Ω , B , μ , T ) be a measure-preserving dynamical system with Ω R . Assume that M is a finite partition of Ω into sets on each of which T is monotone. Then, for all q [ 0 , ]
PE ¯ ( T , q ) h ( T , M , q ) + log 2 .
holds true.
Proof. 
We have
PE ¯ ( T , q ) h ( T , M , q ) + lim sup n 1 n max i I n log ( # Δ ( O P ( n ) | M ( i ) ) )
according to Lemma A3. Lemma A4 provides
# Δ ( O P ( n ) | M ( i ) ) 2 # { s { 0 , 1 , , n 2 } i s = i n 1 } 2 n 1
for all n N and i I n . Combining the above statements yields
PE ¯ ( T , q ) = h ( T , M , q ) + lim sup n ( n 1 ) / n log 2 = h ( T , M , q ) + log 2 .
 □
We use the following fact about the Rényi entropies of stochastic vectors:
Lemma A5.
For all q [ 1 , ] and all c > 0 , there exists a stochastic vector p with
H ( p ) H ( p , q ) c .
Proof. 
Let c > 0 and q > 1 . Choose N N , such that
N 2 · c log ( 2 ) + q q 1
holds true. Consider stochastic vector p = ( p 1 , p 2 , , p 2 N + 1 ) [ 0 , 1 ] 2 N + 1 with
p 1 = 1 2
and
p i = 1 2 N + 1 for i { 2 , 3 , , 2 N + 1 } .
Then,
H ( p ) = i = 1 2 N + 1 p i log ( p i ) = 2 N · 1 2 N + 1 log 1 2 N + 1 1 2 log 1 2 = N + 1 2 log ( 2 ) + 1 2 log ( 2 ) N 2 log ( 2 )
and
H ( p , q ) = 1 q 1 log i = 1 2 N + 1 p i q = 1 q 1 log 2 N · 1 2 ( N + 1 ) q + 1 2 q 1 q 1 log 1 2 q = q q 1 log 2 .
Hence,
H ( p ) H ( p , q ) log ( 2 ) · N 2 q q 1 c .
 □
Recall that a (one-sided) Bernoulli shift σ is a map defined on { 0 , 1 , , N 1 } for some N N by
σ ( ( b n ) n = 1 ) = ( b n ) n = 2
for ( b n ) n = 1 { 0 , 1 , , N 1 } . Given a stochastic vector p = ( p 0 , p 1 , , p N 1 ) [ 0 , 1 ] N , one considers the product measure μ p associated to p . More precisely, μ p is the probability measure on the smallest σ -algebra containing all cylinder sets
C k ( a 1 , a 2 , , a k ) = { ( b n ) n = 1 { 0 , 1 , , N 1 } ( b 1 , b 2 , , b k ) = ( a 1 , a 2 , , a k ) }
for k N ; a 1 , a 2 , , a k { 0 , 1 , , N 1 } defined by
μ p ( C k ( a 1 , a 2 , , a k ) ) = n = 1 k p a n .
for such cylinder sets. σ is μ p -preserving and ergodic.
Lemma A6.
Let σ be a Bernoulli shift on { 0 , 1 , , N 1 } for some N N and p = ( p 0 , p 1 , , p N 1 ) [ 0 , 1 ] N be a stochastic vector. Let μ p be defined as above and, for k N , let
C k = { C k ( a 1 , a 2 , , a k ) a 1 , a 2 , , a k { 0 , 1 , , N 1 } } .
Then,
h ( σ , C k , q ) = H ( p , q )
holds true for all q 1 , k N .
Proof. 
Take k N and q R . Notice that
C k ( n ) = t = 0 n 1 σ 1 ( C k ) = C k + n
for all n N . This implies that
H ( C k ( n ) ) = 1 q 1 log P C k + n μ p ( P ) q = 1 q 1 log a 1 , , a k + n { 0 , 1 , , N 1 } i = 1 k + n p i q = 1 q 1 log a 1 , , a k + n { 0 , 1 , , N 1 } i = 1 k + n p i q = 1 q 1 log a { 0 , 1 , , N 1 } p a q k + n = ( k + n ) · 1 q 1 log a { 0 , 1 , , N 1 } p a q = ( k + n ) · H ( p , q )
for all n N and q 1 . Therefore,
h ( σ , C k , q ) = lim inf n 1 n H i = 0 n 1 σ 1 ( C k ) , q = lim inf n k + n n · H ( p , q ) = H ( p , q )
holds true for all q 1 .  □
Example A1.
Take any q > 1 . Choose some N N and a stochastic vector p = ( p 0 , p 1 , , p N 1 ) [ 0 , 1 ] N such that
H ( p , q ) + log 2 < H ( p )
holds true. This is always possible due to Lemma A5. Further, let μ p be the probability measure defined before Lemma A6.
In order to obtain a map with finitely many monotone parts, identify Bernoulli shift σ on { 0 , 1 , , N 1 } with interval map T on Ω = [ 0 , 1 ] defined by
T ( ω ) = N · ω mod 1
for all ω Ω . This is possible since the correspondence
( b n ) n N { 0 , 1 , , N 1 } n = 1 b n N n
turns σ into T. In this way, measure μ p can be considered as a measure on Ω with the Borel sets. One easily sees that M = { [ ( i 1 ) / N , i / N ] i { 1 , 2 , , N } } is a partition into intervals on each of which T is monotone and continuous. According to the above identification, M can be considered as the partition C 1 in Lemma A6. Now
PE ¯ ( T , q ) h ( T , M , q ) + log 2 = H ( p , q ) + log 2 < H ( p ) = h ( T , q )
holds true. Here the first equality follows from Corollary A1, the last one from Theorem 1, and the first inequality from Lemma A6.

References

  1. Bandt, C.; Pompe, B. Permutation Entropy: A Natural Complexity Measure for Time Series. Phys. Rev. Lett. 2002, 88, 174102. [Google Scholar] [CrossRef] [PubMed]
  2. Zanin, M.; Zunino, L.; Rosso, O.; Papo, D. Permutation Entropy and Its Main Biomedical and Econophysics Applications: A Review. Entropy 2012, 14, 1553. [Google Scholar] [CrossRef]
  3. Amigò, J.M.; Keller, K.; Unakafova, V.A. Ordinal symbolic analysis and its application to biomedical recordings. Philos. Trans. Ser. Math. Phys. Eng. Sci. 2015, 373, 20140091. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Bandt, C.; Keller, G.; Pompe, B. Entropy of interval maps via permutations. Nonlinearity 2002, 15, 1595. [Google Scholar] [CrossRef]
  5. Liang, Z.; Wang, Y.; Sun, X.; Li, D.; Voss, L.J.; Sleigh, J.W.; Hagihira, S.; Li, X. EEG entropy measures in anesthesia. Front. Comput. Neurosci. 2015, 9, 16. [Google Scholar] [CrossRef] [PubMed]
  6. Mammone, N.; Duun-Henriksen, J.; Kjaer, T.W.; Morabito, F.C. Differentiating Interictal and Ictal States in Childhood Absence Epilepsy through Permutation Rényi Entropy. Entropy 2015, 17, 4627–4643. [Google Scholar] [CrossRef] [Green Version]
  7. Zunino, L.; Olivares, F.; Rosso, O.A. Permutation min-entropy: An improved quantifier for unveiling subtle temporal correlations. EPL (Europhys. Lett.) 2015, 109, 10005. [Google Scholar] [CrossRef]
  8. Rivero, C.; Pucheta, J.; Patiño, D.; Laboret, S.; Juárez, G.; Sauchelli, V. A combined approach for long-term series prediction: Renyi permutation entropy with BEA predictor filter. In Proceedings of the 2016 IEEE Biennial Congress of Argentina (ARGENCON), Argentina, South America, 15–17 June 2016; pp. 1–5. [Google Scholar]
  9. Park, H.; Kim, Y.; Kim, S.; Jeon, A.; Kim, S.; Choi, W. Comparison of electroencephalogram between propofol- and thiopental-induced anesthesia for awareness risk in pregnant women. Sci. Rep. 2020, 10, 6192. [Google Scholar] [CrossRef] [PubMed]
  10. Chen, S.; Shang, P.; Wu, Y. Weighted multiscale Rényi permutation entropy of nonlinear time series. Phys. Stat. Mech. Appl. 2018, 496, 548–570. [Google Scholar] [CrossRef]
  11. Zhou, Q.; Shang, P. Weighted multiscale cumulative residual Rényi permutation entropy of financial time series. Phys. Stat. Mech. Appl. 2020, 540, 123089. [Google Scholar] [CrossRef]
  12. Keller, K.; Mangold, T.; Stolz, I.; Werner, J. Permutation Entropy: New Ideas and Challenges. Entropy 2017, 19, 134. [Google Scholar] [CrossRef] [Green Version]
  13. Simpson, E. Measurement of diversity. Nature 1949, 163, 688. [Google Scholar] [CrossRef]
  14. Caballero-Pintado, M.V.; Matilla-García, M.; Marín, M.R. Symbolic correlation integral. Econom. Rev. 2019, 38, 533–556. [Google Scholar] [CrossRef]
  15. Takens, F.; Verbitski, E. Generalized entropies: Rényi and correlation integral approach. Nonlinearity 1998, 11, 771–782. [Google Scholar] [CrossRef]
  16. Walters, P. An introduction to ergodic theory. In Graduate Texts in Mathematics; Springer: New York, NY, USA, 1982; Volume 79. [Google Scholar]
  17. Verbitskiy, E. Generalized Entropies in Dynamical Systems. Ph.D. Thesis, University of Groningen, Groningen, The Netherlands, 2000. [Google Scholar]
  18. Takens, F.; Verbitskiy, E. Rényi entropies of aperiodic dynamical systems. Isr. J. Math. 2002, 127, 279–302. [Google Scholar] [CrossRef] [Green Version]
  19. Gutjahr, T.; Keller, K. Ordinal Pattern Based Entropies and the Kolmogorov–Sinai Entropy: An Update. Entropy 2020, 22, 63. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Misiurewicz, M.; Szlenk, W. Entropy of piecewise monotone mappings. Stud. Math. 1980, 67, 45–63. [Google Scholar] [CrossRef]
  21. Gutjahr, T. Determining the Kolmogorov-Sinai Entropy by Using Ordinal Patterns. Ph.D. Thesis, University of Lübeck, Lübeck, Germany, 2021. [Google Scholar]
  22. Tim Gutjahr, K.K. Equality of Kolmogorov-Sinai and permutation entropy for one-dimensional maps consisting of countably many monotone parts. Discret. Contin. Dyn. Syst. 2019, 39, 4207–4224. [Google Scholar] [CrossRef] [Green Version]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Gutjahr, T.; Keller, K. On Rényi Permutation Entropy. Entropy 2022, 24, 37. https://doi.org/10.3390/e24010037

AMA Style

Gutjahr T, Keller K. On Rényi Permutation Entropy. Entropy. 2022; 24(1):37. https://doi.org/10.3390/e24010037

Chicago/Turabian Style

Gutjahr, Tim, and Karsten Keller. 2022. "On Rényi Permutation Entropy" Entropy 24, no. 1: 37. https://doi.org/10.3390/e24010037

APA Style

Gutjahr, T., & Keller, K. (2022). On Rényi Permutation Entropy. Entropy, 24(1), 37. https://doi.org/10.3390/e24010037

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop