Next Article in Journal
Bearing Fault Diagnosis Based on Multiscale Permutation Entropy and Support Vector Machine
Next Article in Special Issue
Infrared Cloaking, Stealth, and the Second Law of Thermodynamics
Previous Article in Journal
A New Entropy Optimization Model for Graduation of Data in Survival Analysis
Previous Article in Special Issue
Association of Finite-Dimension Thermodynamics and a Bond-Graph Approach for Modeling an Irreversible Heat Engine
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Equivalence of Partition Functions Leads to Classification of Entropies and Means

by
Michel S. Elnaggar
1,† and
Achim Kempf
2,*
1
Department of Applied Mathematics, University of Waterloo, Waterloo, Ontario N2L 3G1, Canada
2
Department of Applied Mathematics, University of Waterloo, 200 University Avenue West, Waterloo, Ontario N2L 3G1, Canada
*
Author to whom correspondence should be addressed.
Present address: Bell Mobility Inc., Mississauga, Ontario L4W 5N2, Canada.
Entropy 2012, 14(8), 1317-1342; https://doi.org/10.3390/e14081317
Submission received: 24 April 2012 / Revised: 26 June 2012 / Accepted: 9 July 2012 / Published: 27 July 2012
(This article belongs to the Special Issue Advances in Applied Thermodynamics)

Abstract

:
We derive a two-parameter family of generalized entropies, Spq, and means mpq. To this end, assume that we want to calculate an entropy and a mean for n non-negative real numbers {x1,,xn}. For comparison, we consider {m1,,mk} where mi = m for all i = 1,,k and where m and k are chosen such that the lp and lq norms of {x1,,xn} and {m1,,mk} coincide. We formally allow k to be real. Then, we define k, log k, and m to be a generalized cardinality kpq, a generalized entropy Spq, and a generalized mean mpq respectively. We show that this family of entropies includes the Shannon and Rényi entropies and that the family of generalized means includes the power means (such as arithmetic, harmonic, geometric, root-mean-square, maximum, and minimum) as well as novel means of Shannon-like and Rényi-like forms. A thermodynamic interpretation arises from the fact that the lp norm is closely related to the partition function at inverse temperature β = p. Namely, two systems possess the same generalized entropy and generalized mean energy if and only if their partition functions agree at two temperatures, which is also equivalent to the condition that their Helmholtz free energies agree at these two temperatures.

1. Introduction

Two of the most basic concepts of thermodynamics are: (a) the average of measurement outcomes and (b) the uncertainty or entropy about measurement outcomes. Consider, for example, a physical system, A, that is in contact with a heat bath at some fixed temperature, i.e., a canonical ensemble. The measurement of the system’s energy can return any one of its energy eigenstates. What then is (a) the mean energy to expect and (b) how uncertain is the prediction of the measured energy state?
We notice that, in principle, many different notions of energy mean and many different measures of entropy could be employed here. Of course, in thermodynamics, the Boltzmann factor weighted mean as well as the Shannon/von Neumann entropy are of foremost importance. In this paper, we show that also other important notions of average such as the harmonic mean, the geometric mean and the arithmetic mean arise naturally, along with generalized notions of entropy including Rényi entropies [1], all unified in a two-parameter family of notions of means and notions of entropies.
To this end, consider systems (canonical ensembles) in a heat bath. We begin by considering the simplest kind of system, namely the type of system which possesses only one energy level, E. Let us denote its degeneracy by k. Unambiguously, we should assign that system the mean m : = E and the entropy S : = log k . Let us denote these simple one-level systems by the term reference system.
Now, let X be a system with arbitrary discrete energy levels. Our aim is to assign X a mean and an entropy by finding that reference system M which is in some sense equivalent to X. Then we assign X the same value for the mean and entropy as the reference system M.
But how do we decide if a reference system is in some sense equivalent to system X? Given that we want the reference system M and system X to share two properties, namely a mean and an entropy, we expect any such condition for the equivalence of two systems to require two equations to be fulfilled. Further, since the properties of systems are encoded in their partition function Z ( X ) , we expect that these two equations can be expressed in terms of the partition functions of the two systems in question.
To this end, let us adopt what may be considered the simplest definition. We choose two temperatures, T1 and T2 and we define that a reference system is (T1, T2)-equivalent to system X if the partition functions of the two systems coincide with each other at these two temperatures. Since the Helmholtz free energy A H obeys A H = K B T log ( Z ) , where K B is the Boltzmann constant, this is the same as saying that two systems are put in the same equivalence class if their Helmholtz free energies coincide at these two temperatures.
This allows us now to assign any system X a mean and an entropy. We simply find its unique (T1 and T2)-equivalent reference system M. Then the mean and entropy of X are defined to be the mean and the entropy of the reference system M.
Clearly, the so-defined mean and entropy of a system X now actually depend on two temperatures, namely (T1, T2). As we will show below, in the limit when we let the two temperatures become the same temperature, we recover the usual Boltzmann factor-weighted mean, i.e., the usual mean energy, along with the usual Shannon/von Neumann entropy.
For general (T1, T2), we cover more, however. Namely, we naturally obtain a unifying 2-parameter family of notions of mean that includes for example the geometric, the harmonic, the arithmetic and the root-mean-square (RMS) means. And we obtain a unifying 2-parameter family of notions of entropy that, for example, includes the Rényi family of entropies.
To be precise, let us assume that a system X has only discrete energy levels, {Ei}, where i enumerates all the energy levels counting also possible degeneracies. Notice that {Ei} is formally what is called a multiset, because its members are allowed to occur more than once. Similarly, let us also collect the exponentiation of the negative energies x i : = exp ( Ε i ) in the multiset { x i } . Either multiset can be used to describe the same thermodynamic system X. Let β : = 1 / K B T denote the inverse temperature, where K B is the Boltzmann constant. The partition function of system X, i.e., the sum of its Boltzmann factors, then reads:
Z β ( X ) = i e β Ε i = i x i β
For later reference, note that the partition function is therefore related to the lp norm of X = {x1,,xn} for p = β through Z β ( X ) = ​​​   | | X | | β β .
Now the key definition is that we call two physical systems ( β 1 , β 2 ) -equivalent if their partition functions coincide at the two inverse temperatures ( β 1 , β 2 ) , i.e., systems X and M are ( β 1 , β 2 ) -equivalent if Z β 1 ( X ) = Z β 1 ( M ) and Z β 2 ( X ) = Z β 2 ( M ) . To be more explicit, one may also call such systems ( β 1 , β 2 ) -partition function equivalent, or also ( β 1 , β 2 ) -Helmholtz free energy equivalent, but we will here use the term ( β 1 , β 2 ) -equivalent for short.
In particular, for any given system X, let us consider the ( β 1 , β 2 ) -equivalent reference system M which possesses just one energy level, with energy E0 and degeneracy k, where we formally allow k to be any positive number. E0 and k are then determined by the two conditions that the partition function of M is to coincide with that of X at the two inverse temperatures β 1 and β 2 . Then, we define S β 1 , β 2 ( X ) : = log k to be the generalized entropy, and m β 1 β 2 ( X ) : = Ε 0 to be the generalized mean energy of system X with respect to the temperatures ( β 1 , β 2 ) .
We will explore the properties of these families of generalized entropies and means in the subsequent sections. First, however, let us consider the special limiting case when the two temperatures coincide (i.e., β 1 = β 2 = β ). As will be detailed in the subsequent sections of the manuscript, in this limiting case, the two equivalence conditions of partition functions can be shown to reduce to:
( Z β ( X ) ) 1 / β = ( Z β ( M ) ) 1 / β
β ( Z β ( X ) ) 1 / β = β ( Z β ( M ) ) 1 / β
which can be shown to be equivalent to:
Z β ( X ) = Z β ( M )
i Ε i e β Ε i j e β Ε j = Ε 0
The conditions (3a) and (3b) physically mean that systems X and M have the same partition function and average energy, respectively, at the inverse temperature β . Notice that this is also the same as saying that the two systems have the same average energy and the same Helmholtz free energy at the inverse temperature β . Now by employing either pair of conditions, (2) or (3), we then recover indeed the usual thermodynamic entropy of the system X, which is given in the Shannon form at the inverse temperature β by
S β β ( X ) = i e β Ε i j e β Ε j log ( e β Ε i j e β Ε j )
The proofs of Equations (2)–(4) are straightforward. We will spell them out in detail in the subsequent sections where the setting is abstract mathematical.
Before we begin the mathematical treatment, let us remark that entropy is not only a cornerstone of thermodynamics but it is also crucial in information theory. Due to its universal significance, measures of uncertainty in the form of an entropy have been proposed by physicists and mathematicians for over a century [2]. Our approach here for deriving a generalized family of entropies was originally motivated by basic questions regarding the effective dimensionality of multi-antenna systems (e.g., [3,4,5]). After initial attempts in [6] and later in [5,7], we here give for the first time a comprehensive derivation with the proofs and we also include the family of generalized means.
The manuscript is organized as follows. In Section 2, we introduce the proposed family of entropies and means mathematically, and also show some special cases thereof. The axiomatic formulation is presented in Section 3, followed by the study of resulting properties in Section 4. Proofs are provided in the appendices.

2. Mathematical Definition of Generalized Entropies and Means

Let X = { x i : i = 1 , 2 , , K X } be a multiset of real non-negative numbers, where K X denotes the cardinality of X. We assume that X possesses at least one non-zero element. Further, let p , q be arbitrary fixed real numbers obeying p q 0 . Let M = { m i : m i = m , m > 0 , i = 1 , 2 , , k } be a reference multiset possessing exactly one real positive element m , which is of multiplicity k 1 . We introduce the following definitions:
  • N X is the number of non-zero elements of X, therefore N X K X . To simplify notation, the subscript X is omitted when dealing with one multiset at hand.
  • K X max is the multiplicity of the maximum elements of X
  • K X min is the multiplicity of the minimum elements of X
Our objective is to determine suitable values for m and k , possibly non-integer, that can serve as mean and effective cardinality of X, respectively, namely by imposing a suitable criterion for the equivalence of X to a reference multiset M. Having two unknowns ( m and k ) in M, we need two equivalence conditions. We choose to impose the equivalence of the p-norms and the q-norms:
X p M p = m k 1 / p X q M q = m k 1 / q
Here, the p-norm p is defined as usual through:
X p : = ( i = 1 K X ( x i ) p ) 1 / p
with the proviso that ( x i ) p is replaced by 0 if x i = 0 and p 0 . We remark that, for p < 1 , (6) is merely a quasi-norm since the triangle inequality does not hold. Note the singularity lim p 0 + X p = .
Solving for k and m in (5), we obtain:
k p q ( X ) = [ X p / X q ] 1 / ( 1 p 1 q )         p q m p q ( X ) = [ X p p / X q q    ] 1 p q             p q
We call k p q ( X ) and m p q ( X ) the norm-induced effective cardinality and generic mean of order p , q for the multiset X, respectively. Let us now express (7) in a logarithmic form and define the entropy S p q ( X ) as follows:
S p q ( X ) : = log k p q ( X ) = log X p log X q 1 p 1 q                      p q
log m p q ( X ) = p log X p q log X q p q             p q
Notice that S p q ( X ) = S q p ( X ) and m p q ( X ) = m q p ( X ) , i.e., both the entropy and mean are symmetrical with respect to the order p , q .
Next, we express S p q ( X ) and m p q ( X ) in the limiting case when p q . For S q q ( X ) we find:
S q q ( X ) : = log k q q ( X )                   = lim p q S p q ( X )                   = lim p q log X p log X q 1 p 1 q                = l ' H o ^ pital q 2 q X q X q                    = i = 1 K x i q j = 1 K x j q log ( x i q j = 1 K x j q )
where the last step is obtained by straightforward manipulations. Similarly, we find for m q q ( X ) :
m q q ( X ) = lim p q   m p q ( X ) = X q / k q q 1 / q log m q q ( X ) = lim p q p log X p q log X q p q                      = l ' H o ^ pital log X q + q q X q X q                         = q X q q X q q                         = i = 1 K x i q j = 1 K x j q log x i
where in the second to last step, we used the fact that d d x g ( x ) f ( x ) = g f { d f d x log g + f g d g d x } , and the last step is obtained by straightforward manipulations.
It is worthwhile to mention the following useful relation linking S q q ( X ) , X q and m q q ( X ) , which is readily deduced from (9) and (10):
log m q q ( X ) = log X q S q q ( X ) / q
We remark that in the early phase of this work [5], each author independently suggested either k p q ( X ) or k q q ( X ) as two possible distinct notions for the effective cardinality. In [7], it was reported that the average energy and Shannon entropy of a thermodynamic system are obtained by starting from equivalence of partition functions of two systems at two temperatures when the two temperatures coincide as mentioned in the introduction. Clearly, the limiting operation in (9) makes the connection and establishes (7) as the general definition of this norm-induced family of entropies and means.
In fact, for the case of degenerate order, ( p q ), the quantities k q q ( X ) , S q q ( X ) , and m q q ( X ) , could have been obtained as well through a differential equivalence of the q -norm. To see this, we impose the following two conditions:
        X q M q q X q q M q
After employing d d x g ( x ) f ( x ) = g f { d f d x log g + f g d g d x } and solving for k and m , we ultimately obtain k q q ( X ) and m q q ( X ) as given by (9) and (10). The condition (12) is the mathematical equivalent of the aforementioned physical condition (2) imposed on the two thermodynamic systems, which yielded the Shannon entropy form (4).
From (9), it is obvious that S q q ( X ) is the Shannon entropy of the distribution { x i q / i = 1 K x i q } , which is called the escort distribution of order q [8] of { x i / i = 1 K x i } . On the other hand, S p q ( X ) is a more general expression of the Rényi entropy of order α . For a probability distribution P = { p i } , the Rényi entropy of order α is given by [9]:
S α ( R ) ( P ) : = 1 1 α log i p i α i p i ,       α > 0 , α 1
By setting the order p = 1 in S p q , we obtain from (8):
S 1 q ( X ) = log i = 1 K x i q q log j = 1 K x j 1 q            q > 0 , q 1                   = 1 1 q log i = 1 K ( x i j = 1 K x j ) q
By comparing (13) and (14), we readily identify S 1 q ( X ) as the Rényi entropy of order q for a complete statistical distribution given by { x i / i = 1 K x i } , where the multiset elements add to 1. Formally:
S 1 q ( X ) = S q ( R ) ( { x i / i = 1 K x i } )
In the degenerate case (when q 1 ), S 11 ( X ) is the Shannon entropy of the latter distribution. For p 1 , S p q ( X ) from (8) can be rearranged as a generalization of (13):
S p q ( X ) = p p q log i = 1 K ( x i p j = 1 K x j p ) q / p                 = 1 1 q / p log i = 1 K ( x i p j = 1 K x j p ) q / p                 = S q / p ( R ) ( { x i p / i = 1 K x i p } )
which can be viewed as the Rényi entropy of order q / p for the p t h order escort distribution { x i p / i = 1 K x i p } .
Rényi defined his entropy for p , q 0 . We relax this condition further and allow S p q ( X ) and m p q ( X ) to be defined for any real indices p , q such that p q 0 . Accordingly, we obtain the following properties:
S p , q ( { x i } ) = S p , q ( { x i 1 : x i 0 } ) m p , q ( { x i } ) = m p , q 1 ( { x i 1 : x i 0 } )
When at least one order p , q is zero, we find the interesting results:
k 0 q ( X ) = N X
S 0 q ( X ) = log N X
m 0 q ( X ) = { ( i = 1 N X x i q N X ) 1 / q x i 0 , q 0 ( i = 1 N X x i ) 1 / N X x i 0 , q 0
We recognize S 0 q ( X ) in (19) as the Hartley entropy [10], which will be shown later to be the maximum value of any entropy. From (20), we obtain a famous family of generic p-means of the non-zero elements of X: particularly m 0 , , m 0 , 1 , m 0 , 1 , m 0 , 2 , m 0 , are the minimum, harmonic mean, arithmetic mean, root-mean-square mean, and maximum, respectively. In the limiting case q 0 , we obtain m 0 , 0 , which is the geometric mean. In Table 1, we summarize these and other particular cases of means and entropies at specific p , q . The key point is that each p , q uniquely defines an entropy S p q with a corresponding mean m p q , such that each pair of S p q and m p q is coupled in this sense.
Table 1. Special cases of S p q and m p q . Note that S p q = S q p and m p q = m q p (Property 4.3).
Table 1. Special cases of S p q and m p q . Note that S p q = S q p and m p q = m q p (Property 4.3).
Order p , q S p q ( X ) S p q name m p q ( X ) m p q name
0 , q     q 0 log N X Boltzmann-Hartley entropy ( i = 1 N X x i q N X ) 1 / q Generic q mean. Specific q values are harmonic (-1), arithmetic (1), root-mean-square (2), maximum ( ), minimum ( )
0,0 log N X Boltzmann-Hartley entropy ( i = 1 N X x i ) 1 / N X Geometric mean
, log K X max max ( X ) maximum
, q    q > 0 log ( K X max + x i max ( x i ) ( x i max ( x i ) ) q ) max ( X ) maximum
, log K X min min ( X ) minimum
, q    q < 0 log ( K X min + x i min ( x i ) ( x i min ( x i ) ) q ) min ( X ) minimum
1 , q    q > 0 1 1 q log i = 1 K ( x i j = 1 K x j ) q Rényi entropy, order q , of the complete distribution { x i / i = 1 K x i } i = 1 K x i exp S 1 q ( X ) “Rényi-like” mean
1,1 i = 1 K x i j = 1 K x j log ( x i j = 1 K x j ) Gibbs-Shannon entropy i = 1 K x i exp S 11 ( X ) “Shannon-like” mean
A typical plot for X p , S p p and m p p is shown in log scale in Figure 1, illustrating some of the properties to be discussed hereafter. In particular, we notice:
  • log X p has a two-sided singularity at p = 0 .
  • S p p is non-decreasing/non-increasing for negative/positive p , respectively, and is guaranteed to be maximized at p = 0 . This is discussed more generally in Property 4.6.
  • m p p ranges from min ( X ) to max ( X ) and is always non-decreasing with respect to p . This is discussed more generally in Property 4.6.
  • p = 1 has a specific property of making m p p k p p = X p .
Figure 1. Typical plot for X p , S p p and m p p in log scale.
Figure 1. Typical plot for X p , S p p and m p p in log scale.
Entropy 14 01317 g001

3. An Axiomatic Approach to the Generalized Entropies and Means

In order to simplify the conceptual underpinnings, let us now describe the generalized entropies S p q and means m p q , through two simple axioms.

3.1. Axioms for the Generalized Entropy

Let p , q be fixed real numbers obeying p q 0 . Consider a map, S p q , which maps multisets of positive real numbers into the real numbers. We call S p q a generalized entropy of order p , q if it obeys the following two axioms:
Entropy Axiom 1: 
S p q ( M ) of a uniform multiset M = { m 0 , m 0 , , m 0 } , m 0 > 0 , with multiplicity k , equals log k , (where the base of the logarithm is arbitrarily chosen), i.e.,
M = { m i : m i = m 0 , m 0 > 0 , i = 1 , 2 , , k } S p q ( M ) = log k
Entropy Axiom 2: 
If p ​​​   q , the map S p q depends only on the ratio of the multiset’s p and q norms, i.e., S p q is some function, f p q , of this ratio:
S p q ( X ) : = f p q ( X p X q )                         p q
If p = q , the map S q q depends only on the ratio of the multiset’s q norm to its derivative, i.e., S p q is some function, f q , of this ratio:
S q q ( X ) : = f q ( q X q X q )              p q
To see that (22b) arises in the limit from (22a) we notice that, since the logarithm is strictly monotone, (22a) is equivalent to saying that S p q is some function h p q of some finite number times the logarithm of the ratio of norms:
S p q ( X )     = h p q ( 1 p q log X p X q )                    = h p q ( log X p log X q p q )
Choosing p = q + ε and taking the limit ε 0 we obtain (22b) with f q = h q q .
Proposition: The two entropy axioms in (21) and (22) uniquely define S p q ( X ) and S q q ( X ) , namely as given in section 2 by (8) and (9).
Proof: 
Entropy Axiom 2 implies that for any two multisets X and Y :
X p X q = Y p Y q                            S p q ( X ) = S p q ( Y )           f o r     p q
q X q X q = q Y q Y q       S q q ( X ) = S q q ( Y )            f o r     p = q
Choosing for Y the uniform multiset M of Axiom 1, and taking the logarithm of both sides yields:
log X p X q = ( 1 p 1 q ) log k            p q q X q X q = log k q 2                       p = q
Using that S p q ( X ) = S p q ( Y ) = S p q ( M ) = log k , we now uniquely obtain the formulas for S p q ( X ) and S q q ( X ) given in (8) and (9), respectively. We note that the functions f p q and f q are therefore:
f p q ( · ) = log ( · ) 1 p 1 q            p q f q ( · ) = q 2 ( · )             p = q
Remarks:
  • Even though k is treated as an integer representing the multiplicity in Axiom 1, this condition is tacitly relaxed in Axiom 2 to include non-integer values, which we may call the effective cardinality (or effective dimensionality) of order p , q .
  • The logarithmic measure of Axiom 1 is directly connected to the celebrated Boltzmann entropy formula S ( B ) = K B log W , where K B is the Boltzmann constant and W is the number of the microstates in the system. The logarithmic measure is also connected to the so-called “Hartley’s measure” [11,12], which indicates the non-specificity [13] and does not require a probability distribution assumption. In Axiom 1, a multiset of equal positive numbers is all that is required. In fact, Axiom 1 encompasses the additivity and monotonicity axioms [12,13], which are equivalent to the earlier Khinchin’s axioms of additivity, maximality and expansibility [14].
  • In Axiom 2, note that the p-norm definition is relaxed to include the values p < 1 , which would result in the triangle inequality to be violated should the multisets be treated as vectors.

3.2. Axioms for the Generalized Mean

We define the pth moment of the multiset X = { x i } as:
p t h m o m e n t ( X ) : = X p p = i x i p
The nomenclature “pth moment” is motivated by the fact that for the density function r ( x ) : = i δ ( x x i ) , where δ ( x ) is the Dirac delta function, the pth moment is indeed:
p t h m o m e n t ( X ) : = x p r ( x ) d x                                     = x p i δ ( x x i ) d x                                     = i x i p = X p p
Let p , q be fixed real numbers obeying p q 0 . Consider a map, m p q , which maps multisets of positive real numbers into the real numbers. We call m p q a generalized mean of order p , q if it obeys the following two axioms:
Mean Axiom 1: 
m p q ( M ) of a uniform multiset M = { m 0 , m 0 , , m 0 } , m 0 > 0 is m 0 . Formally:
M = { m i : m i = m 0 , m 0 > 0 , i = 1 , 2 , , k } m ( M ) : = m 0
Mean Axiom 2: 
If p ​​​   q , the map m p q depends for any multiset X only on the ratio of the multiset’s pth and qth moments, | | X | | p p and | | X | | q q , i.e., m p q is some function, g p q , of their ratio:
m p q ( X ) : = g p q ( X p p X q q )                         p q
If p = q , the map m q q is a function only of the ratio:
m q q ( X ) : = g ( q X q q X q q )              p = q
The fact that (30b) is the limit of (30a) follows by the same reasoning as in (23).
Proposition: 
The two mean axioms in (29) and (30) uniquely define m p q , namely as given in (7) and (10).
Proof: 
Axiom 2 implies for any two multisets X and Y that:
X p p X q q = Y p p Y q q                            m p q ( X ) : = m p q ( Y )            p q q X q q X q q = q Y q q Y q q       m q q ( X ) : = m q q ( Y )            p = q
Choosing the multiset Y to be the uniform multiset M from (29) we obtain
X p p X q q = m 0 p q                          p q q X q q X q q = log m 0            p = q
We can now use that m ( X ) = m ( M ) = m 0 , to obtain m p q ( X ) and m q q ( X ) as given in (7) and (10), respectively. Accordingly the functions g p q and g are found to be
g p q ( · ) = ( · ) 1 p q            p q g ( · ) = exp ( · )              p = q
We have obtained axiomatizations of the generalized entropies and means which revealed, in particular, that the generalized entropies can be characterized as those entropies that cover the reference multiset case (the multiset of equal elements) and that are functions of only the ratio of the multisets’ l p and l q norms. Similarly, the axiomatization also revealed that the generalized means can be characterized as those means which cover the reference multiset case and which are functions of only the ratio of the multisets’ pth and qth moments | | X | | p p and | | X | | q q . We will now develop an axiomatization that links up with Section 2, yielding simultaneously a unique family of generalized entropies and means.

3.3. Unifying Axioms for Generalized Entropies and Means

We notice that, as is straightforward to verify:
X p X q = Y p Y q  and  X p p X q q = Y p p Y q q          X p = Y p  and  X q = Y q
This means that we can describe the generalized entropies and means also through a unifying set of axioms. To this end, let p , q be fixed real numbers obeying p q 0 . Consider maps S p q and m p q , which map multisets of positive real numbers into the real numbers. We call S p q and m p q generalized entropies and means of order p , q respectively, if they obey the following two axioms:
Unifying Axiom 1: 
S p q and m p q applied to a multiset of k equal elements M = { m 0 , m 0 , , m 0 } , m 0 > 0 yield the values log k and m 0 , respectively.
Unifying axiom 2: 
X p = Y p     a n d     X q = Y q                             S p q ( X ) : = S p q ( Y )     a n d     m p q ( X ) : = m p q ( Y )       p q X q = Y q     a n d     q X q = q Y q         S q q ( X ) : = S q q ( Y )     a n d     m q q ( X ) : = m q q ( Y )       p = q
Proposition: 
The maps S p q and m p q are unique and given by Equations (7)–(10).
Proof: 
The proofs are straightforward and proceed similarly to the proofs of the propositions related to the entropy and mean axioms.

4. Properties of S p q and m p q

We list in this section useful properties of m p q , S p q and k p q with proofs in the appendix. The definitions of m p q , S p q and k p q are given by (7)–(10). We also add two plots for an example multiset X = { 10 ; 9 ; 8 ; 7 ; 6 ; 0.5 ; 0.4 ; 0.3 ; 0.2 ; 0.1 } in Figure 2 and Figure 3 in order to provide some numerical illustration of the properties hereunder.
Figure 2. Numerical example showing the multiset elements { x i } versus their index i ; with their corresponding mean m p q and effective cardinality k p q for different values of p , q .
Figure 2. Numerical example showing the multiset elements { x i } versus their index i ; with their corresponding mean m p q and effective cardinality k p q for different values of p , q .
Entropy 14 01317 g002
Figure 3. Numerical example for the multiset X showing its mean m p q and effective cardinality k p q plotted versus q for fixed values of p .
Figure 3. Numerical example for the multiset X showing its mean m p q and effective cardinality k p q plotted versus q for fixed values of p .
Entropy 14 01317 g003

4.1. Scaling

Given a multiset X = { x i : i = 1 , 2 , , K } and a constant γ > 0 , we have:
S p q ( { γ x i : i = 1 , 2 , , K } ) = S p q ( X ) m p q ( { γ x i : i = 1 , 2 , , K } ) = γ m p q ( X )
That is the entropy is invariant of the scaling, whereas the mean varies linearly with scaling. The proof is straightforward, based on (7).

4.2. Symmetry with Respect to the Elements of X

S p q ( X ) does not depend on the order of the elements of X .

4.3. Symmetry with Respect to the Order p , q

By exchanging p and q in (8), we readily find that:
S p q ( X ) = S q p ( X )

4.4. Sign Change of the Order p , q

From (17), for p q 0 , we obtain:
S p , q ( { x i } ) = S p , q ( { x i 1 : x i 0 } ) m p , q ( { x i } ) = m p , q 1 ( { x i 1 : x i 0 } )

4.5. Range of S p q ( X )

Let N X 1 be the number of the non-zero elements in X . Therefore:
0 S p q ( X ) log N X            p q 0
The minimum value occurs when X has exactly one non-zero element. For p , q 0 , the maximum value occurs when all the non-zero elements of X are equal. When either p or q is zero, S p q ( X ) yields the maximum value, log N X , for any distribution of X , which is physically intuitive since the zeroth order renders all the non-zero multiset elements to an equal value and thus we reach the equiprobable case leading to maximum entropy. The proof of this property is in Appendix A.3.

4.6. Monotonicity of S p q ( X ) and m p q ( X ) with respect to p , q

We have the following results for the monotonicity of S p q ( X ) and m p q ( X ) with respect to p , q . The proofs are in Appendix B and Appendix C, respectively:
q p S p q 0        p q ,     p q 0 p p S p p 0        p q
p m p q 0            p q ,    p q 0 p m p p 0            p q
where equality holds when all the non-zero elements x i are equal. Accordingly, for p q and p q 0 , by fixing one order (say q ), m p q is always non-decreasing with respect to the other order ( p ); whereas S p q is non-decreasing for p < 0 , non-increasing for p > 0 , with a maximum value at p = 0 . The result is true when switching p , q from the symmetry Property 4.3.
Similarly, for the degenerate case p q , m p p is always non-decreasing with respect to p ; whereas S p p is non-decreasing for p < 0 , non-increasing for p > 0 , with a maximum value at p = 0 . In all cases, both the mean and entropy are invariant with respect to the order p , q if and only if all the non-zero elements x i are equal.
This property explains the monotonicity of the curves in Figure 1 and Figure 3.

4.7. Range of m p q ( X )

From Property 4.6, m p q ( X ) is non-decreasing with respect to p , q . From Table 1, we know that m p , ( X ) = max ( X ) and m p , ( X ) = min ( X ) . Therefore, min ( X ) m p q ( X ) max ( X ) , which is an intuitive range for a mean. This is also true in the degenerate case p q , i.e.,
min ( X ) m p p ( X ) max ( X )
where, as usual, equality holds when all the non-zero elements x i are equal.

4.8. Additivity of the Joint Multiset Entropy

For the two probability distribution multisets X = { x i : i = 1 , 2 , , K X } and Y = { y i : i = 1 , 2 , , K Y } , we define the joint multiset X Y : = { x i y j : i = 1 , 2 , , K X j = 1 , 2 , , K Y } . Therefore, we have:
S ( X Y ) = S ( X ) + S ( Y ) m ( X Y ) = m ( X ) m ( Y )
The proof is straightforward by using the fact that X Y p = X p Y p along with (7)–(10), Property 4.8 is true for both p q and for the degenerate case p q .

4.9. Sub-Additivity of the Effective Cardinality Subject to the Multiset Additive Union Operation

Let denote a multiset additive union operation [15] (page 50), e.g., { 2 , 2 } { 1 , 1 , 2 } = { 1 , 1 , 2 , 2 , 2 } . Let X = { x i : i = 1 , 2 , . K X } and Y = { y j : j = 1 , 2 , . K Y } be two multisets of non-negative real numbers. Moreover, let ξ and η be two positive real scaling factors of the elements of X and Y , respectively. Then:
k p q ( { ξ x i } { η y j } ) k p q ( X ) + k p q ( Y )
where the equality holds under the following condition for the value ξ / η :
ξ 0 η 0 = { ( j y j q / j y j p i x i q / i x i p ) 1 q p p q Y q X q ( k q q ( X ) k q q ( Y ) ) 1 / q p q
This property is a generalization of the effective alphabet size of two disjoint alphabets mixture as discussed in [16] (Problem 2.10). To see this, set p = q = 1 and note that X 1 = Y 1 = 1 for a complete probability distribution. Accordingly, k p q ( X ) represents the effective alphabet size of X corresponding to the entropy S p q ( X ) . The proof is in Appendix D.

4.10. Effective Rank of a Matrix

Let X be the multiset of the singular values of a matrix M . Then, k 0 q ( X ) = rank ( M ) . Accordingly, for general p , q , k p q ( X ) can be viewed as a biased effective rank of M corresponding to the order p , q . From Property 4.5, we have 1 k p q ( X ) rank ( M ) . The minimum value occurs when M has exactly one non-zero singular value (the rank of M is 1). The maximum value, rank ( M ) , is reached for any p q > 0 when all the non-zero singular values are equal. The effective rank can be helpful to determine, in a well-defined p , q sense, how to view an ill-conditioned matrix, which is a full-rank from a mathematical perspective, but is effectively behaving as if possessing a lower rank. Such ill-conditioned matrices often arise in problems of oversampling or determining the degrees of freedom, where the singular values, ordered in non-increasing order by definition, exhibit some sort of “knee cut-off”, similar to { x i } in Figure 2. A biased effective rank can help to compare matrices when the knee cut-off is not sharp, thus giving more weight to the small or large singular values according to the order p , q in a consistent manner for different matrices. An example thereof is the evaluation of the degrees of freedom of some applications such as multi-antenna systems [3,4,5,6], optical imaging systems [20], or in general any case of similar limitation to space-bandwidth product [21].

4.11. Geometrical Interpretation of S p q and m p q on Log Scale, with Thermodynamics Analogy

For the multiset X = { x i } , after taking the logarithm of both sides of (5), we obtain a simple relation between X q , S p q and m p q as in (11):
log X q q = log M q q = S p q + q log m p q log X p p = log M p p = S p q + p log m p q
Accordingly, a secant cutting the function f ( q ) = log X q q at q = p 0 and q = q 0 will have a slope and intercept of log m p 0 q 0 and S p 0 q 0 , respectively, as shown in Figure 4. Based on the discussion of Section 1, (46) readily yields the following analogous expressions for a thermodynamics system described by the Boltzmann factors { e Ε i } :
log Z β 1 ( X ) = S β 1 , β 2 β 1 E 0 log Z β 2 ( X ) = S β 1 , β 2 β 2 E 0
In the limiting case, when p q , β 1 β 2 , the secant in Figure 4 becomes a tangent and we get the Gibbs–Shannon entropy at the inverse temperature β . Consequently, (47) can be re-written after introducing the Boltzmann constant K B and using the absolute temperature T = 1 / K B β as:
K B T log Z β : = A H                               = E 0 T S β
where A H is the Helmholtz free energy of the system.
Figure 4. A secant of log X q q versus q .
Figure 4. A secant of log X q q versus q .
Entropy 14 01317 g004

5. Discussion and Conclusions

A two-parameter family of cardinalities, entropies and means has been derived for multisets of non-negative elements. Rather than starting from thermodynamic or information theoretic considerations to derive entropies and means, see e.g., [9,17,18,19], we here defined the generalized entropies and means through simple abstract axioms. There are other families of entropies in the literature (e.g., [23]), which are a generalization of Shannon entropy. The generalized entropy in this manuscript is shown to preserve the additivity (Property 4.8), which is not the case with the generalized entropies based on Tsallis non-additive entropy as in [23].
Our first two axiomatizations treat the generalized entropies and means separately. It revealed that the generalized entropies are exactly those entropies that are functions of only the ratio of the multisets’ l p and l q norms. It also revealed that the generalized means are exactly those means that are functions only of the ratio of the multisets’ pth and qth moments, | | X | | p p and | | X | | q q . Subsequently, our unifying axiomatization characterized the generalized entropies and means together. This showed that if two multisets have exactly the same l p and l q norms, then they share the same generalized entropy and mean.
We presented several key features of the new families of generalized entropies and means, for example, that the family of generalized entropies contains and generalizes the Rényi family of entropies, of which the Shannon entropy is a special case, thus including some of the desiderata for entropies [22]. We also showed the monotonicity with respect to p , q , extreme values, symmetry with respect to p , q , and additivity preservation. The effective cardinality k p q measures the distribution uniformity of the multiset elements in the sense of the p- and q-norm equivalence to a reference flat multiset. From an information theory perspective, S p q and k p q represent a two-parameter entropy of order p , q and its corresponding effective alphabet size, respectively, when a probability distribution is constructed after proper normalization of the multiset elements. Furthermore, we recall that knowing the p and q norms of a multiset is to know the multiset’s pth and qth moments. Our findings here therefore imply that knowledge of a multiset’s pth and qth moments is exactly enough information to deduce the multiset’s (p,q) entropy and (p,q) mean. Further, knowledge of sufficiently many moments of a multiset can be sufficient to reconstruct the multiset. Conversely, it should be interesting to examine how many (p,q)-entropies and/or (p,q)-means are required to completely determine the multiset.
Regarding the thermodynamic interpretation, we noticed that to require that the p and q norms of multisets coincide is mathematically equivalent to requiring that the two partition functions of two thermodynamic systems coincide at two temperatures. This in turn is equivalent to requiring that the Helmholtz free energy of the two thermodynamic systems coincide at two temperatures. The Helmholtz free energy represents the maximum mechanical work that can be extracted from a thermodynamic system under certain idealized circumstances. This suggests that there perhaps exists a thermodynamic interpretation of the generalized entropies and means in terms of the extractability of mechanical work. In this case, the fact that the generalized entropies and means depend on two rather than one temperature could be related to the fact that the maximum efficiency of a heat engine, obtained in Carnot cycles, is a function of two temperatures. We did show that in the limiting case, when the two temperatures become the same, one recovers the usual Boltzmann factor weighted mean energy as well as the usual Shannon/von Neumann entropy.

Acknowledgments

The authors gratefully acknowledge the research support by the Canada Research Chair, Discovery and Postdoctoral fellowship programs of the Natural Sciences and Engineering Research Council of Canada (NSERC).The first author is grateful for the kind hospitality at the department of Applied Mathematics at the University of Waterloo during the course of this research.

References

  1. Rényi, A. On the foundations of information theory. Rev. Int. Stat. Inst. 1965, 33, 1–14. [Google Scholar] [CrossRef]
  2. Beck, C. Generalized information and entropy measures in physics. Contemp. Phys. 2009, 50, 495–510. [Google Scholar] [CrossRef]
  3. Poon, A.S.Y.; Brodersen, R.W.; Tse, D.N.C. Degrees of freedom in multiple-antenna channels: A signal space approach. IEEE Trans. Inform. Theor. 2005, 51, 523–536. [Google Scholar] [CrossRef]
  4. Migliore, M.D. On the role of the number of degrees of freedom of the field in MIMO channels. IEEE Trans. Antenn. Propag. 2006, 54, 620–628. [Google Scholar] [CrossRef]
  5. Elnaggar, M.S. Electromagnetic Dimensionality of Deterministic Multi-Polarization MIMO Systems. Ph.D. Thesis; University of Waterloo: Waterloo, ON, Canada, 2007. Available online: http://uwspace.uwaterloo.ca/handle/10012/3434 (accessed on 18 April 2012).
  6. Elnaggar, M.S.; Safavi-Naeini, S.; Chaudhuri, S.K. A novel dimensionality metric for multi-antenna systems. In Proceedings of Asia-Pacific Microwave Conference (APMC 2006), Yokohama, Japan, December 2006; pp. 242–245.
  7. Elnaggar, M.S.; Kempf, A. On a Generic Entropy Measure in Physics and Information. In Proceedings of the 7th International Symposium on Modeling and Optimization in Mobile, Ad Hoc, and Wireless Networks (WiOPT 2009), Seoul, Korea, June 2009.
  8. Beck, C.; Schlogl, F. Thermodynamics of Chaotic Systems: An Introduction; Cambridge University Press: Cambridge, UK, 1995. [Google Scholar]
  9. Rényi, A. On measures of entropy and information. In Proceedings of the Fourth Berkeley Symposium on Mathematics, Statistics and Probability, University of California, Berkeley, CA, USA; 1961; Volume 1, pp. 541–561. [Google Scholar]
  10. Aczél, J.; Forte, B.; Ng, C.T. Why the Shannon and Hartley entropies are ‘natural’. Adv. Appl. Probab. 1974, 6, 131–146. [Google Scholar] [CrossRef]
  11. Hartley, R.V.L. Transmission of information. Bell Syst. Tech. J. 1928, 7, 535–564. [Google Scholar] [CrossRef]
  12. Rényi, A. Probability Theory; North-Holland Pub. Co.: Amsterdam, The Netherlands, 1970. [Google Scholar]
  13. Klir, G.J.; Folger, T.A. Fuzzy Sets, Uncertainty and Information; Prentice Hall: Upper Saddle River, NJ, USA, 1988. [Google Scholar]
  14. Khinchin, A.I. Mathematical Foundations of Information Theory; Dover: New York, NY, USA, 1957. [Google Scholar]
  15. Blizard, W.D. Multiset theory. Notre Dame J. Formal Logic 1989, 30, 36–66. [Google Scholar] [CrossRef]
  16. Cover, T.M.; Thomas, J.A. Elements of Information Theory; John Wiley & Sons: Hoboken, NJ, USA, 1991. [Google Scholar]
  17. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423, 623–656. [Google Scholar] [CrossRef]
  18. Ash, R.B. Information Theory; Interscience Publishers: New York, NY, USA, 1965. [Google Scholar]
  19. Reza, F.M. An Introduction to Information Theory; Dover: New York, NY, USA, 1994. [Google Scholar]
  20. Gori, F.; Guattari, G. Shannon number and degrees of freedom of an image. Opt. Commun. 1973, 7, 163–165. [Google Scholar] [CrossRef]
  21. Landau, H.; Pollak, H. Prolate spheroidal wave functions, Fourier analysis and uncertainty: Part III: The dimension of the space of essentially time- and band-limited signals. Bell Syst. Tech. J. 1962, 41, 1295–1336. [Google Scholar] [CrossRef]
  22. Aczel, J.; Daroczy, Z. On Measures of Information and Their Characterizations; Academic Press: New York, NY, USA, 1975. [Google Scholar]
  23. Shigeru, F. An axiomatic characterization of a two-parameter extended relative entropy. J. Math. Phys. 2010, 51, 123302. [Google Scholar]

Appendix A. Proof of Property 4.5: Range of S p q

A.1. Non-Increasing X p with Respect to p

We want to show that X p is non-increasing with respect to p. This is equivalent to showing that p ( log X p ) = p ( log i = 1 N x i p p ) 0 .
p 2 p ( log X p ) = p i x i p j x j p log x i log j x j p                                  = i x i p j x j p log x i p i x i p j x j p log j x j p                                  = i x i p j x j p log x i p j x j p    0
Therefore, X p X q for p<q. The equality occurs only when x i p = x i p , i.e., when { x i } includes exactly one non-zero element. Note that there is a singularity for log X p such that lim p 0 + log X p = and lim p 0 log X p = . The property herein applies over the intervals < p < 0 and 0 < p <

A.2. Non-Decreasing Generalized p-Mean with Respect to p

Let X be a multiset of N strictly positive numbers. For p < q, we want to show that ( i = 1 N x i p N ) 1 / p ( i = 1 N x i q N ) 1 / q . To this end, we define f ( x ) = x q / p , where x > 0 and p / q > 0 , i.e., p , q 0 have the same sign. Clearly, d 2 d x 2 f ( x ) > 0 and thus f ( x ) is convex. From Jensen’s inequality,
f ( w i x i ) w i f ( x i ) ,  where  i = 1 n w i = 1 .  Accordingly , f ( w i x i p ) w i f ( x i p ) ,  hence
( w i x i p ) q / p w i ( x i p ) q / p = w i x i q
( w i x i p ) 1 / p ( w i x i q ) 1 / q
By setting w i = 1 / N , the proof is complete. Equality holds when all x i are equal.

A.3. Proof of Property 4.5: 0 S p q log N

We start by proving that S p q 0 . From Lemma A.1, for p < q, we have log X p log X q . Therefore, log X p log X q 0 , p , q 0 and have the same sign. Since 1 / p 1 / q > 0 , we get log X p log X q 1 / p 1 / q = S p q 0 . Following the same steps for p > q yields the same result. Finally, when p q , we have S q q = i x i q j x j q log ( x i q j x j q ) 0 . In all cases, the equality holds when { x i } includes exactly one non-zero element.
Next, we show that S p q log N . From Lemma A.2, assuming both p , q 0 have the same sign, for p < q, we have ( i = 1 N x i p N ) 1 / p ( i = 1 N x i q N ) 1 / q . Taking the logarithm yields log X p log X q 1 / p 1 / q = S p q log N . For p > q, we obtain the same result.
When p q , we have:
S q q log N = i = 1 N x i q j = 1 N x j q log ( x i q j = 1 N x j q ) i = 1 N x i q j = 1 N x j q log N                              = i = 1 N x i q j = 1 N x j q log ( j = 1 N x j q N x i q )                               i = 1 N x i q j = 1 N x j q ( j = 1 N x j q N x i q 1 ) = i = 1 N ( 1 N x i q j = 1 N x j q ) = 0 ,
where log x ( x 1 ) was used. In all cases, equality holds when all the non-zero elements of X are equal.

Appendix B. Proof of Property 4.6: Monotonicity of S p q

B.1. Non-Degenerate Case p q

We have S p q = log k p q = q log i x i p p log i x i q q p . Accordingly:
( q p ) 2 p S p q = q ( q p ) i x i p j x j p log x i ( q p ) log i x i q + q log i x i p p log i x i q                                 = q ( q p ) i x i p j x j p log x i q log i x i q + q log i x i p                                 = q ( q p ) i x i p j x j p log x i q log j x i p + q log j x i p
where we inserted another summation index for the last two terms in the last step. Therefore:
( q p ) 2 q p S p q = ( q p ) i x i p j x j p log x i log j x i q + log j x i p                                   = i x i p j x j p log x i q i x i p j x j p log x i q log j x i p + log j x i p                                   = i x i p j x j p log ( x i q j x j p x i p j x j q )
where we have multiplied the last two terms in the second step by i x i p j x i p . Consequently:
( q p ) 2 q p S p q i x i p j x j p ( x i q j x j p x i p j x j q 1 ) = 0 ,
where we have used the fact that log x ( x 1 ) . The equality occurs only when x i q j x i p = x i p j x j p for every x i , implying that each x i is either zero or some fixed value x 0 .
Accordingly, after multiplying the last inequality by q 2 ( q p ) 2 , we obtain:
q p S p q 0       p q ,
where the equality holds when all the non-zero elements x i are equal.
Since p and q are assumed to have the same sign (from the condition p q 0 ), we deduce that, with respect to either p or q (while fixing the other order), S p q is non-decreasing for negative order p,q, non-increasing for positive order p,q, with a maximum value at either p , q = 0 (note that p S p q | q = 0 = 0 ). S p q is invariant to p,q if and only if all the non-zero elements x i are equal.

B.2. Degenerate Case p q

From Appendix B.1, if 0 a b c d , then S c d S b d S b c S a c S a b . In the limiting case, when c d and b a + , and knowing from (9) that this limit exists, we readily obtain S d d S a a . In a similar fashion, for negative indices values d c b a 0 , we get S a a S d d . In any case, the equality occurs when all the non-zero elements x i are equal. S p p has a maximum value when p = 0 . Accordingly:
p p S p p 0 .

Appendix C. Proof of Property 4.6: Monotonicity of m p q

C.1. Non-Degenerate Case p q

We have from (5) log X q = log m p q + S p q / q . Differentiating with respect to p, we get:
p m p q = m p q p S p q q .
Since k p q > 0 (Property 4.5), therefore (5) confirms that m p q > 0 as well because the norm is defined to discard any zero elements (6). From Appendix B, p S p q q 0 . Therefore:
p m p q 0      p , q 0       p q
with equality when all the non-zero elements x i are equal.

C.2. Degenerate Case p q

From (9), log m p p ( X ) = log X p S p p ( X ) / p , accordingly:
p m p p m p p = p X p X p p p S p p S p p p 2
Moreover, from (9), S q q ( X ) = q 2 q X q X q . Consequently, p 2 p m p p m p p = p p S p p 0 , where we used the result of Appendix B.2. Therefore, since m p p > 0 , we get:
p m p p 0 .

Appendix D. Proof of Property 4.9: Sub-Additivity of the Effective Cardinality

For p q , we want to show that k p q ( { ξ x i } { η y j } ) k p q ( X ) + k p q ( Y ) , where ξ , η 0 . Our objective is to find a condition in terms of ξ and η that maximizes S p q ( { ξ x i } { η y j } ) . We have:
( p q ) S p q ( { ξ x i } { η y j } ) = p log ( i ξ q x i q + j η q y j q ) q log ( i ξ p x i q + j η p y j p )
By setting ξ S p q = 0 and η S p q = 0 , we obtain the following two conditions:
i ξ 0 q x i q i ξ 0 q x i q + j η 0 q y j q = i ξ 0 p x i q i ξ 0 p x i q + j η 0 p y j p
j η 0 q y j q i ξ 0 q x i q + j η 0 q y j q = j η 0 p y j p i ξ 0 p x i q + j η 0 p y j p
From (50) and (51), we have:
i ξ 0 q x i q i ξ 0 p x i q = j η 0 q y j q j η 0 p y j p
Accordingly, the required proportionality condition is:
ξ 0 η 0 = ( j y j q / j y j p i x i q / i x i p ) 1 q p
From (49), we have:
k p q ( { ξ 0 x i } { η 0 y j } ) = ( ( i ξ 0 q x i q + j η 0 q y j q ) p ( i ξ 0 p x i p + j η 0 p y j p ) q ) 1 p q
Substituting i ξ 0 p x i p + j η 0 p y j p from (50) in (54) yields:
k p q ( { ξ 0 x i } { η 0 y j } ) = ( ( i ξ 0 q x i q ) q ( i ξ 0 p x i p ) q ( i ξ 0 q x i q + j η 0 q y j q ) p q ) 1 p q                                             = ( i ξ 0 q x i q ) q p q ( i ξ 0 p x i p ) q p q ( i ξ 0 q x i q + j η 0 q y j q )                                             = ( i ξ 0 q x i q ) 1 + q p q ( i ξ 0 p x i p ) q p q + ( j η 0 q y j q ) ( i ξ 0 q x i q i ξ 0 p x i p ) q p q                                             = ( i ξ 0 q x i q ) p p q ( i ξ 0 p x i p ) q p q + ( j η 0 q y j q ) ( j η 0 q y j q j η 0 p y j p ) q p q                                             = ( i ξ 0 q x i q ) p p q ( i ξ 0 p x i p ) q p q + ( j η 0 q y j q ) p p q ( j η 0 p y j p ) q p q
where we employed (52) to obtain the second to last step. Consequently, since ξ 0 and η 0 cancel out in the last step, we get:
k p q ( ξ 0 X η 0 Y ) = ( ( i x i q ) p ( i x i p ) q ) 1 p q + ( ( j y j q ) p ( j y j p ) q ) 1 p q                                     = k p q ( X ) + k p q ( Y )
.
In order to confirm that (53) is indeed a maximizing condition of k p q ( ξ 0 X η 0 Y ) , consider the following example. Let X = { x 0 } and Y = { y 0 } with multiplicity K X and K Y , respectively. From the range of k p q (Property 4.5), we know that max ( k p q ( ξ 0 X η 0 Y ) ) = K X + K Y only when all the elements are equal, i.e., when ξ 0 x 0 = η 0 y 0 . Indeed, (53) confirms this maximizing condition and thus k p q ( ξ X η Y ) k p q ( X ) + k p q ( Y ) .
By taking the limit when p q , (53) becomes
ξ 0 η 0 = Y q X q ( k q q ( X ) k q q ( Y ) ) 1 / q                p q
and k q q ( ξ X η Y ) k q q ( X ) + k q q ( Y ) .

Share and Cite

MDPI and ACS Style

Elnaggar, M.S.; Kempf, A. Equivalence of Partition Functions Leads to Classification of Entropies and Means. Entropy 2012, 14, 1317-1342. https://doi.org/10.3390/e14081317

AMA Style

Elnaggar MS, Kempf A. Equivalence of Partition Functions Leads to Classification of Entropies and Means. Entropy. 2012; 14(8):1317-1342. https://doi.org/10.3390/e14081317

Chicago/Turabian Style

Elnaggar, Michel S., and Achim Kempf. 2012. "Equivalence of Partition Functions Leads to Classification of Entropies and Means" Entropy 14, no. 8: 1317-1342. https://doi.org/10.3390/e14081317

APA Style

Elnaggar, M. S., & Kempf, A. (2012). Equivalence of Partition Functions Leads to Classification of Entropies and Means. Entropy, 14(8), 1317-1342. https://doi.org/10.3390/e14081317

Article Metrics

Back to TopTop