Next Article in Journal / Special Issue
Solution of Excited Non-Linear Oscillators under Damping Effects Using the Modified Differential Transform Method
Previous Article in Journal
Coefficient Inequalities of Second Hankel Determinants for Some Classes of Bi-Univalent Functions
Previous Article in Special Issue
Microtubules Nonlinear Models Dynamics Investigations through the exp(−Φ(ξ))-Expansion Method Implementation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Note on Burg’s Modified Entropy in Statistical Mechanics

1
Department of Mathematics, Rajyadharpur Deshbandhu Vidyapith, Serampore, Hooghly 712203, West Bengal, India
2
Department of Mathematics, Indian Institute of Engineering Science and Technology (IIEST), Shibpur, Howrah 711103, West Bengal, India
*
Author to whom correspondence should be addressed.
Mathematics 2016, 4(1), 10; https://doi.org/10.3390/math4010010
Submission received: 3 December 2015 / Revised: 2 February 2016 / Accepted: 16 February 2016 / Published: 27 February 2016
(This article belongs to the Special Issue Applied Mathematics and Mechanics)

Abstract

:
Burg’s entropy plays an important role in this age of information euphoria, particularly in understanding the emergent behavior of a complex system such as statistical mechanics. For discrete or continuous variable, maximization of Burg’s Entropy subject to its only natural and mean constraint always provide us a positive density function though the Entropy is always negative. On the other hand, Burg’s modified entropy is a better measure than the standard Burg’s entropy measure since this is always positive and there is no computational problem for small probabilistic values. Moreover, the maximum value of Burg’s modified entropy increases with the number of possible outcomes. In this paper, a premium has been put on the fact that if Burg’s modified entropy is used instead of conventional Burg’s entropy in a maximum entropy probability density (MEPD) function, the result yields a better approximation of the probability distribution. An important lemma in basic algebra and a suitable example with tables and graphs in statistical mechanics have been given to illustrate the whole idea appropriately.

1. Introduction

The concept of entropy [1] figured strongly in the physical sciences during the 19th century, especially in thermodynamics and statistical mechanics [2], as a measure of equilibrium and evolution of thermodynamic systems. Two main views were developed which were the macroscopic view formulated originally by Clausius and Carnot and the microscopic approach associated with Boltzmann and Maxwell. Since then, both the approaches have made introspection in natural thermodynamic and microscopically probabilistic systems possible. Entropy is defined as the measure of a system’s thermal energy per unit temperature that is unavailable for doing useful work. Because work is obtained from ordered molecular motion, the amount of entropy is also a measure of molecular disorder or randomness of a system. The concept of entropy provides deep insight into the direction of spontaneous change for many day-to-day phenomena. Now, how entropy was developed by Rudolf Clausius [3] is discussed below.

1.1. Clausius’s Entropy

To provide a quantitative measure for the direction of spontaneous change, Clausius introduced the concept of entropy as a precise way of expressing the second law of thermodynamics. The Clausius form of the second law states that spontaneous change for an irreversible process [4] in an isolated system (that is, one that does not exchange heat or work with its surroundings) always proceeds in the direction of increasing entropy. By the Clausius definition, if an amount of heat Q flows into a large heat reservoir at temperature T above absolute zero, then Δ s = Q T . This equation effectively gives an alternate definition of temperature that agrees with the usual definition. Assume that there are two heat reservoirs R 1 and R 2 at temperatures T 1 and T 2 . (such as the stove and the block of ice). If an amount of heat Q flows from R 1 to R 2 , then the net entropy change for the two reservoirs is Δ s = Q ( 1 T 2 1 T 1 ) which is positive, provided that T 1 T 2 . Thus, the observation that heat never flows spontaneously from cold to hot is equivalent to requiring the net entropy change to be positive for a spontaneous flow of heat. When the system is in thermodynamic equilibrium, then d S = 0 , i.e., if T 1 = T 2 , then the reservoirs are in equilibrium, no heat flows, and Δ s = 0 . If the gas absorbs an incremental amount of heat d Q from a heat reservoir at temperature T and expands reversibly against the maximum possible restraining pressure P , then it does the maximum work and d W = P d V . The internal energy of the gas might also change by an amount d U as it expands. Then, by conservation of energy, d Q = d U + P d V . Because the net entropy change for the system plus reservoir is zero when maximum work is done and the entropy of the reservoir decreases by an amount d S r e s e r v o i r = d Q T , this must be counterbalanced by an entropy increase of d S s y s t e m = d U + P d V T = d Q T for the working gas so that d S s y s t e m + d S r e s e r v o i r = 0 . For any real process, less than the maximum work would be done (because of friction, for example), and so the actual amount of heat d Q absorbed from the heat reservoir would be less than the maximum amount d Q . For example, the gas could be allowed to expand freely into a vacuum and do no work at all. Therefore, it can be stated that d S s y s t e m = d U + P d V T d Q T with, d Q = d Q in the case of maximum work corresponding to a reversible process. This equation defines S s y s t e m as a thermodynamic state variable, meaning that its value is completely determined by the current state of the system and not by how the system reached that state. Entropy is a comprehensive property in that its magnitude depends on the amount of material in the system.
In one statistical interpretation of entropy, it is found that for a very large system in thermodynamic equilibrium, entropy S is proportional to the natural logarithm of a quantity Ω corresponding to S and can be realized; that is, S = K ln Ω , in which K is related to molecular energy. On the other hand, entropy generation analysis [5,6,7,8,9,10,11] is used to optimize the thermal engineering devices for higher energy efficiency; it has attracted wide attention to its applications and rates in recent years. In order to access the best thermal design of systems, by minimizing the irreversibility, the second law of thermodynamics could be employed. Entropy generation is a criterion for the destruction of a systematized work.The development of the theory followed two conceptually different lines of thought. Nevertheless, they are symbiotically related, in particular through the work of Boltzmann.

1.2. Boltzmann’s Entropy

In addition to thermodynamic (or heat-change) entropy, physicists also study entropy statistically [12,13]. The statistical or probabilistic study of entropy is presented in Boltzmann’s law. Boltzmann’s equation is somewhat different from the original Clausius (thermodynamic) formulation of entropy. Firstly, the Boltzmann formulation is structured in terms of probabilities, while the thermodynamic formulation does not consist in the calculation of probabilities. The thermodynamic formulation can be characterized as a mathematical formulation, while the Boltzmann formulation is statistical. Secondly, the Boltzmann equation yields a value of entropy S while the thermodynamic formulation yields only a value for the change in entropy ( d S ) . Thirdly, there is a shift in content, as the Boltzmann equation was developed for research on gas molecules rather than thermodynamics. Fourthly, by incorporating probabilities, the Boltzmann equation focuses on microstates, and thus explicitly introduces the question of the relationship between macrostates and microstates. Boltzmann investigated such microstates and defined entropy in a new way such that the macroscopic maximum entropy state corresponded to a thermodynamic configuration which could be formulated by the maximum number of different microstates. He noticed that the entropy of a system can be considered as a measure of the disorder in the system and that in a system having many degrees of freedom, the number measuring the degree of disorder also measured the uncertainty in a probabilistic sense about the particular microstates.
The value W was originally intended to be proportional to the Wahrscheinlichkeit (means probability) of a macrostate for some probability distribution of a possible microstate, in which the thermodynamic state of a system can be realized by assigning different ξ and ρ of different molecules. The Boltzmann formula is the most general formula for thermodynamic entropy; however, his hypothesis was for an ideal gas of N identical particles, of which N i ( i = 1 , 2 , . , l ) are the i th microscopic condition of position and momentum of a given distribution D i = ( N 1 , N 2 , . , N l ) . Here, D 1 = ( N , 0 , . , 0 ) , D 2 = ( N 1 , 1 , . , 0 ) ,…etc. For this state, the probability of each microstate system is equal, so it was equivalent to calculating the number of microstates associated with a macrostate. Then the statistical disorder is given by [14] W = N ! N 1 ! N 2 ! . . N l ! , [ N 1 + N 2 + . . + N l = N ] . Therefore, the entropy given by Boltzmann is: S B = K ln W Where ln W = ln N ! N 1 ! N 2 ! . . N l ! = ln N ! i = 1 l ln ( N i ! )
Let us now take an approximate value of W for a large N ! . Using Stirling’s approximation ln N ! N ln N N , we have:
ln W N ln N N [ i = 1 l N i ln ( N i ) i = 1 l ( N i ) ] = i = 1 l N i ln N i = N i = 1 l p i ln p i
S B = K i = 1 n p i ln p i where p i = N i N is the probability of the occurrence of i th microstates. Boltzmann was the first to emphasize the probabilistic meaning of entropy and the probabilistic nature of thermodynamics.

1.3. Information Theory andShannon’s Entropy

Unlike the first two entropy approaches (thermodynamic entropy by Clausius and Boltzmann’s entropy), the third major form of entropy did not fall within the field of physics, but was developed instead in a new field known as information theory [15,16,17] (also known as communication theory). A fundamental step in using entropy in new contexts unrelated to thermodynamics was provided by Shannon [18], who came to conclude that entropy could be used to measure types of disorder other than that of thermodynamic microstates. Shannon was interested in information theory [19,20], particularly in the ways in which information can be conveyed via a message. This led him to examine probability distributions in a very general sense and he worked to find a way of measuring the level of uncertainty in different distributions.
For example, suppose the probability distribution for the outcome of a coin toss experiment is P(H) = 0.999 and P(T) = 0.001. One is likely to notice that there is much more “certainty” than “uncertainty” about the outcome of this experiment and, consequently, the probability distribution. If, on the other hand, the probability distribution governing that same experiment were P(H) = 0.5 and P(T) = 0.5, then there is much less “certainty” and much more “uncertainty” when compared to the previous distribution. However, how can these uncertainties can be quantified? Is there some algebraic function which measures the amount of uncertainty in any probabilistic distribution in terms of the individual probabilities? From these types of simple examples and others, Shannon was able to devise a set of criteria which any measure of uncertainty may satisfy. He then tried to find an algebraic form which would satisfy his criteria and discovered that there was only one formula which fit. Let the probabilities of n possible outcomes E 1 , E 2 , . , E n of an experiment be p 1 , p 2 , . . , p n , giving rise to the probability distribution P = ( p 1 , p 2 , , p n ) ; i = 1 n p i = 1 ; p i s 0 . There is an uncertainty as to the outcome when the experiment is performed. Shannon suggested the measure i = 1 n p i ln p i , which is identical to the previous entropy relation if the constant of probability is taken as the Boltzmann constant K . Thus, Shannon showed that entropy, which measures the amount of disorder in a thermodynamic system, also measures the amount of uncertainty in any probability distribution. Let us now give the formal definition of Shannon’s entropy as follows: Consider a random experiment P = ( p 1 , p 2 , , p n ) whose possible outcomes have probabilities p i , i = 1 , 2 , , n that are known. Can we guess in advance which outcome we shall obtain? Can we measure the amount of uncertainty? We shall denote such an uncertainty measure by H ( P ) = H n ( p 1 , p 2 , , p n ) . The most common as well as the most useful measure of uncertainty is Shannon's informational entropy (which should satisfy some basic requirements), which is defined as follows:
Definition I: 
Let ( p 1 , p 2 , , p n ) be the probability of the occurrence of the events E 1 , E 2 , , E n associated with a random experiment. The Shannon’s entropy probability distribution ( p 1 , p 2 , , p n ) of the random experiment system P is defined by H ( P ) = H n ( p 1 , p 2 , , p n ) = i = 1 n p i ln p i where, 0 ln 0 0 . The above definition is generalized straightforwardly as the definition of entropy of a random variable.
Definition II: Let X R be a discrete random variable which takes the value x i ( i = 1 , 2 , 3 , . , n ) with the probability p i , i = 1 , 2 , , n ; then the entropy H ( X ) of X is defined by the expression H ( X ) = i = 1 n p i ln p i Examination of H or H n ( P ) reveals why Shannon’s measure is the most satisfactory measure of entropy because of the following:
(i)
H n ( P ) is a continuous function of p 1 , p 2 , , p n .
(ii)
H n ( P ) is a symmetric function of its arguments.
(iii)
H n + 1 ( p 1 , p 2 , , p n , 0 ) = H n ( p 1 , p 2 , , p n ) , i.e., it should not change if there is an impossible outcome to the probability.
(iv)
Its minimum is 0 when there is no uncertainty about the outcome. Thus, it should vanish when one of the outcomes is certain to happen so that
H n ( p 1 , p 2 , , p n ) = 0 ; p i = 1 , p j = 0 , i j , i = 1 , 2 , , n
(v)
It is the maximum when there is maximum uncertainty, which arises when the outcomes are equally likely so that H n ( P ) is the maximum when p 1 = p 2 = . . = p n = 1 n .
(vi)
The maximum value of H n ( P ) increases with n .
(vii)
For two independent probability distributions P ( p 1 , p 2 , , p n ) and Q ( q 1 , q 2 , , q m ) , the uncertainty of the joint scheme P Q should be the sum of their uncertainties: H n + m ( P Q ) = H n ( P ) + H m ( Q )
Shannon’s entropy has various applications in the field of portfolio analysis, the measurement of economic analysis, transportation, and urban and regional planning as well as in the fields of statistics, thermodynamics, queuing theory, parametric estimation, etc. It has been used in the field non-commensurable and conflicting criteria [21] and in the nonlinear complexity of random sequences [22] as well.

2. Discussion

2.1. Jaynes’ Maximum Entropy (MaxEnt) Principle

Let the random variable of an experiment be X , and assume the probability mass associated with the value x i is p i , i.e., P X ( i ) P ( X ) = p i , i = 1 , 2 , . , n . The set ( X , P ) = { ( x 1 , p 2 ) , ( x 2 , p 2 ) , . , ( x n , p n ) } , i = 1 n p i = 1 is called the source ensemble as described by Karmeshu [23]. In general, we may find expected values of the functions g 1 ( X ) , g 2 ( X ) , . . , g z ( X ) to get i = 1 n p i g r i = a r , r = 1 , 2 , . , z , p i 0 and with natural constraint i = 1 n p i = 1 , given a number of constraints. Thus, we have z + 1 relations between ( p 1 , p 2 , , p n ) . There may be infinite probability distributions ( p 1 , p 2 , , p n ) satisfying the above equation. If we know only g 1 ( x ) , g 2 ( x ) , . . , g z ( x ) , then we get a family of max entropy distributions. If, in addition, we know the values of a 1 , a 2 , , a z , we get a specific member of this family and we call it the max entropy probability distribution. According to a great article by Jaynes [24,25,26], we choose the probability distribution out of all these which maximizes the measure of entropy as shown by Shannon’s equation, S ( P ) = i = 1 n p i ln p i .
Any distribution of the form p i = exp [ λ 0 λ 1 g 1 ( x i ) λ 2 g ( x i ) 2 . . λ z g z ( x i ) ] , i = 1 , 2 , , n may be regarded as the maximum entropy distribution where λ 0 , λ 1 , . , λ z are determined as functions of a 1 , a 2 , , a z ; then the maximum entropy S max is given by S max = λ 0 + λ 1 a 1 + λ 2 a 2 + + λ z a z .
Kapur [27,28] showed that there is always a concave function of a 1 , a 2 , , a m . We also note that all the probabilities given by p i are always positive. We naturally want to know whether there is another measure of entropy other than Shannon’s entropy which, when maximized, subject to, i = 1 n p i = 1 , i = 1 n p i g r i = a r , r = 1 , 2 , . , z , p i 0 , gives positive probabilities and for which S max is possibly a concave function [29] of parameters. Kapur [30] studied that Burg’s [31] measure of entropy, which has been very successfully used in spectral analysis, does always give positive probabilities. The maximum entropy principal of Jaynes has been used frequently to derive the distribution of statistical mechanics by maximizing the entropy of the system subject to some given constraints. The Maxwell-Boltzman distribution is obtained when there is only one constraint on a system which prescribes the expected energy per particle of the system by Bose-Einstein (B.E.) distribution, Fermi-Dirac (F.D.) distribution and intermediate statistics (I.S.) distributions; these are obtained by maximizing the entropy subject to two constraints by Kapur and Kesavan, and Kullback [32,33] and also by the present authors [34].

2.2. Formulation of MEPD in Statistical Mechanics Using Shannon’s Measure of Entropy

Let p 1 , p 2 , . . , p n be the probabilities of a particle having energy levels ε 1 , ε 2 , …., ε n , respectively, and let the expected value of energy be prescribed as ε ¯ ; then, to get MEPD, we maximize the Shannon’s measure of entropy:
S ( P ) = i = 1 n p i ln p i
Subject to
i = 1 n p i = 1 ,   i = 1 n p i ε i = ε ¯
Let the Lagrangian be
L = i = 1 n p i ln p i ( λ 1 ) ( i = 1 n p i 1 ) μ ( i = 1 n p i ε i ε ¯ )
Differentiating with respect to p i ’s, we get:
ln p i + λ + μ ε i = 0 p i = exp ( λ μ ε i ) where, λ , μ are to be determined by using Equation (2) so that
p i = exp ( μ ε i ) i = 1 n exp ( μ ε i ) , i = 1 , 2 , . , n ,
Where
i = 1 n ε i exp ( μ ε i ) i = 1 n exp ( μ ε i ) = ε ¯
Equation (4) is the well-known Maxwell-Boltzmann distribution from statistical mechanics which is used in many areas [35,36,37].

2.3. Burg’s Entropy Measure and MEPD

When S ( P ) = i = 1 n p i ln p i was replaced by Burg’s measure of entropy B ( P ) = i ln p i , it gave interesting results as shown by Kapur. Burg’s measure of entropy is always negative, but this does not matter in entropy maximization, where it has been found that a probability distribution with maximum entropy satisfies the same constraint and it does not matter if all the entropies are negative. So, in Equation (1) when we use
B ( P ) = i ln p i ,  with the constraints  i = 1 n p i = 1 , i = 1 n i p i = m
we get
p i = 1 λ + μ i ,   i = 1 , 2 , 3 , . , n
where λ & μ are obtained by solving the equations
i 1 λ + μ i = 1 ,   i i λ + μ i = m
Multiplying first and second Equation (8) by λ & μ respectively then adding
We get
i λ + μ i λ + μ i = λ + μ m  or ,   n = λ + μ m
so that from Equation (8),
i = 1 n 1 n μ ( m i ) = 1
Then, μ = 0 is an obvious solution but that will give us λ = n and this will satisfy the second equation of (8)
if m = n + 1 2 . Now, Equation (10) is the n th degree polynomial in μ , and one of its roots is zero. Its non-zero solutions will be obtained by solving an equation of ( n 1 ) th degree in μ . Lemma has been proved by Kapur as the following:
Lemma
All the roots of r = 1 n 1 n μ ( m i ) = 1 are real; in other words, none of the roots can be complex.
Proof: 
Let μ = α + i β , i = 1 be a pair of complex conjugate roots of Equation (10). Then,
r = 1 n 1 n ( α + i β ) ( m i ) = 1 and r = 1 n 1 n ( α i β ) ( m i ) = 1 ; subtracting the second from the first, we get
r = 1 n 2 i β ( m r ) { n α ( m r ) } 2 + β 2 = 0 , which gives β = 0 , or r = 1 n ( m r ) { n α ( m r ) } 2 + β 2 = 0 . The second possibility can easily be ruled out. To find the actual location of the n real roots, let us assume
f ( μ ) r 1 n μ ( m r ) 1 and this function is discontinuous at the following points:
n n m , n n ( m 1 ) , . , n 1 l , n l , , n m 2 , n m 1 , where m = k + l , k = [ m ] and l is a “+” fraction. More precisely, f ( μ ) + when μ points from one side and f ( μ ) when μ points from other side. Again,
f ( μ ) r 1 n μ ( m r ) 1 = 1 n [ r { 1 μ ( m r n ) } 1 ] 1
= 1 n [ r { 1 + μ ( m r n ) } + O ( μ 2 ) ] 1 = 1 n [ n + μ n { m n n ( n + 1 ) 2 } ] 1 + O ( μ 2 )
= μ n { m ( n + 1 ) 2 } + O ( μ 2 )

2.4. d Burg’s Modifie Entropy (MBE) Measure and MEPD

2.4.1. Monotonic Character of MBE

We propose use of Burg’s modified entropy instead of Burg’s entropy. Maximizing the Burg’s modified measure of entropy:
B mod ( P ) = i = 1 n ln ( 1 + a p i ) ln ( 1 + a ) , a 0
d d a ( B mod ( P ) ) = i = 1 n p i 1 + a p i 1 1 + a = i = 1 n p i [ 1 1 + a p i 1 1 + a ] = i = 1 n a p i ( 1 p i ) ( 1 + a ) ( 1 + a p i ) 0
Therefore, B mod ( P ) is the monotonic increasing function of a . For the probability distribution
( p , 1 p )
d d a ( B mod ( P ) ) = i = 1 n p i 1 + a p i 1 1 + a = i = 1 n p i [ 1 1 + a p i 1 1 + a ] = i = 1 n a p i ( 1 p i ) ( 1 + a ) ( 1 + a p i ) 0
it is showed (see: Table 1.) that M a x P ( B mod ( P ) ) = ln ( 1 + a 2 4 ( 1 + a ) ) = ln ( ( a + 2 ) 2 4 ( 1 + a ) ) .
The measure of entropy M a x P ( B mod ( P ) ) is the Burg’s modified entropy. This is a better measure than the standard Burg’s measure since it is always positive and there is no computational problem when p i is very small. In the above case, the maximum value increases with the number of possible outcomes n .

2.4.2. MBE and Its Relation with Burg’s Entropy

B mod ( P ) = i = 1 n ln ( 1 + a p i ) ln ( 1 + a ) = i = 1 n [ ln a + ln ( p i + 1 a ) ] ln a ln ( 1 + 1 a ) = ( n 1 ) ln a + i = 1 n ln ( p i + 1 a ) ln ( 1 + 1 a )
So, when a , maximizing ( B mod ( P ) ) and ( B ( P ) ) will give the same result in both cases;again, if ( B mod ( P ) ) is maximized under the constraints
i = 1 n p i = 1 , i = 1 n p i g r i = a r , r = 1 , 2 , . , m , p i 0
we get 1 1 + a p i = [ λ 0 + λ 1 g 1 ( x i ) + λ 2 g ( x i ) 2 + . . + λ m g m ( x i ) ] , i = 1 , 2 , , n . Letting, we have
1 p i = λ 0 + λ 1 g 1 ( x i ) + λ 2 g ( x i ) 2 + . . + λ m g m ( x i )
The λ i ’s are determined by using Constraints (14) and (15) and this gives the MEPD when Burg’s entropy is maximized as subject to Equation (14).
Therefore, when a , the MEPD of BME Burg’s MEPD; in fact:
( B mod ( P ) ) ( B ( P ) ) = ( n 1 ) ln a + i = 1 n ln ( p i + 1 a ) p i ln ( 1 + 1 a )  as
a .

2.4.3. MBE and Its Concavity of S max under Prescribed Mean

Maximize  B mod ( P ) = i = 1 n ln ( 1 + a p i ) ln ( 1 + a )
subjectto i = 1 n p i = 1 , i = 1 n i p i = m .
We obtain this using Lagranges multiplier mechanics:
( p i + 1 a ) = 1 λ + μ i  Where ,   i 1 λ + μ i = i ( p i + 1 a ) = 1 + n a
From the above equation:
i i .1 λ + μ i = i ( p i + 1 a ) = i = 1 n ( i p i + i a )
i i λ + μ i = m + n ( n + 1 ) 2 a , i.e., n ( n + 1 ) 2 a + m = i i λ + μ i and
μ n ( n + 1 ) 2 a + μ m = i μ i λ + μ i ,  therefore  i λ λ + μ i = i ( p i + 1 a ) = λ ( 1 + n a )
So
i λ λ + μ i + i μ i λ + μ i = λ ( 1 + n a ) + μ n ( n + 1 ) 2 a + μ m = i λ + μ i λ + μ i = i n 1 = n
λ ( 1 + n a ) + μ n ( n + 1 ) 2 a + μ m = n λ = n μ ( m + n ( n + 1 ) 2 a ) 1 + n a
Therefore,
λ + μ i = n + μ [ ( 1 + n a ) i m n ( n + 1 ) 2 a ] 1 + n a
and
p i = a + n a n + μ [ ( n + a ) i a m n ( n + 1 ) 2 ] 1 a , i = 1 , 2 , . , n ,
where μ is determined as a function of m and that is
1 + n a = i n a + n a n + μ [ ( n + a ) i a m n ( n + 1 ) 2 ] &
1 a = i n 1 a n + μ [ ( n + a ) i a m n ( n + 1 ) 2 ] ,
Therefore,
S max = i ln ( 1 + a p i ) ln ( 1 + a ) = i n ln a λ + μ i ln ( 1 + a ) = ln a n 1 + a i ln ( λ + μ i ) = ln a n 1 + a i n ln [ n + μ ( ( 1 + n a ) i m n ( n + 1 ) 2 ) ] + n ln ( 1 + n a )
d S max d m = i = 1 n [ d μ d m { ( 1 + n a ) i m n ( n + 1 ) 2 a } μ ] n + μ [ ( 1 + n a ) i m n ( n + 1 ) 2 a ] = ( p i + 1 a ) a + n a [ d μ d m [ ( 1 + n a ) i m n ( n + 1 ) 2 a ] μ ] = d μ d m [ m + n ( n + 1 ) 2 m n ( n + 1 ) 2 ] + μ a a + n ( 1 + n a )
So that
d S max d m = d μ d m & d 2 S max d m 2 = d μ d m = μ i n ( p i + 1 a ) 2 i n ( p i + 1 a ) 2 [ ( 1 + n a ) i m n ( n + 1 ) 2 a ]
S max will be a concave function of m if d μ d m 0 , that is, if either μ 0 , 0 when the denominator 0 , 0 , respectively.
In the above case, when a = 1 we get S max and the derivative of S max as follows:
d S max d m = i 1 λ + μ i [ d d m ( λ + μ i ) ] = i [ n 2 d μ d m + 1 n + 1 ( m d μ d m + μ ) d μ d m i ]
= d μ d m i ( 1 + p i ) ( n 2 + m n + 1 i ) + μ n + 1 i ( 1 + p i )
= d μ d m i ( n 2 + m n + 1 i ) + d μ d m ( n 2 + m n + 1 ) d μ d m i i p i + μ n n + 1 + μ n + 1
= d μ d m i ( n 2 . n + m n + 1 . n n ( n + 1 ) 2 ) + d μ d m ( n 2 + m n + 1 ) d μ d m m + μ n n + 1 + μ n + 1
= d μ d m ( m n n + 1 m n + 1 m ) + μ n n + 1 + μ n + 1 = μ ( n n + 1 + 1 n + 1 ) = μ
So that
d 2 S max d m 2 = d μ d m
Additionally, we have
0 = i ( 1 + p i ) 2 [ n 2 d μ d m + 1 n + 1 ( m d μ d m + μ ) d μ d m i ]
So that
d μ d m i ( 1 + p i ) 2 ( n 2 + m n + 1 i ) + μ n + 1 i ( i + p i ) 2 = 0
Therefore,
d μ d m = μ n + 1 i ( 1 + p i ) 2 i ( 1 + p i ) 2 ( i n 2 m n + 1 )
So,
d 2 S max d m 2 0 d μ d m 0
Therefore, from Equations (27) and (30),
S max will be a concave function of m if either
( i )   μ 0 , i ( 1 + p i ) 2 ( i n 2 m n + 1 ) 0
Or
( ii )   μ 0 , i ( 1 + p i ) 2 ( i n 2 m n + 1 ) 0
Additionally, when μ = 0 , d 2 S max d m 2 = 0 , all the probabilities are equal, and from Equation (24) we have:
We have,
m + n ( n + 1 ) 2 = 1 λ n ( n + 1 ) 2  or ,   m = n ( n + 1 ) 2 [ 1 λ 1 ] m = n + 1 2 .
Now, if we proceed algebraically as done Section 2.3, we get:
1 + p i = 1 λ + μ i = 1 1 n + 1 ( n μ m ) μ ( n 2 i ) , i = 1 , 2 , . , n ,
i = 1 n 1 1 n + 1 ( n μ m ) μ ( n 2 i ) = 1 n ( 1 + p i ) = n + 1 . The obvious solution of the above problem is μ = 0 which will give λ = 1 p i = n , i.e., uniform distribution, and thus we get, m = n + 1 2 .

3. An Illustrative Example in Statistical Mechanics

3.1. Example

Let p 1 , p 2 , . . , p 10 be the probabilities of a particle having energy levels ε 1 , ε 2 , …., ε 10 ; respectively, and let the expected value of energy be prescribed as m ; then, we get the maximum entropy probability distribution (MEPD) with MBE as follows:
Maximize  i = 1 10 ln ( 1 + p i ) ln 2  such that ,   i = 1 10 p i = 1  and  i = 1 10 i p i = m  for  m = 1 ( 0.25 ) 10
Solution: 
Maximizing the measure of entropy subject to the given constraints, we have ( 1 + p i ) = 1 λ + μ i .
When i 1 λ + μ i = 10 + 1 = 11 , 10 ( 10 + 1 ) 2 + m = i i λ + μ i i.e., i i λ + μ i = 55 + m , we get the probability distribution in the form of a table and also get the values of S max as described by Kapur and Kesavan.
There may be two cases,
Case (i) when m n + 1 2 = 5.5 .
In this case, when m lies between 1 and 5.5, μ 0 implies d S max d m 0 and then S max is increasing.
Case (ii) when m n + 1 2 = 5.5 .
In this case, when m lies between 5.5 and 10, μ 0 implies d S max d m 0 and then S max is decreasing.
When m = n + 1 2 = 5.5 , μ = 0 , d S max d m = 0 .
it can be shown that S max will be concave if we prescribe E ( g ( x ) ) instead of E ( x ) , where g ( x ) is a monotonic increasing function of x ; then we will apply the necessary changes. Again, since the concavity of S max has already been proven, this will enable us to handle the inequality constraint of the type m 1 i = 1 n i p i m 2 .

3.2. Simulated Results

Following Table (Table 2) Found Using LINGO Software 2011 where Different Max-Entropy Values Are Given for Different m Values:
Graphs obtained from the above table are given on the next page (Figure 1, Figure 2 and Figure 3):
If p i is plotted against i , we get rectangular hyperbolic types of curves (Table 3 and Figure 4).

4. Conclusions

In the present paper we have presented different MEPDs and respective entropy measures with their properties. It has been found that MBE is a better measure than Burg’s entropy when the maximized subject to the mean is prescribed, and it also has been shown that unlike Burg’s entropy, the maximum value of MBE increases with n . The main problem here will consist of solving z + 1 simultaneous transcendental equations for the Lagranges multipliers. An application in statistical mechanics with simulated data has been studied with the help of Lingo11 software and corresponding graphs are provided. Now, one question arises: Will this result continue to hold for other moment constraints also? When we take generalized moment expectation of g ( x ) instead of expectation of x , then g ( x ) must be a monotonic increasing function of x , and if p i becomes negative for some values of the moments, then we have to set those probabilities to zero and reformulate the problem for the remaining probabilities over the remaining range and solve it.

Acknowledgments

Authors would like to express their sincere gratitude and appreciation to the reviewers for their valuable comments and good suggestions to make this paper enriched and impactful, and a special thank you to the Assistant Editor.

Author Contributions

A. Ray and S. K. Majumder conceived and designed the experiments; A. Ray performed the experiments, analyzed the data and both A. Ray and S. K. Majumder contributed in analysis tools; A. Ray wrote the paper.

Conflicts of Interest

The authors declare no conflict of interest.

Nomenclature

  • S = Entropy
  • p i = N i N , probability of i th event
  • H ( P ) = H n ( p 1 , p 2 , , p n ) = Information entropy
  • S ( P ) o r H n ( P ) = i = 1 n p i ln p i = Shannon’s entropy
  • B ( P ) = i ln p i = Burg’s entropy
  • B mod ( P ) = Burg’s modified entropy
  • n = Number of energy levels/number of possible outcome
  • μ = 1 K T
  • K = Boltzmann constant
  • T = Absolute temperature
  • N = Identical particle of ideal gas
  • Δ s = Increase of entropy
  • d S = Change in entropy
  • [ m ] = Greatest integer value of m
  • Where m is the mean value:
  • E ( x ) = Expectation
  • d V = Change in volume
  • P Q = Union of two sets
  • m = 1 ( 0.25 ) 10 , m ranges from 1 to 10 with step length of 0.25
  • S max = Maximum entropy
  • M a x P ( B mod ( P ) ) = Maximum value under the given probability distribution
  • MBE=Modified Burg’s Entropy

Greek Symbols

  • Ω = The maximum number of microscopic ways in the macroscopic state
  • ξ = Position of the molecule
  • ρ = Momentum of the molecule
  • μ , λ =Lagrangian constant
  • ε 1 , ε 2 , . =Different energy levels
  • ε ¯ =Mean energy

Subscripts 

  • B=Boltzmann
  • max =Maximum
  • mod =modified

Superscript 

  • λ =New constant different from λ

References

  1. Andreas, G.; Keller, G.; Warnecke, G. Entropy; Princeton University Press: Princeton, NJ, USA, 2003. [Google Scholar]
  2. Phil, A. Thermodynamics and Statistical Mechanics: Equilibrium by Entropy Maximisation; Academic Press: Cambridge, MA, USA, 2002. [Google Scholar]
  3. Rudolf, C. Ueber verschiedene für die Anwendung bequeme Formen der Hauptgleichungen der mechanischen Wärmetheorie. Annalen der Physik 1865. (In German) [Google Scholar] [CrossRef]
  4. Wu, J. Three Factors Causing the Thermal Efficiency of a Heat Engine to Be Less than Unity and Their Relevance to Daily Life. Eur. J. Phys. 2015, 36, 015008. [Google Scholar] [CrossRef]
  5. Rashidi, M.M.; Shamekhi, L. Entropy Generation Analysis of the Revised Cheng-Minkowycz Problem for Natural Convective Boundary Layer Flow of Nanofluid in a Porous Medium. J. Thermal Sci. 2015, 19. [Google Scholar] [CrossRef]
  6. Rashidi, M.M.; Mohammadi, F. Entropy Generation Analysis for Stagnation Point Flow in a Porous Medium over a Permeable Stretching Surface. J. Appl. Fluid Mech. 2014, 8, 753–763. [Google Scholar]
  7. Rashidi, M.M.; Mahmud, S. Analysis of Entropy Generation in an MHD Flow over a Rotating Porous Disk with Variable Physical Properties. Int. J. Energy 2014. [Google Scholar] [CrossRef]
  8. Abolbashari, M.H.; Freidoonimehr, N. Analytical Modeling of Entropy Generation for Casson Nano-Fluid Flow Induced by a Stretching Surface. Adv. Powder Technol. 2015, 231. [Google Scholar] [CrossRef]
  9. Baag, S.S.R.; Dash, M.G.C.; Acharya, M.R. Entropy Generation Analysis for Viscoelastic MHD Flow over a Stretching Sheet Embedded in a Porous Medium. Ain Shams Eng. J. 2016, 23. [Google Scholar] [CrossRef]
  10. Shi, Z.; Tao, D. Entropy Generation and Optimization of Laminar Convective Heat Transfer and Fluid Flow in a Microchannel with Staggered Arrays of Pin Fin Structure with Tip Clearance. Energy Convers. Manag. 2015, 94, 493–504. [Google Scholar] [CrossRef]
  11. Hossein, A.M.; Ahmadi, M.A.; Mellit, A.; Pourfayaz, F.; Feidt, M. Thermodynamic Analysis and Multi Objective Optimization of Performance of Solar Dish Stirling Engine by the Centrality of Entransy and Entropy Generation. Int. J. Electr. Power Energy Syst. 2016, 78, 88–95. [Google Scholar] [CrossRef]
  12. Giovanni, G. Statistical Mechanics: A Short Treatise; Springer Science & Business Media: New York, NY, USA, 2013. [Google Scholar]
  13. Reif, F. Fundamentals of Statistical and Thermal Physics; Waveland Press: Long Grove, IL, USA, 2009. [Google Scholar]
  14. Rudolf, C.; Shimony, A. Two Essays on Entropy; University of California Press: Berkeley, CA, USA, 1977. [Google Scholar]
  15. Shu-Cherng, F.; Rajasekera, J.R.; Tsao, H.S.J. Entropy Optimization and Mathematical Programming; Springer Science & Business Media: New York, NY, USA, 2012. [Google Scholar]
  16. Robert, M.G. Entropy and Information Theory; Springer Science & Business Media: New York, NY, USA, 2013. [Google Scholar]
  17. Silviu, G. Information Theory with New Applications; MacGraw-Hill Books Company: New York, NY, USA, 1977. [Google Scholar]
  18. Shannon, C.E. A Mathematical Theory of Communication. SigmobileMob. Comput. Commun. Rev. 2001, 5, 3–55. [Google Scholar] [CrossRef]
  19. Thomas, M.C.; Thomas, J.A. Elements of Information Theory; John Wiley & Sons: Hoboken, NJ, USA, 2012. [Google Scholar]
  20. Christodoulos, A.F.; Pardalos, P.M. Encyclopedia of Optimization; Springer Science & Business Media: New York, NY, USA, 2008. [Google Scholar]
  21. Arash, M.; Azarnivand, A. Application of Integrated Shannon’s Entropy and VIKOR Techniques in Prioritization of Flood Risk in the Shemshak Watershed, Iran. Water Resour. Manag. 2015, 30, 409–425. [Google Scholar] [CrossRef]
  22. Liu, L.; Miao, S.; Liu, B. On Nonlinear Complexity and Shannon’s Entropy of Finite Length Random Sequences. Entropy 2015, 17, 1936–1945. [Google Scholar] [CrossRef]
  23. Karmeshu. Entropy Measures, Maximum Entropy Principle and Emerging Applications; Springer: Berlin, Germany; Heidelberg, Germany, 2012. [Google Scholar]
  24. Jaynes, E.T. Information Theory and Statistical Mechanics. II. Phys. Rev. 1957, 108, 171–190. [Google Scholar] [CrossRef]
  25. Jaynes, E.T. On the Rationale of Maximum-Entropy Methods. Proc. IEEE 1982, 70, 939–952. [Google Scholar] [CrossRef]
  26. Jaynes, E.T. Prior Probabilities. IEEE Trans. Syst. Sci. Cybernet. 1968, 4, 227–241. [Google Scholar] [CrossRef]
  27. Kapur, J.N. Maximum-Entropy Models in Science and Engineering; John Wiley & Sons: New York, NY, USA, 1989. [Google Scholar]
  28. Kapur, J.N. Maximum-Entropy Probability Distribution for a Continuous Random Variate over a Finite Interval. J. Math. Phys. Sci. 1982, 16, 693–714. [Google Scholar]
  29. Ray, A.; Majumder, S.K. Concavity of maximum entropy through modified Burg’s entropy subject to its prescribed mean. Int. J. Math. Oper. Res. 2016, 8. to appear. [Google Scholar]
  30. Kapur, J.N. Measures of Information and Their Applications; Wiley: New York, NY, USA, 1994. [Google Scholar]
  31. Burg, J. The Relationship between Maximum Entropy Spectra and Maximum Likelihood Spectra. Geophysics 1972, 37, 375–376. [Google Scholar] [CrossRef]
  32. Narain, K.J.; Kesavan, H.K. Entropy Optimization Principles with Applications; Academic Press: Cambridge, MA, USA, 1992. [Google Scholar]
  33. Solomon, K. Information Theory and Statistics; Courier Corporation: North Chelmsford, MA, USA, 2012. [Google Scholar]
  34. Amritansu, R.; Majumder, S.K. Derivation of some new distributions in statistical mechanics using maximum entropy approach. Yugoslav J. Oper. Res. 2013, 24, 145–155. [Google Scholar]
  35. Ulrych, T.J.; Bishop, T.N. Maximum Entropy Spectral Analysis and Autoregressive Decomposition. Rev. Geophys. 1975, 13, 183–200. [Google Scholar] [CrossRef]
  36. Michele, P.; Ferrante, A. On the Geometry of Maximum Entropy Problems. 2011. arXiv:1112.5529. [Google Scholar]
  37. Ke, J.-C.; Lin, C.-H. Maximum Entropy Approach to Machine Repair Problem. Int. J. Serv. Oper. Inform. 2010, 5, 197–208. [Google Scholar] [CrossRef]
Figure 1. Maximum Burg’s modified entropy.
Figure 1. Maximum Burg’s modified entropy.
Mathematics 04 00010 g001
Figure 2. Maximum Burg’s entropy.
Figure 2. Maximum Burg’s entropy.
Mathematics 04 00010 g002
Figure 3. Maximum Shannon’s entropy.
Figure 3. Maximum Shannon’s entropy.
Mathematics 04 00010 g003
Figure 4. Rectangular hyperbolic types of graphs. (a) when m n + 1 2 , μ 0 ; (b) when m n + 1 2 , μ 0 .
Figure 4. Rectangular hyperbolic types of graphs. (a) when m n + 1 2 , μ 0 ; (b) when m n + 1 2 , μ 0 .
Mathematics 04 00010 g004
Table 1. The a values & maximum values of Burg’s Modified Entropy.
Table 1. The a values & maximum values of Burg’s Modified Entropy.
a M a x P ( B mod ( P ) )
0.50.04082
1.00.11778
1.50.20294
2.00.28768
5.00.71376
10.01.1856
20.01.7512
30.02.1111
40.02.3754
50.02.5843
Table 2. Comparative Maximum Entropy Values.
Table 2. Comparative Maximum Entropy Values.
mMaximum Entropy Value of B mod ( P ) Maximum Entropy Value of B ( P ) Maximum Entropy Value of S ( P )
1.000−158.2250
1.250.0896122−45.135050.6255029
1.500.1345336−38.984780.9547543
1.750.1633092−35.431131.194875
2.000.1836655−32.946201.385892
2.250.1989097−31.052191.542705
2.500.2107459−29.537351.675885
2.750.2201939−28.289821.790029
3.000.2279397−27.243941.888477
3.250.2344146−26.358511.973503
3.500.2399041−25.606492.046725
3.750.24461−24.969582.109324
4.000.248682−24.435122.162186
4.250.252127−23.994192.205980
4.500.2549453−23.640472.241209
4.750.257137−23.369422.268253
5.000.2587024−23.177892.287386
5.250.2596416−23.063762.298794
5.500.259954623.025852.302585
5.750.2596416−23.063762.298794
6.000.2587024−23.177892.287386
6.250.257137−23.369422.268253
6.500.2549453−23.640472.241209
6.750.252127−23.994192.205980
7.000.248682−24.435122.162186
7.250.24461−24.969582.109324
7.500.2399041−25.606492.046725
7.750.2344146−26.358511.973503
8.000.2279397−27.243941.888477
8.250.2201939−28.289821.790029
8.500.2107459−29.537351.675885
8.750.1989097−31.052191.542705
9.000.1836655−32.946201.385892
9.250.1633092−35.431131.194875
9.500.1345336−38.984780.9547543
9.750.0896122−45.135050.6255029
10.00−141.252850
Table 3. p i and i graph.
Table 3. p i and i graph.
i p i
11.000000
20.2899575
30.1747893
40.1253783
50.1027619
60.1027620
70.1253782
80.1747891
90.2899574
101.000000

Share and Cite

MDPI and ACS Style

Ray, A.; Majumder, S.K. A Note on Burg’s Modified Entropy in Statistical Mechanics. Mathematics 2016, 4, 10. https://doi.org/10.3390/math4010010

AMA Style

Ray A, Majumder SK. A Note on Burg’s Modified Entropy in Statistical Mechanics. Mathematics. 2016; 4(1):10. https://doi.org/10.3390/math4010010

Chicago/Turabian Style

Ray, Amritansu, and S. K. Majumder. 2016. "A Note on Burg’s Modified Entropy in Statistical Mechanics" Mathematics 4, no. 1: 10. https://doi.org/10.3390/math4010010

APA Style

Ray, A., & Majumder, S. K. (2016). A Note on Burg’s Modified Entropy in Statistical Mechanics. Mathematics, 4(1), 10. https://doi.org/10.3390/math4010010

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop