Next Article in Journal
Towards Realising Secure and Efficient Image and Video Processing Applications on Quantum Computers
Next Article in Special Issue
Objective Bayesianism and the Maximum Entropy Principle
Previous Article in Journal
Maximum Entropy Production and Time Varying Problems: The Seasonal Cycle in a Conceptual Climate Model
Previous Article in Special Issue
Classification of Knee Joint Vibration Signals Using Bivariate Feature Distribution Estimation and Maximal Posterior Probability Decision Criterion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Relative Entropy Derivative Bounds

Universidad de los Andes, Facultad de Ingeniería y Ciencias Aplicadas, Monseñor Álvaro del Portillo 12455, Las Condes, Santiago, Chile
*
Author to whom correspondence should be addressed.
Entropy 2013, 15(7), 2861-2873; https://doi.org/10.3390/e15072861
Submission received: 24 May 2013 / Revised: 12 July 2013 / Accepted: 16 July 2013 / Published: 23 July 2013
(This article belongs to the Special Issue Maximum Entropy and Bayes Theorem)

Abstract

:
We show that the derivative of the relative entropy with respect to its parameters is lower and upper bounded. We characterize the conditions under which this derivative can reach zero. We use these results to explain when the minimum relative entropy and the maximum log likelihood approaches can be valid. We show that these approaches naturally activate in the presence of large data sets and that they are inherent properties of any density estimation process involving large numbers of random variables.

1. Introduction

Given a large ensemble of i.i.d.random variables, all of them generated according to some common density function, the asymptotic equipartition principle [1] guarantees that only a fraction of the possible ensembles, which is called the typical set, gathers almost all the probability ([2] [p. 226]). This fact opens interesting possibilities: What if the properties of the typical set impose conditions on any estimation process, such that the parameter search is focused on just a small subset of the available parameter set? We explore this using generalizations of the Shannon differential entropy [1] and the Fisher information [3,4,5] under typical set considerations. There are other works that take into account both concepts [6,7,8]. However, this work focuses on something new: defining algebraic bounds on the behavior of the derivative of the relative entropy with respect to their parameters. Furthermore, we characterize the conditions under which seeking for the extreme of this expression is a valid approach. Most importantly, we prove that these conditions automatically activate if large data sets are used. We finish this work discussing the relation between these bounds and the density function estimation problem. Under the conditions that activate these bounds, we characterize when the minimum relative entropy and maximum log likelihood approaches render the same solution. Furthermore, we show that these equivalences become true by the sole fact of having a large number of random variables.

2. Known Density Functions Case

Let us assume for the ensuing discussion that there exists a known source density function, f X , with support, S . We also know a set of i.i.d. samples, { x } n { x 1 , , x n } , that was generated by f X . There is also another density function, f Y | θ ( y ) , with the same support, S . This density function is indexed by the parameter vector, θ Θ R m .
The first part of this work deals with this case, where all the density functions are known. The second part of this work studies the density function estimation problem, where the source density function, f X , is not known.

3. Mixed Entropy Derivative: A Proxy for the Relative Entropy Derivative

We present the following definition:
Definition 1 
The relative entropy ([2][p. 231]), also called Kullback-Leibler divergence, is defined by:
D R E f X f Y | θ S f X ( u ) ln f X ( u ) f Y | θ ( u ) d u
Notice that lim u 0 + u ln u = 0 . Even though the definition is valid for any pair of density functions, we are using those useful for the ensuing analysis. The domain of this integral, and that of all the integrals in this work, corresponds to the support, S .
Definition 2 
The mixed entropy, h M ( f X , f Y | θ ) of two density functions is defined by:
h M ( f X , f Y | θ ) - S f X ( u ) ln f Y | θ ( u ) d u
when the integral exists.
Definition 3 
The Shannon differential entropy [1] is defined by:
h S ( f X ) - S f X ( u ) ln f X ( u ) d u
From these definitions, when f Y | θ = f X , from a functional point of view, the mixed entropy, h M ( f X , f Y | θ ) , equals the differential entropy, h S ( f X )
From examination of the relative entropy definition:
D K L f X f Y | θ = f X ( u ) ln f X ( u ) d u - f X ( u ) ln f Y | θ ( u ) d u (4) = - h S ( f X ) + h M ( f X , f Y | θ )
Hence:
D K L f X f Y | θ θ k = h M ( f X , f Y | θ ) θ k
In the previous equation it was assumed that the θ k are defined in the interior of Θ for k { 1 , , m } . In this work, where we assume that f X does not depend on any θ k , we use extensively the fact that the relative entropy derivative is equal to the mixed entropy derivative. Hence, in the following, we focus on studying the properties of the mixed entropy in order to be able to say something about the properties of the relative entropy derivatives.

4. Mixed Entropy Typical Set

An interesting insight related to sequences of i.i.d. random variables, discovered by Shannon [1], is the usefulness of the weak law of large numbers to characterize large ensembles of random variables. In the context of this work, this law implies:
Defining
Definition 4 
The likelihood, f { Y } n | θ , is defined by:
f { Y } n | θ k = 1 n f Y | θ ( x k )
we can state:
Theorem 1 
For any positive real number, ε, and fixed parameter vector, θ :
P - 1 n ln f { Y } n | θ - h M ( f X , f Y | θ ) ε n 1
Proof: Use that the random variables are i.i.d., and the weak law of large numbers. ☐
Hence, it makes sense to define the following set ([2][p. 226]):
Definition 5 
For any positive real number, ε, fixed parameter vector, θ , and any n 1 , the mixed entropy typical set, M ε n , with respect to the density function, f X , is defined by:
M ε n = { x } n S n : - 1 n ln f { Y } n | θ - h M ( f X , f Y | θ ) ε
Furthermore, using the weak law of large numbers, it is possible to prove:
Lemma 1 
For any positive real number, ε, and any parameter vector, θ , it is true that:
P { x } n M ε n n 1
Proof: Use the weak law of large numbers in the mixed entropy typical set definition. ☐
This proves that for large values of n, almost all the probability is contained by the mixed entropy typical set and that the probability of being outside this set becomes negligible.
Then, it is possible to prove that:
Lemma 2 
For any positive real number, ε, and any fixed parameter vector, θ , it exists n 0 n 0 ( ε , θ ) , such that for all n > n 0 , the total probability of the typical set is lower bounded by:
1 - ε P M ε n
Proof: Check ([2][p. 226], [9][p. 118]) for details. ☐

5. Micro-Differences in Mixed Entropy Typical Sets

Assuming that { x } n M ε n and from the definition of mixed entropy typical set:
ln f { Y } n | θ ] - n h M ( f X , f Y | θ ) - n ε , - n h M ( f X , f Y | θ ) + n ε [
The width of this interval is 2 ε n . This allows us to compare ln f { Y } n | θ with - n h M ( f X , f Y | θ ) , the value that is obtained in the limit thanks to the weak law of large numbers. Their ratio is:
2 ε n - n h M ( f X , f Y | θ ) = 2 ε h M ( f X , f Y | θ )
In other words, given that ε can be chosen as small as needed, the range of values that ln f { Y } n | θ can take may be indistinguishable from - n h M ( f X , f Y | θ ) , with high probability. This is another way of stating the weak law of large numbers.
The previous facts allow us to present the following definition:
Definition 6 
When { x } n M ε n , its associated micro-difference is defined by the following function:
δ δ ( f X , n , { x } n , f Y | θ , θ , ε ) ] 0 , 1 [
such that:
ln f { Y } n | θ = - n ( h M ( f X , f Y | θ ) + ε ) + 2 ε n δ
hence:
δ = 1 2 ε n ln f { Y } n | θ + n ( h M ( f X , f Y | θ ) + ε )
The analysis performed in the preceding paragraphs shows that the behavior of the micro-differences might be negligible in the case of large values of n. However, the following sections of this work deal with the the derivative of the micro-difference value with respect to the parameters; so, we cannot neglect it, and we need to study its behavior.
From Equation (14):
ln f { Y } n | θ θ k = - n h M ( f X , f Y | θ ) θ k + 2 ε n δ θ k
for all k [ 1 , , m ] . Thus:
2 ε δ θ k = 1 n ln f { Y } n | θ θ k + h M ( f X , f Y | θ ) θ k
= θ k 1 n ln f { Y } n | θ + h M ( f X , f Y | θ )
= - θ k - 1 n ln f { Y } n | θ - h M ( f X , f Y | θ )
This last expression helps us to understand that the derivative of δ with respect to the parameters is related to changes experienced by quantities upper bounded by ε.

6. Mixed Information

We define:
Definition 7 
The mixed information, i F f { X } n , f { Y } n | θ k , is defined by:
i F f { X } n , f { Y } n | θ k S n f { X } n ln f { Y } n | θ θ k 2 d u
with:
f { X } n k = 1 n f X ( x k )
when the integral exists.

7. Mixed Entropy Derivative Bounds

The main contribution of this work is the following pair of inequalities, which are obtained thanks to a combination of the mixed information and the mixed entropy.
Theorem 2 
Given a positive real number, ε, any parameter vector, θ , and an i.i.d. sequence with n elements, then the components of θ h M ( f X , f Y | θ ) comply with:
I R k - 1 n i F f { X } n , f { Y } n | θ k I L h M ( f X , f Y | θ ) θ k I R k + 1 n i F f { X } n , f { Y } n | θ k I L
where:
I R k M ε n f { X } n 2 ε δ θ k 2 d u
I L M ε n f { X } n d u
Proof: From the mixed information definition:
i F f { X } n , f { Y } n | θ k = I M ε n k + I M ε n k I M ε n k > 0
where:
I M ε n k M ε n f { X } n ln f { Y } n | θ θ k 2 d u
and the symbol, ∼, denotes the complement set.
Replacing Equation (16) in Equation (26), it is obtained:
I M ε n k = n 2 h M ( f X , f Y | θ ) θ k 2 M ε n f { X } n d u - 2 n h M ( f X , f Y | θ ) θ k M ε n f { X } n 2 ε n δ θ k d u (27) + M ε n f { X } n 2 ε n δ θ k 2 d u
Using the Cauchy-Bunyakovsky-Schwarz inequality:
M ε n f { X } n 2 ε n δ θ k d u 2 = M ε n f { X } n f { X } n 2 ε n δ θ k d u 2 M ε n f { X } n d u · M ε n f { X } n 2 ε n δ θ k 2 d u (28) = I L · n 2 I R k
Therefore:
I M ε n k n 2 h M ( f X , f Y | θ ) θ k 2 I L - 2 n 2 h M ( f X , f Y | θ ) θ k I L I R k + n 2 I R k (29) = n 2 h M ( f X , f Y | θ ) θ k I L - I R k 2
which implies:
1 n 2 i F f { X } n , f { Y } n | θ k h M ( f X , f Y | θ ) θ k I L - I R k 2
Hence, using the negative solution of the square root:
1 n i F f { X } n , f { Y } n | θ k - h M ( f X , f Y | θ ) θ k I L + I R k
Therefore:
I R k - 1 n i F f { X } n , f { Y } n | θ k I L h M ( f X , f Y | θ ) θ k
Taking into account the positive solution of Equation (30):
h M ( f X , f Y | θ ) θ k I R k + 1 n i F f { X } n , f { Y } n | θ k I L
The previous result allows us to state the following remarks:
Remark 1 
This theorem states algebraic bounds on the derivative of the mixed entropy that are valid for any value of n.
Remark 2 
The expression:
I R k M ε n f { X } n 2 ε δ θ k 2 d u = M ε n f { X } n - θ k - 1 n ln f { Y } n | θ - h M ( f X , f Y | θ ) 2 d u
refers to the accumulation of derivatives of very small differences. This differences become smaller and smaller as the weak law of large numbers is enacted.

8. Unknown Source Density Function, f X

Now, we study what happens when the source density function, f X , is not known.
It can be proven that D K L f X f Y | θ 0 ([2][p. 232]). Thus, from the corresponding definitions given at the start of this work:
h M ( f X , f Y | θ ) h S ( f X )
Therefore, the density estimation problem can be framed as the following optimization program:
θ θ ( f X , f Y | θ ) = arg min θ h M ( f X , f Y | θ )
Remark 3 
The solutions of this optimization program pose the following options:
  • h M ( f X , f Y | θ ) = h S ( f X ) , hence f Y | θ = f X .
  • h M ( f X , f Y | θ ) > h S ( f X ) , thus f Y | θ f X . This happens when the estimator cannot implement the desired density function, which is beyond its reach. One way of checking whether the global minimum has been reached is to compare the mixed entropy value to that of the Shannon differential entropy. If they differ, the global minimum has not been reached yet.
According to this result, it is straightforward to search for parameters that make true the following expression:
θ h M ( f X , f Y | θ ) θ = θ = 0
However, this is out of the limits in our density estimation problem, given that we do not have access to f X . Thank to the weak law of large numbers:
lim n - 1 n ln f { Y } n | θ = h M ( f X , f Y | θ )
in probability. Thus, for sufficiently large values of n:
- 1 n ln f { Y } n | θ h M ( f X , f Y | θ )
Hence, the optimization program posed in Equation (36) can be approximated by:
θ θ ( n , { x } n , f Y | θ ) = arg min θ - 1 n ln f { Y } n | θ
or:
θ θ ( n , { x } n , f Y | θ ) = arg max θ 1 n ln f { Y } n | θ
which is the maximum log likelihood framework ([3][p. 261], [4][260], [5][p. 65]).
Remark 4 
As in the mixed entropy case, the maximum log likelihood framework can exhibit the following cases:
  • The optimization program finds a parameter combination that makes the estimator equal to the unknown source density function, thus making true the following expression for a sufficiently large n:
    - 1 n ln f { Y } n | θ h M ( f X , f Y | θ ) h S ( f X )
  • The estimator does not find the desired target function. This happens when the family of functions that can be implemented by the estimator does not include the target function. It also happens when the maximum log likelihood program is riddled with multiple local minima, and the optimization process gets trapped in one of them. Only under very specific conditions, it is possible to obtain the global maximum of the maximum log likelihood problem [10]. In this case, we cannot use a comparison between the value of the log likelihood term and that of the Shannon differential entropy to determine the goodness of our solution, as we could do in the case of the mixed entropy, because calculating the latter is analogous to the very problem we are trying to solve.
Hence, it is natural to work with this approximated program and look for parameters that make the following expression true:
θ 1 n ln f { Y } n | θ θ = θ = 0

9. An Example: Mismatched Density Functions

9.1. Known Density Functions Case

Let us define a source density function:
f X ( x ) = λ e - λ x H ( x )
with x R and H ( x ) , the Heaviside step function. This density function is completely defined by the parameter, λ. Its differential entropy is given by:
h S ( f X ) = 1 - ln λ = 1 + ln 1 λ
We also have a data set, { x } n , generated by the source density function.
The density function that depends on parameters is:
f Y | μ , σ = 1 σ 2 π exp - μ - x 2 2 σ 2
with x R , μ its mean and σ its standard deviation. Its differential entropy is given by:
h S ( f Y | μ , σ ) = ln σ 2 π e = 1 + ln σ + ln 2 π e
which does not depend on the mean, μ.

9.2. Compliance with Theorem 2

We first analyze the bounds associated with parameter μ. For these density functions:
1 n ln f { Y } n | μ , σ = 1 n k = 1 n ln f Y | μ , σ ( x k ) = ln 1 σ 2 π - 1 2 n σ 2 k = 1 n μ - x k 2
Hence:
1 n ln f { Y } n | μ , σ μ = - 1 σ 2 μ - 1 n k = 1 n x k
We calculate the mixed information for μ:
i F f { X } n , f { Y } n | μ , σ μ = R + n f { X } n ln f { Y } n | μ , σ μ 2 d u = R + n f { X } n n σ 2 μ - 1 n k = 1 n u k 2 d u (50) = n 2 σ 4 R + n f { X } n μ - 1 n k = 1 n u k 2 d u
Hence:
1 n i F f { X } n , f { Y } n | μ , σ μ = 1 σ 2 R + n f { X } n μ - 1 n k = 1 n u k 2 d u
Also:
h M ( f X , f Y | μ , σ ) = - - f X ( u ) ln f Y | μ , σ ( u ) d u = - - λ e - λ u H ( u ) ln f Y | μ , σ ( u ) d u = - 0 λ e - λ u ln 1 σ 2 π - μ - u 2 2 σ 2 d u = ln σ 2 π 0 λ e - λ u d u + λ 2 σ 2 0 μ - u 2 e - λ u d u (52) = ln σ 2 π + 1 σ 2 μ 2 2 - μ λ + 1 λ 2
and:
h M ( f X , f Y | μ , σ ) μ = 1 σ 2 μ - 1 λ
Finally, we calculate the micro-structure term using Equation (17):
2 ε δ μ = 1 σ 2 1 n k = 1 n x k - 1 λ
This allows us to calculate:
I R μ = M ε n f { X } n 2 ε δ μ 2 d u = M ε n f { X } n 1 σ 2 1 n k = 1 n u k - 1 λ 2 d u (55) = 1 σ 4 M ε n f { X } n 1 n k = 1 n u k - 1 λ 2 d u
Now, for σ, we have:
1 n ln f { Y } n | μ , σ σ = 1 σ 3 - σ 2 + 1 n k = 1 n μ - u k 2
The mixed information for σ is defined by:
i F f { X } n , f { Y } n | μ , σ μ = R + n f { X } n ln f { Y } n | μ , σ σ 2 d u = R + n f { X } n n σ 3 - σ 2 + 1 n k = 1 n μ - u k 2 2 d u (57) = n 2 σ 6 R + n f { X } n - σ 2 + 1 n k = 1 n μ - u k 2 2 d u
Hence:
1 n i F f { X } n , f { Y } n | μ , σ μ = 1 σ 3 R + n f { X } n - σ 2 + 1 n k = 1 n μ - u k 2 2 d u
Again, the derivative of the mixed entropy with respect to σ:
h M ( f X , f Y | μ , σ ) σ = 1 σ 3 σ 2 - 1 λ 2 - μ - 1 λ 2
Hence, the associated micro-differences expression is:
2 ε δ σ = 1 σ 3 1 n k = 1 n μ - u k 2 - 1 λ 2 - μ - 1 λ 2
which allows us to calculate:
I R σ = M ε n f { X } n 2 ε δ σ 2 d u = M ε n f { X } n 1 σ 3 1 n k = 1 n μ - u k 2 - 1 λ 2 - μ - 1 λ 2 2 d u (61) = 1 σ 6 M ε n f { X } n 1 n k = 1 n μ - u k 2 - 1 λ 2 - μ - 1 λ 2 2 d u

9.3. Unknown Source Density Function, f X

If f X is unknown, then it is not possible to calculate the previous bounds. In this case, we resort to the optimization program specified in Equation (41). Hence, we use Equations (49) and (56), in order to obtain:
μ n = 1 n k = 1 n x k
σ n 2 = 1 n k = 1 n μ n - x k 2
Using this choice of values immediately makes both mixed information expressions described by Equations (51) and (58) equal to zero, without the need for large values of n.
Furthermore, if we use:
μ = lim n μ n = 1 λ
σ = lim n σ n = 1 λ
then the micro-differences integrals also become equal to zero.
This example shows several things:
  • The use of the maximum log likelihood solutions and large values of n guarantees that the derivative of the mixed entropy equals zero; hence, the derivative of the relative entropy with respect to its parameters is zero, too.
  • However, this is not enough to guarantee that the density functions will match each other (a Gaussian cannot estimate an exponential). The maximum log likelihood solution only guarantees that the estimator will do the best it can do.
  • If not enough data is available, the derivative of the mixed entropy will be different from zero, because the micro-differences terms are not zero.

10. Discussion

If one wanted to estimate a density function out of a data set, one could minimize the relative entropy: once we reach its minimum, we can guarantee that the density function implemented by the estimator will be equal to that which originated the data. This is true only if the family of functions that the estimator can effectively implement includes that of the source density function. Moreover, realizing that it is not possible to measure the relative entropy, one would quickly settle to maximize the log likelihood and obtain the desired density function. Keeping this in mind, the results presented in this work seem just another way of saying the same. However, this is not so. The difference is subtle. Whereas there are many density function estimation principles and it could be discussed whether minimizing the relative entropy is the most convenient one, the bounds presented in this work show that minimizing the relative entropy automatically activates in the presence of large data sets and that one does not need to decide whether that estimation framework is the most convenient or not. Something similar happens with the weak law of large numbers: there are many ways of determining the expected value of the density function that generated the data, but everybody knows that the weak law of large numbers guarantees that in the presence of large data sets, this value can be effectively approximated by the average of the values. Our bounds, also based on the weak law of large numbers, state something similar: in the presence of large data sets, the source density function is perceived as the solution of the maximum log likelihood solution.

11. Conclusions

Theorem 2, which is the main result of this work, consists in a set of bounds on the partial derivative of the mixed entropy, which is a proxy to the relative entropy, with respect to the parameters of the estimator. These bounds relate the mixed information value, the capacity of the estimator to produce micro-differences and the number of random variables present in the analysis in such a way that it is possible to determine that these bounds activate automatically when the data set is large. If the micro-differences integral is zero, the mixed information derivative is zero for large data sets, independently of any other considerations. In other words, minimizing the mixed entropy is the preferred density estimation framework when large data sets are available.
Given the impossibility of estimating the mixed entropy, it is shown that the correct numerical approximation to this problem is finding the solution to the maximum log likelihood problem. Again, when the estimator is associated to a micro-differences integral equal to zero, this framework becomes the optimal one in the presence of large data sets.

Acknowledgments

The authors thank Jaime Cisternas, Jorge Silva and Jose Principe for their helpful insights about this work.

Conflict of Interest

The authors declare no conflict of interest.

References

  1. Shannon, C. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  2. Cover, T.; Thomas, J. Elements of Information Theory; John Wiley and Sons, Inc.: Hoboken, NJ, USA, 1991. [Google Scholar]
  3. Hogg, R.V.; Craig, A.T. Introduction to Mathematical Statistics; Prentice Hall: Upper Saddle River, NJ, USA, 1995. [Google Scholar]
  4. Papoulis, A. Probability, Random Variables, and Stochastic Processes; McGraw-Hill: New York, NY, USA, 1991. [Google Scholar]
  5. Van Trees, H.L. Detection, Estimation, and Modulation Theory: Part 1; John Wiley and Sons, Inc.: Hoboken, NJ, USA, 2001. [Google Scholar]
  6. Vignat, C.; Bercher, J. Analysis of signals in the Fisher-Shannon information plane. Phys. Lett. A 2003, 312. [Google Scholar] [CrossRef]
  7. Romera, E.; Dehesa, J. The Fisher-Shannon information plane, an electron correlation tool. J. Chem. Phys. 2004, 120, 8906–8912. [Google Scholar] [CrossRef] [PubMed]
  8. Dimitrov, V. On Shannon-Jaynes Entropy and Fisher Information. In Proceedings of the 27th International Workshop on Bayesian Inference and Maximum Entropy Methods in Science and Engineering, Saratoga Springs, New York, NY, USA, 8–13 July 2007.
  9. Taubman, D.; Marcellin, M. JPEG2000: Image Compression Fundamentals, Standards, and Practice; Kluwer Academic Publishers: Dordrecht, The Netherlands, 2002. [Google Scholar]
  10. Boyd, S.; Vandenberghe, L. Convex Optimization; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar]

Share and Cite

MDPI and ACS Style

Zegers, P.; Fuentes, A.; Alarcón, C. Relative Entropy Derivative Bounds. Entropy 2013, 15, 2861-2873. https://doi.org/10.3390/e15072861

AMA Style

Zegers P, Fuentes A, Alarcón C. Relative Entropy Derivative Bounds. Entropy. 2013; 15(7):2861-2873. https://doi.org/10.3390/e15072861

Chicago/Turabian Style

Zegers, Pablo, Alexis Fuentes, and Carlos Alarcón. 2013. "Relative Entropy Derivative Bounds" Entropy 15, no. 7: 2861-2873. https://doi.org/10.3390/e15072861

APA Style

Zegers, P., Fuentes, A., & Alarcón, C. (2013). Relative Entropy Derivative Bounds. Entropy, 15(7), 2861-2873. https://doi.org/10.3390/e15072861

Article Metrics

Back to TopTop