Next Article in Journal
Adaptive Computation of Multiscale Entropy and Its Application in EEG Signals for Monitoring Depth of Anesthesia During Surgery
Previous Article in Journal
On Chirality of the Vorticity of the Universe
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Some Further Results on the Minimum Error Entropy Estimation

1
Department of Precision Instruments and Mechanology, Tsinghua University, Beijing, 100084, China
2
Department of Electrical and Computer Engineering, University of Florida, Gainesville, FL 32611, USA
*
Author to whom correspondence should be addressed.
Entropy 2012, 14(5), 966-977; https://doi.org/10.3390/e14050966
Submission received: 1 April 2012 / Revised: 2 May 2012 / Accepted: 10 May 2012 / Published: 21 May 2012

Abstract

:
The minimum error entropy (MEE) criterion has been receiving increasing attention due to its promising perspectives for applications in signal processing and machine learning. In the context of Bayesian estimation, the MEE criterion is concerned with the estimation of a certain random variable based on another random variable, so that the error’s entropy is minimized. Several theoretical results on this topic have been reported. In this work, we present some further results on the MEE estimation. The contributions are twofold: (1) we extend a recent result on the minimum entropy of a mixture of unimodal and symmetric distributions to a more general case, and prove that if the conditional distributions are generalized uniformly dominated (GUD), the dominant alignment will be the MEE estimator; (2) we show by examples that the MEE estimator (not limited to singular cases) may be non-unique even if the error distribution is restricted to zero-mean (unbiased).
MSC Codes:
62B10

1. Introduction

A central concept in information theory is entropy, which is a mathematical measure of the uncertainty or the amount of missing information [1]. Entropy has been widely used in many areas, including physics, mathematics, communication, economics, signal processing, machine learning, etc. The maximum entropy principle is a powerful and widely accepted method for statistical inference or probabilistic reasoning with incomplete knowledge of probability distribution [2]. Another important entropy principle is the minimum entropy principle, which decreases the uncertainty associated with a system. In particular, the minimum error entropy (MEE) criterion can be applied in problems like estimation [3,4,5], identification [6,7], filtering [8,9,10], and system control [11,12]. In recent years, the MEE criterion, together with the nonparametric Renyi entropy estimator, has been successfully used in information theoretic learning (ITL) [13,14,15].
In the scenario of Bayesian estimation, the MEE criterion aims to minimize the entropy of the estimation error, and hence decrease the uncertainty in estimation. Given two random variables: X n , an unknown parameter to be estimated, and Y m , the observation (or measurement), the MEE estimation of X based on Y can be formulated as:
g * = arg min g G H ( X g ( Y ) )      = arg min g G E [ log p g ( X g ( Y ) ) ]      = arg min g G n p g ( x ) log p g ( x ) d x
where g ( Y ) denotes an estimator of X based on Y , g is a measurable function, G stands for the collection of all measurable functions of Y , H ( X g ( Y ) ) denotes the Shannon entropy of the estimation error X g ( Y ) , and p g ( x ) denotes the probability density function (PDF) of the estimation error. Let p ( x | y ) be the conditional PDF of X given Y . Then:
p g ( x ) = m p ( x + g ( y ) | y ) d F ( y )
where F ( y ) denotes the distribution function of Y . From (2), one can see the error PDF p g ( x ) is actually a mixture of the shifted conditional PDF.
Different from conventional Bayesian risks, like mean square error (MSE) and risk-sensitive cost [16], the “loss function” in MEE is log p g ( . ) , which is directly related to the error’s PDF, transforming nonlinearly the error by its own PDF. Some theoretical aspects of MEE estimation have been studied in the literature. In an early work [3], Weidemann and Stear proved that minimizing the error entropy is equivalent to minimizing the mutual information between the error and the observation, and also proved that the reduced error entropy is upper-bounded by the amount of information obtained by the observation. In [17], Janzura et al. proved that, for the case of finite mixtures ( Y is a discrete random variable with finite possible values), the MEE estimator equals the conditional median provided that the conditional PDFs are conditionally symmetric and unimodal (CSUM). Otahal [18] extended Janzura’s results to finite-dimensional Euclidean space. In a recent paper, Chen and Geman [19] employed a “function rearrangement” to study the minimum entropy of a mixture of CSUM distributions where no restriction on Y was imposed. More recently, Chen et al. have investigated the robustness, non-uniqueness (for singular cases), sufficient condition, and the necessary condition involved in the MEE estimation [20]. Chen et al. have also presented a new interpretation on the MSE criterion as a robust MEE criterion [21].
In this work, we continue the study on the MEE estimation, and obtain some further results. Our contributions are twofold. First, we extend the results of Chen and Geman to a more general case, and show that when the conditional PDFs are generalized uniformly dominated (GUD), the MEE estimator equals the dominant alignment. Second, we show by examples that, the unbiased MEE estimator (not limited to singular cases) may be non-unique, and there can even be infinitely many optimal solutions. The rest of the paper is organized as follows. In Section 2, we study the minimum entropy of a mixture of generalized uniformly dominated conditional distributions. In Section 3, we present two examples to show the non-uniqueness of the unbiased MEE estimation. Finally, we give our conclusions in Section 4.

2. MEE Estimator for Generalized Uniformly Dominated Conditional Distributions

Before presenting the main theorem of this section, we give the following definitions.
Definition 1: Let = { f t ( x ) : n + , t T } be a set of nonnegative, integrable functions, where T denotes an index set (possibly uncountable). Then is said to be uniformly dominated in x n if and only if ε [ 0 , ) , there exists a measurable set D ε n , satisfying λ ( D ε ) = ε and:
D ε f t ( x ) d x = sup A : λ ( A ) = ε A f t ( x ) d x , t T
where λ is Lebesgue measure. The set D ε is called the ε -volume dominant support of .
Definition 2: The nonnegative, integrable function set is said to be generalized uniformly dominated (GUD) in x n if and only if there exists a function π : T n , such that π is uniformly dominated, where:
π = { f t π : f t π ( x ) = f t ( x + π ( t ) ) , t T }
The function π ( t ) is called the dominant alignment of .
Remark 1: The dominant alignment is, obviously, non-unique. If π ( t ) is a dominant alignment of , then c n , π ( t ) + c will also be a dominant alignment of .
When regarding y as an index parameter, the conditional PDF p ( x | y ) will represent a set of nonnegative and integrable functions, that is:
𝒫 = { p ( x | y ) : n + , y m }
If the above function set is (generalized) uniformly dominated in x n , then we say that the conditional PDF p ( x | y ) is (generalized) uniformly dominated in x n .
Remark 2: The GUD is much more general than CSUM. Actually, if the conditional PDF is CSUM, it must also be GUD (with the conditional mean as the dominant alignment), but not vice versa. In Figure 1 we show two examples where two PDFs (solid and dotted lines) are uniformly dominated but not CSUM.
Figure 1. Uniformly dominated PDFs: (a) non-symmetric, (b) non-unimodal.
Figure 1. Uniformly dominated PDFs: (a) non-symmetric, (b) non-unimodal.
Entropy 14 00966 g001
Theorem 1: Assume the conditional PDF p ( x | y ) is generalized uniformly dominated in x n , with dominant alignment π ( y ) . If H ( X π ( Y ) ) exists (here “exists” means “exists in the extended sense” as defined in [19]) then H ( X π ( Y ) ) H ( X g ( Y ) ) for all g : m n for which H ( X g ( Y ) ) also exists.
Proof of Theorem 1: The proof presented below is similar to that of the Theorem 1 in [19], except that the discretization procedure is avoided. In the following, we give a brief sketch of the proof, and consider only the case n = 1 (the proof can be easily extended to n > 1 ). First, one needs to prove the following proposition.
Proposition 1: Assume that the function f ( x | y ) , x , y m (not necessarily a conditional PDF) satisfies the following conditions:
(1)
non-negative, continuous, and integrable in x for each y m ;
(2)
generalized uniformly dominated in x , with dominant alignment π ( y ) ;
(3)
uniformly bounded in ( x , y ) .
Then for any g : m , we have:
H ( f π ) H ( f g )
where H ( f ) = f ( x ) log f ( x ) d x (here we extend the entropy definition to nonnegative L 1 functions), and:
{ f π ( x ) = m f ( x + π ( y ) | y ) d F ( y ) f g ( x ) = m f ( x + g ( y ) | y ) d F ( y )
Remark 3: It is easy to verify that f π d x = f g d x sup ( x , y ) f ( x | y ) < (not necessarily f π d x = 1 ).
Proof of Proposition 1: The above proposition can be readily proved using the following two lemmas.
Lemma 1 [19]: Assume the nonnegative function h : [ 0 , ) is bounded, continuous, and integrable, and define the function O h ( z ) by ( λ is Lebesgue measure):
O h ( z ) = λ { x : h ( x ) z }
Then the following results hold:
(a)
Define m h ( x ) = sup { z : O h ( z ) x } , x ( 0 , ) , and m h ( 0 ) = sup x h ( x ) . Then m h ( x ) is continuous and non-increasing on [ 0 , ) , and m h ( x ) 0 as x .
(b)
For any function G : [ 0 , ) with | G ( h ( x ) ) | d x <
G ( h ( x ) ) d x = 0 G ( m h ( x ) ) d x
(c)
For any x 0 [ 0 , )
0 x 0 m h ( x ) d x = sup A : λ ( A ) = x 0 A h ( x ) d x
Proof of Lemma 1: See [19].
Remark 4: Denote m π = m f π , and m g = m f g . Then by Lemma 1, we have H ( m π )   = H ( f π ) and H ( m g ) = H ( f g ) (let G ( x ) = x log x ). Therefore, to prove Proposition 1, it suffices to prove:
H ( m π ) H ( m g )  
Lemma 2: Functions m π and m g satisfy:
(a)
0 m π ( x ) d x = 0 m g ( x ) d x <
(b)
0 ε m g ( x ) d x 0 ε m π ( x ) d x ,    ε [ 0 , )
Proof of Lemma 2: (a) comes directly from the fact f π d x = f g d x < . We only need to prove (b). We have:
sup A : λ ( A ) = ε A f g ( x ) d x = sup A : λ ( A ) = ε A m f ( x + g ( y ) | y ) d F ( y ) d x = sup A : λ ( A ) = ε m A f ( x + g ( y ) | y ) d x d F ( y ) m { sup A : λ ( A ) = ε A f ( x + g ( y ) | y ) d x } d F ( y ) = m { D ε f ( x + π ( y ) | y ) d x } d F ( y ) = D ε m f ( x + π ( y ) | y ) d F ( y ) d x = sup A : λ ( A ) = ε A m f ( x + π ( y ) | y ) d F ( y ) d x = sup A : λ ( A ) = ε A f π ( x ) d x
where D ε is the ε -volume dominant support of f ( x + π ( y ) | y ) . By Lemma 1 (c):
{ 0 ε m g ( x ) d x = sup A : λ ( A ) = ε A f g ( x ) d x 0 ε m π ( x ) d x = sup A : λ ( A ) = ε A f π ( x ) d x
Thus ε [ 0 , ) , we have 0 ε m g ( x ) d x 0 ε m π ( x ) d x :
We are now in position to prove (11):
H ( m g ) = 0 m g ( x ) log m g ( x ) d x             = 0 m g ( x ) 0 log m g ( x ) d y d x             = 0 ( 0 m g ( x ) I [ y log m g ( x ) ] d y ) d x             = 0 ( inf { x : log m g ( x ) y } m g ( x ) d x ) d y             L e m m a 2 0 ( inf { x : log m g ( x ) y } m π ( x ) d x ) d y             = 0 m π ( x ) log m g ( x ) d x             ( a ) 0 m π ( x ) log m π ( x ) d x             = H ( m π )  
where I [ . ] denotes the indicator function, and (a) follows from x > 0 , log x 1 x , that is:
0 m π ( x ) log m g ( x ) d x + 0 m π ( x ) log m π ( x ) d x = 0 m π ( x ) log ( m g ( x ) m π ( x ) ) d x 0 m π ( x ) ( 1 m g ( x ) m π ( x ) ) d x = 0 m π ( x ) d x 0 m g ( x ) d x = L e m m a 2 0
In the above proof, we adopt the convention 0 = 0 .
Now the proof of Proposition 1 has been completed. To finish the proof of Theorem 1, we have to remove the conditions of continuity and uniform boundedness imposed in Proposition 1. This can be easily accomplished by approximating p ( x | y ) by a sequence of functions { f n ( x | y ) } , n = 1 , 2 , , which satisfy these conditions. The remaining proof is omitted here, since it is exactly the same as the last part of the proof for Theorem 1 in [19].
Example 1: Consider an additive noise model:
X = φ ( Y ) + ξ
where ξ is an additive noise that is independent of Y . In this case, we have p ( x | y ) = p ξ ( x φ ( y ) ) , where p ξ ( . ) denotes the noise PDF. It is clear that p ( x | y ) is generalized uniformly dominated, with dominant alignment π ( y ) = φ ( y ) . According to Theorem 1, we have H ( X φ ( Y ) ) H ( X g ( Y ) ) . In fact, this result can also be proved by:
H ( X φ ( Y ) ) = H ( ξ )        ( b ) H ( ξ + ( φ ( Y ) g ( Y ) ) )        = H ( X g ( Y ) )
where (b) comes from the fact that ξ and ( φ ( Y ) g ( Y ) ) are independent (For independent random variables X and Y , the inequality H ( X + Y ) max { H ( X ) , H ( Y ) } holds). In this example, the conditional PDF p ( x | y ) is, obviously, not necessarily CSUM.
Example 2: Suppose the joint PDF of random variables X , Y ( X , Y ) is:
p ( x , y ) = exp ( 1 x exp ( y ) )
where y 0 ,   x exp ( y ) 1 . Then the conditional PDF p ( x | y ) will be:
p ( x | y ) = p ( x , y ) p ( y )            = exp ( 1 x exp ( y ) )   exp ( y ) exp ( 1 x exp ( y ) ) d x            = exp ( 1 x exp ( y ) + y )
One can easily verify that the above conditional PDF is non-symmetric but generalized uniformly dominated, with dominant alignment π ( y ) = exp ( y ) (the ε-volume dominant support of p ( x + π ( y ) | y ) is D ε = [ 0 , ε ] ). By Theorem 1, the function exp ( y ) is the minimizer of error entropy.

3. Non-Uniqueness of Unbiased MEE Estimation

Because entropy is shift-invariant, the MEE estimator is obviously non-unique. In practical applications, in order to yield a unique solution, or to meet the desire for small error values, the MEE estimator is usually restricted to be unbiased, that is, the estimation error is restricted to be zero-mean [15]. The question of interest in this paper is whether the unbiased MEE estimator is unique. In [20], it has been shown that, for the singular case (in which the error entropy approaches minus infinity), the unbiased MEE estimation may yield non-unique (even infinitely many) solutions. In the following, we present two examples to show that this result still holds even for nonsingular case.
Example 3: Let the joint PDF of X and Y ( X , Y ) be a mixed-Gaussian density [20]:
p ( x , y ) = 1 4 π 1 ρ 2 { exp { ( y 2 2 ρ y ( x μ ) + ( x μ ) 2 ) 2 ( 1 ρ 2 ) } + exp { ( y 2 + 2 ρ y ( x + μ ) + ( x + μ ) 2 ) 2 ( 1 ρ 2 ) } }
where μ > 0 , 0 | ρ | < 1 . The conditional PDF of X given Y will be:
p ( x | y ) = 1 2 2 π ( 1 ρ 2 ) { exp ( ( x μ ρ y ) 2 2 ( 1 ρ 2 ) ) + exp ( ( x + μ + ρ y ) 2 2 ( 1 ρ 2 ) ) }
y , p ( x | y ) is symmetric around zero (but not unimodal in x). It can be shown that for some values of μ , ρ , the MEE estimator of X based on Y does not equal zero (see [20], Example 3). In these cases, the MEE estimator will be non-unique, even if the error’s PDF is restricted to zero-mean (unbiased) distribution. This can be proved as follows:
Let g * be an unbiased MEE estimator of X based on Y . Then g * will also be an unbiased MEE estimator, because:
p g * ( x ) = p ( x + g * ( y ) | y ) d F ( y )            = ( c ) p ( x g * ( y ) | y ) d F ( y )            = p g * ( x )
where (c) comes from the fact that p ( x | y ) is symmetric around zero, and further:
H ( p g * ) = p g * ( x ) log p g * ( x ) d x              = p g * ( x ) log p g * ( x ) d x              = p g * ( x ) log p g * ( x ) d x              = H ( p g * )
If the unbiased MEE estimator is unique, then we have g * = g * g * = 0 , which contradicts the fact that g * does not equal zero. Therefore, the unbiased MEE estimator must be non-unique. Obviously, the above result can be extended to more general cases. In fact, we have the following proposition.
Proposition2: The unbiased MEE estimator will be non-unique if the conditional PDF p ( x | y ) satisfies:
(1)
Symmetric in x n around the conditional mean μ ( y ) = E [ X | Y = y ] for each y m ;
(2)
There exists a function ζ : m n such that ζ ( Y ) μ ( Y ) , H ( X ζ ( Y ) ) H ( X μ ( Y ) ) .
Proof: Similar to the proof presented above (Omitted).
In the next example, we show that, for some particular situations, there can be even infinitely many unbiased MEE estimators.
Example 4: Suppose Y is a discrete random variable with Bernoulli distribution:
Pr ( Y = 0 ) = Pr ( Y = 1 ) = 0.5
The conditional PDF of X given Y is (see Figure 2):
p ( x | Y = 0 ) = { 1 2 a     i f   | x | a 0        o t h e r
p ( x | Y = 1 ) = { 1 a     i f   | x | 1 2 a 0        o t h e r
where a > 0 . Note that the above conditional PDF is uniformly dominated in x .
Figure 2. Conditional PDF of X given Y : (a) p ( x | Y = 0 ) , (b) p ( x | Y = 1 ) .
Figure 2. Conditional PDF of X given Y : (a) p ( x | Y = 0 ) , (b) p ( x | Y = 1 ) .
Entropy 14 00966 g002
Given an estimator X ^ = g ( Y ) , the error’s PDF will be:
p g ( x ) = 1 2 { p ( x + g ( 0 ) | Y = 0 ) + p ( x + g ( 1 ) | Y = 1 ) }
Let g be an unbiased estimator, then x p g ( x ) d x = 0 , and hence g ( 0 ) = g ( 1 ) . In the following, we assume g ( 0 ) 0 (due to symmetry, one can obtain similar results for g ( 0 ) < 0 ), and consider three cases:
Case 1: 0 g ( 0 ) 1 4 a . In this case, the error PDF is:
p g ( x ) = { 1 4 a ,        a g ( 0 ) x < a 2 + g ( 0 ) 3 4 a ,          a 2 + g ( 0 ) x a 2 + g ( 0 ) 1 4 a ,             a 2 + g ( 0 ) < x a g ( 0 )
Then the error entropy can be calculated as:
H ( p g ( x ) ) = 3 4 log ( 3 ) + log ( 4 a )
Case 2: 1 4 a < g ( 0 ) 3 4 a . In this case, we have:
p g ( x ) = { 1 4 a ,        a g ( 0 ) x < a 2 + g ( 0 ) 3 4 a ,          a 2 + g ( 0 ) x a g ( 0 ) 1 2 a ,             a g ( 0 ) < x a 2 + g ( 0 )
And hence:
H ( p g ( x ) ) = ( 9 8 + 3 g ( 0 ) 2 a ) log ( 3 ) + ( 1 4 g ( 0 ) a ) log ( 2 ) + log ( 4 a )
Case 3: g ( 0 ) > 3 4 a . In this case:
p g ( x ) = { 1 4 a ,        a g ( 0 ) x a g ( 0 ) 1 2 a ,       a 2 + g ( 0 ) x a 2 + g ( 0 )
Thus:
H ( p g ( x ) ) = 1 2 log ( 2 ) + log ( 4 a )
One can easily verify that the error entropy achieves its minimum value when 0 g ( 0 ) a / 4 (the first case). There are, therefore, infinitely many unbiased estimators that minimize the error entropy.

4. Conclusion

Two issues involved in the minimum error entropy (MEE) estimation have been studied in this work. The first issue is about which estimator minimizes the error entropy. In general there is no explicit expression for the MEE estimator unless some constraints on the conditional distribution are imposed. In the past, several researchers have shown that, if the conditional density is conditionally symmetric and unimodal (CSUM), then the conditional mean (or median) will be the MEE estimator. We extend these results to a more general case, and show that if the conditional densities are generalized uniformly dominated (GUD), then the dominant alignment will minimize the error entropy. The second issue is about the non-uniqueness of the unbiased MEE estimation. It has been shown in a recent paper that for the singular case (in which the error entropy approaches minus infinity), the unbiased MEE estimation may yield non-unique (even infinitely many) solutions. In this work, we show by examples that this result still holds even for nonsingular case.

Acknowledgments

This work was supported by National Natural Science Foundation of China (No. 60904054), NSF grant ECCS 0856441, and ONR N00014-10-1-0375.

References

  1. Cover, T.M.; Thomas, J.A. Element of Information Theory; Wiley & Son, Inc.: New York, NY, USA, 1991. [Google Scholar]
  2. Kapur, J.N.; Kesavan, H.K. Entropy Optimization Principles with Applications; Academic Press, Inc.: Boston, MA, USA, 1992. [Google Scholar]
  3. Weidemann, H.L.; Stear, E.B. Entropy analysis of estimating systems. IEEE Trans. Inform. Theor. 1970, 16, 264–270. [Google Scholar] [CrossRef]
  4. Tomita, Y.; Ohmatsu, S.; Soeda, T. An application of the information theory to estimation problems. Inf. Control 1976, 32, 101–111. [Google Scholar] [CrossRef]
  5. Wolsztynski, E.; Thierry, E.; Pronzato, L. Minimum-entropy estimation in semiparametric models. Signal Process. 2005, 85, 937–949. [Google Scholar] [CrossRef]
  6. Guo, L.Z.; Billings, S.A.; Zhu, D.Q. An extended orthogonal forward regression algorithm for system identification using entropy. Int. J. Control 2008, 81, 690–699. [Google Scholar] [CrossRef]
  7. Chen, B.; Zhu, Y.; Hu, J.; Principe, J.C. △-entropy: Definition, properties and applications in system identification with quantized data. Inf. Sci. 2011, 181, 1384–1402. [Google Scholar] [CrossRef]
  8. Kalata, P.; Priemer, R. Linear prediction, filtering and smoothing: an information theoretic approach. Inf. Sci. 1979, 17, 1–14. [Google Scholar] [CrossRef]
  9. Feng, X.; Loparo, K.A.; Fang, Y. Optimal state estimation for stochastic systems: An information theoretic approach. IEEE Trans. Automat. Contr. 1997, 42, 771–785. [Google Scholar] [CrossRef]
  10. Guo, L.; Wang, H. Minimum entropy filtering for multivariate stochastic systems with non-Gaussian Noises. IEEE Trans. Autom. Control 2006, 51, 695–700. [Google Scholar] [CrossRef]
  11. Wang, H. Minimum entropy control of non-Gaussian dynamic stochastic systems. IEEE Trans. Autom. Control 2002, 47, 398–403. [Google Scholar] [CrossRef]
  12. Yue, H.; Wang, H. Minimum entropy control of closed-loop tracking errors for dynamic stochastic systems. IEEE Trans. Autom. Control 2003, 48, 118–122. [Google Scholar]
  13. Erdogmus, D.; Principe, J.C. An error-entropy minimization algorithm for supervised training of nonlinear adaptive systems. IEEE Trans. Signal Process. 2002, 50, 1780–1786. [Google Scholar] [CrossRef]
  14. Erdogmus, D.; Principe, J.C. From linear adaptive filtering to nonlinear information processing—The design and analysis of information processing systems. IEEE Signal Process. Mag. 2006, 23, 14–33. [Google Scholar] [CrossRef]
  15. Principe, J.C. Information Theoretic Learning: Renyi's Entropy and Kernel Perspectives; Springer: New York, NY, USA, 2010. [Google Scholar]
  16. Boel, R.K.; James, M.R.; Petersen, I.R. Robustness and risk-sensitive filtering. IEEE Trans. Autom. Control 2002, 47, 451–461. [Google Scholar] [CrossRef]
  17. Janzura, M.; Koski, T.; Otahal, A. Minimum entropy of error principle in estimation. Inf. Sci. 1994, 79, 123–144. [Google Scholar] [CrossRef]
  18. Otahal, A. Minimum entropy of error estimate for multi-dimensional parameter and finite-state-space observations. Kybernetika 1995, 31, 331–335. [Google Scholar]
  19. Chen, T.-L.; Geman, S. On the minimum entropy of a mixture of unimodal and symmetric distributions. IEEE Trans. Inf. Theory 2008, 54, 3166–3174. [Google Scholar] [CrossRef]
  20. Chen, B.; Zhu, Y.; Hu, J.; Zhang, M. On optimal estimations with minimum error entropy criterion. J. Frankl. Inst.-Eng. Appl. Math. 2010, 347, 545–558. [Google Scholar] [CrossRef]
  21. Chen, B.; Zhu, Y.; Hu, J.; Zhang, M. A new interpretation on the MMSE as a robust MEE criterion. Signal Process. 2010, 90, 3313–3316. [Google Scholar] [CrossRef]

Share and Cite

MDPI and ACS Style

Chen, B.; Principe, J.C. Some Further Results on the Minimum Error Entropy Estimation. Entropy 2012, 14, 966-977. https://doi.org/10.3390/e14050966

AMA Style

Chen B, Principe JC. Some Further Results on the Minimum Error Entropy Estimation. Entropy. 2012; 14(5):966-977. https://doi.org/10.3390/e14050966

Chicago/Turabian Style

Chen, Badong, and Jose C. Principe. 2012. "Some Further Results on the Minimum Error Entropy Estimation" Entropy 14, no. 5: 966-977. https://doi.org/10.3390/e14050966

APA Style

Chen, B., & Principe, J. C. (2012). Some Further Results on the Minimum Error Entropy Estimation. Entropy, 14(5), 966-977. https://doi.org/10.3390/e14050966

Article Metrics

Back to TopTop