Estimating the Major Cluster by Mean-Shift with Updating Kernel
Abstract
:1. Introduction
2. General Mean-Shift Method
2.1. General Mean-Shift Method
- Letting the mean of sample be the initial value of the mean estimator of major cluster, then
- Consider a Gaussian distribution with the mean and standard deviation as the kernel function in the value direction. Here, the mean of kernel function is found by the mean estimator of major clusterThe standard deviation is assigned to be an appropriate size as discussed later in Section 2.2.
- Weight for each sample weighted by such a Gaussian kernel isHowever, A in Equation (3) above is a normalization coefficient for which the sum of the weight is equal to 1, asWe use this weight to calculate the sample mean with as
- The value of mean estimator of the major cluster is updated by the following equation:
- If the variation of the value of mean estimator is equal to or less than the predetermined fixed value, then the update process is terminated. Otherwise, return to 2 and repeat the iteration.
2.2. Shortcomings and Solution of the General Mean-Shift Method
3. One-Dimensional Mean-Shift with Updating Kernel
3.1. Derivation of Major Cluster Standard Deviation from Sample Standard Deviation
3.2. Mean-Shift with Updating Kernel
- Let the mean of sample be the initial value of the mean estimator of the major cluster and let standard deviation of this sample be the initial value of the standard deviation estimator of the major cluster as
- Consider a Gaussian distribution with mean and standard deviation as the kernel function in the value direction. Here, the mean and the standard deviation are given respectively by the estimated value of the mean and the estimated value of the standard deviation of the major cluster:Here, mean and variance of the Gaussian kernel are not estimators, although they change when the kernel updates.
- Weight for each sample weighted by such a Gaussian kernel is calculated using Equations (3) and (4). We use this weight to calculate the sample mean and standard deviation with as shown below:
- The values of mean estimator , standard deviation estimator , and number of samples estimator of the sample are updated, respectively, by the following equations:
- If the variations of the values of these estimators are equal to or less than the predetermined fixed value, then the update process is terminated. Otherwise, return to 2 and repeat the iteration.
4. Numerical Experiment
4.1. Update Process of Mean-Shift with an Updatable Kernel
4.2. Influence of Kernel Bandwidth on Estimation Accuracy (Unbiasedness)
4.3. Influence of the Scale Factor r Value on Estimation Accuracy
4.4. Verification of Consistency
4.5. Estimation Precisions of the Proposed and General Mean-Shift Methods
4.6. Discussion
5. Application
6. Conclusions
Author Contributions
Funding
Acknowledgments
Conflicts of Interest
Appendix A. General Mean-Shift for a Multi-Dimensional Situation
- Let the mean vector of sample be the initial value of the mean estimator of the major cluster
- Consider a M-dimensional Gaussian distribution with mean vector and covariance matrix as the kernel function in value direction. Here, the mean mean vector of kernel function is ascertained by the mean estimator of major clusterIn addition, covariance matrix is assigned to be an appropriate size as discussed in Section 2.2.
- The weight for each sample weighted by such a Gaussian kernel isHowever,We use this weight to calculate the sample mean vector with as
- The value of mean vector estimator for the major cluster is updated using the following equation:
- If the value variation of mean vector estimator is equal to or less than the predetermined fixed value, the update process is terminated. Otherwise, return to 2 and repeat the iteration.
Appendix B. Proof of Equation (17)
Expectation | S.D. | |
---|---|---|
gamma | ||
k | ||
exponential | ||
Erlang | ||
Rayleigh | ||
log-normal | ||
Pareto |
Appendix C. Multi-Dimensional Mean-Shift with Updating Kernel
Appendix C.1. Derivation of Standard Deviation of a Major Cluster from the Sample
Appendix C.2. Mean-Shift Method with Updating Kernel
- The mean vector and the covariance matrix of the whole samples are determined using the following equations:The initial values of the mean vector and the covariance matrix of the major cluster are assigned as
- One can consider a multi-dimensional Gaussian distribution with mean vector and covariance matrix as the kernel function in the value direction. Here, the mean vector and covariance matrix of the kernel function are determined asActually, in the above equation is derived from the fact that the covariance matrix has the squared order of the standard deviation.
- Weight for each sample weighted by such a Gaussian kernel is calculated using Equations (A3) and (A4). The mean vector and the covariance matrix are determined using the following equations:
- The value of mean vector estimator is updated using the following equation:Let
- If the value variations of are equal to or less than the predetermined fixed value, then the update process is terminated. Otherwise, return to 2 and repeat the iteration.
References
- Fukunaga, K.; Hostetler, L. The estimation of the gradient of a density function, with applications in pattern recognition. IEEE Trans. Inf. Theory 1975, 21, 32–40. [Google Scholar] [CrossRef]
- Cheng, Y. Mean shift, mode seeking, and clustering. IEEE Trans. Pattern Anal. Mach. Intell. 1995, 17, 790–799. [Google Scholar] [CrossRef]
- Comaniciu, D.; Meer, P. Mean shift analysis and applications. In Proceedings of the Seventh IEEE International Conference on Computer Vision, Kerkyra, Greece, 20–27 September 1999; pp. 1197–1203. [Google Scholar]
- Comaniciu, D.; Meer, P. Mean shift: A robust approach toward feature space analysis. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 603–619. [Google Scholar] [CrossRef]
- Wand, M.P.; Jones, M.C. Kernel Smoothing; Springer: London, UK, 1995. [Google Scholar]
- Sheather, S.J.; Jones, M.C. A reliable data-based bandwidth selection method for kernel density estimation. J. R. Stat. Soc. Ser. B 1991, 53, 683–690. [Google Scholar] [CrossRef]
- Chen, S. Optimal bandwidth selection for kernel density functionals estimation. J. Probab. Stat. 2015, 70, 1–21. [Google Scholar] [CrossRef]
- Slaoui, Y. Data-Driven Bandwidth Selection for Recursive Kernel Density Estimators Under Double Truncation. Sankhya B 2018, 80, 341–368. [Google Scholar] [CrossRef]
- Comaniciu, D.; Ramesh, V.; Meer, P. The variable bandwidth mean shift and data-driven scale selection. In Proceedings of the 1991 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Maui, HI, USA, 3–6 June 1991; pp. 438–445. [Google Scholar]
- Okada, K.; Comaniciu, D.; Krishnan, A. Scale selection for anisotropic scale-space. In Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Washington, DC, USA, 27 June–2 July 2004; pp. 594–601. [Google Scholar]
- Comaniciu, D. An algorithm for data-driven bandwidth selection. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 281–288. [Google Scholar] [CrossRef]
- Li, X.; Hu, Z.; Wu, F. A note on the convergence of the mean shift. Pattern Recognit. 2007, 40, 1756–1762. [Google Scholar] [CrossRef]
- Ghassabeh, Y.A.; Rudzicz, F. Modified mean shift algorithm. IET Image Process. 2018, 12, 2171–2177. [Google Scholar]
- Birchfield, S.T.; Rangarajan, S. Mean shift blob tracking with kernel histogram filtering and hypothesis testing. Pattern Recognit. Lett. 2005, 26, 605–614. [Google Scholar]
- Leichter, I. Mean shift trackers with cross-bin metrics. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 34, 695–706. [Google Scholar] [CrossRef] [PubMed]
- Vojir, T.; Noskova, J.; Matasa, J. Robust scale-adaptive mean-shift for tracking. Pattern Recognit. Lett. 2014, 49, 250–258. [Google Scholar] [CrossRef]
- Silverman, B.W. Density Estimation for Statistics and Data Analysis; Chapman & Hall: London, UK, 1986. [Google Scholar]
- Kawamura, Y.; Yokota, Y.; Matsumaru, N.; Shirai, K. 24 Hours monitoring system of heart rate variability to predict septic shock. IEICE Tech. Rep. 2012, 112, 29–34. [Google Scholar]
- Yokota, Y.; Ye, T. Quasi-Gaussian distributed power spectrum series by radical root transform and application to robust power spectrum density estimation against for sudden noise. IEICE Trans. Fundam. (Jpn. Ed.) 2016, 3, 149–158. [Google Scholar]
- Lotter, T.; Vary, P. Noise reduction by maximum a posteriori spectral amplitude estimation with super-Gaussian speech modelling. In Proceedings of the International Workshop on Acoustic Echo and Noise Control, Kyoto, Japan, 8 September 2003; pp. 83–88. [Google Scholar]
- Ye, T.; Yokota, Y. Noise estimation for speech enhancement based on quasi-Gaussian distributed power spectrum series by radical root transformation. IEICE Trans. Fundam. Electron. Commun. Comput. Sci. 2017, 6, 1306–1314. [Google Scholar] [CrossRef]
- Grubbs, F.E. Sample criteria for testing outlying observations. Ann. Math. Stat. 1950, 21, 27–58. [Google Scholar] [CrossRef]
- Grubbs, F.E.; Beck, G. Extension of sample sizes and percentage points for significance tests of outlying observations. Technometrics 1972, 14, 847–854. [Google Scholar] [CrossRef]
- Zeller, C.B.; Lachos, V.H.; Labra, F.V. Influence diagnostics for Grubbs’s model with asymmetric heavy-tailed distributions. Stat. Pap. 2014, 55, 671–690. [Google Scholar] [CrossRef]
- Thompson, R. A note on restricted maximum likelihood estimation with an alternative outlier model. J. R. Stat. Soc. Ser. B (Methodol.) 1985, 47, 53–55. [Google Scholar] [CrossRef]
- Rasmussen, C.E. The infinite Gaussian mixture model. In Advances in Information Processing Systems 12; MIT Press: Cambridge, MA, USA, 2000; pp. 554–560. [Google Scholar]
- Blomer, J.; Bujna, K. Adaptive seeding for Gaussian mixture models. In Proceedings of the 20th Pacific-Asia Conference, PAKDD 2016, Auckland, New Zealand, 19–22 April 2016; Volume 9652, pp. 296–308. [Google Scholar]
- Viroli, C.; McLachlan, G.J. Deep Gaussian mixture models. Stat. Comput. 2017, 29, 43–51. [Google Scholar] [CrossRef] [Green Version]
- Dempster, A.P.; Laird, N.M.; Rubin, D.B. Maximum likelihood from incomplete data via the EM algorithm. J. R. Stat. Soc. Ser. B (Stat. Methodol.) 1977, 39, 1–38. [Google Scholar] [CrossRef]
- Melnykov, V.; Melnykov, I. Initializing the EM algorithm in Gaussian mixture models with an unknown number of components. J. Comput. Stat. Data Anal. 2012, 56, 1381–1395. [Google Scholar] [CrossRef]
© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Tian, Y.; Yokota, Y. Estimating the Major Cluster by Mean-Shift with Updating Kernel. Mathematics 2019, 7, 771. https://doi.org/10.3390/math7090771
Tian Y, Yokota Y. Estimating the Major Cluster by Mean-Shift with Updating Kernel. Mathematics. 2019; 7(9):771. https://doi.org/10.3390/math7090771
Chicago/Turabian StyleTian, Ye, and Yasunari Yokota. 2019. "Estimating the Major Cluster by Mean-Shift with Updating Kernel" Mathematics 7, no. 9: 771. https://doi.org/10.3390/math7090771
APA StyleTian, Y., & Yokota, Y. (2019). Estimating the Major Cluster by Mean-Shift with Updating Kernel. Mathematics, 7(9), 771. https://doi.org/10.3390/math7090771