Abstract
In this paper, we propose an incremental and local adaptive gaussian mixture for online density estimation (LAIM). Using a similarity threshold based criterion, the method is able to allocate components incrementally to accommodate novel data points without affecting previously learned components. A local adaptive learning strategy is presented for estimating density with complex structure in an online way. We also adopt a denoising scheme to make the algorithm more robust to noise. We compared the LAIM to the state-of-art methods for density estimation in both artificial and real data sets, the results show that our method outperforms the compared online counterpart and produces comparable results to the compared batch algorithms.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
McLachlan, G., Peel, D.: Finite mixture models. Wiley-Interscience (2000)
Parzen, E.: On estimation of a probability density function and mode. Annals of Math. Statistics 33, 1065–1076 (1962)
Wand, M.P., Jones, M.C.: Kernel Smoothing. Chapman Hall/CRC (1995)
Hall, P., Sheater, S.J., Jones, M.C., Marron, J.S.: On optimal data-based bandwidth selection in kernel density estimation. Biometrika 78(2), 263–269 (1991)
Silverman, B.W.: Density Estimation. Chapman and Hall, London (1986)
Hall, P.: Large sample optimality of least squares cross-validation in density estimation. The Annals of Statistics, 1156–1174 (1983)
Girolami, M., He, C.: Probability density estimation from optimally condensed data samples. IEEE Trans. Pattern Anal. Mach. Intell. 25(10), 1253–1264 (2003)
Botev, Z.I., Grotowski, J.F., Kroese, D.P.: Kernel density estimation via diffusion. The Annals of Statistics 38(5), 2916–2957 (2010)
Vlassis, N., Likas, A.: A greedy EM algorithm for Gaussian mixture learning. Neural processing letters 15(1), 77–87 (2002)
Engel, P.M., Heinen, M.R.: Incremental learning of multivariate gaussian mixture models. In: da Rocha Costa, A.C., Vicari, R.M., Tonidandel, F. (eds.) SBIA 2010. LNCS, vol. 6404, pp. 82–91. Springer, Heidelberg (2010)
Yin, H., Allinson, N.M.: Self-organizing mixture networks for probability density estimation. IEEE Transactions on Neural Networks 12(2), 405–411 (2001)
Kristan, M., Leonardis, A., Skoaj, D.: Multivariate online kernel density estimation with Gaussian kernels. Pattern Recognition 44(10), 2630–2642 (2011)
Bishop, C.M.: Pattern recognition and machine learning, vol. 1. Springer, New York (2006)
Ouyang, Q., Shen, F., Zhao, J.: A local distribution net for data clustering. In: Anthony, P., Ishizuka, M., Lukose, D. (eds.) PRICAI 2012. LNCS, vol. 7458, pp. 411–422. Springer, Heidelberg (2012)
Furao, S., Ogura, T., Hasegawa, O.: An enhanced self-organizing incremental neural network for online unsupervised learning. Neural Networks 20(8), 893–903 (2007)
Asuncion, A., Newman, D.: UCI machine learning repository (2007)
Chang, C.-C., Lin, C.-J.: LIBSVM: a library for support vector machines. ACM Transactions on Intelligent Systems and Technology (TIST) 2(3), 27 (2011)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer International Publishing Switzerland
About this paper
Cite this paper
Qiu, T., Shen, F., Zhao, J. (2015). Local Adaptive and Incremental Gaussian Mixture for Online Density Estimation. In: Cao, T., Lim, EP., Zhou, ZH., Ho, TB., Cheung, D., Motoda, H. (eds) Advances in Knowledge Discovery and Data Mining. PAKDD 2015. Lecture Notes in Computer Science(), vol 9077. Springer, Cham. https://doi.org/10.1007/978-3-319-18038-0_33
Download citation
DOI: https://doi.org/10.1007/978-3-319-18038-0_33
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-18037-3
Online ISBN: 978-3-319-18038-0
eBook Packages: Computer ScienceComputer Science (R0)