Abstract
Mixture models, especially mixtures of Gaussian, have been widely used due to their great flexibility and power. Non-Gaussian clusters can be approximated by several Gaussian components, however, it can not always acquire appropriate results. By cancelling the nonnegative constraint to mixture coefficients and introducing a new concept of “negative components”, we extend the traditional mixture models and enhance their performance without increasing the complexity obviously. Moreover, we propose a parameter estimation algorithm based on an iteration mechanism, which can effectively discover patterns of “negative components”. Experiments on some synthetic data testified the reasonableness of the proposed novel model and the effectiveness of the parameter estimation algorithm.
This work is supported by the project (60475001) of the National Natural Science Foundation of China.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
McLachlan, G., Peel, D.: Finite Mixture Models. John Wiley & Sons, New York (2000)
Jain, A.K., Dubes, R.: Algorithm for Clustering Data. Prentice Hall, Englewood Cliffs (1988)
Hinton, G., Dayan, P., Revow, M.: Modeling the manifolds of images of handwritten digits. IEEE Trans. on Neural Networks 8, 65–74 (1997)
Dasgupta, A., Raftery, A.E.: Detecting features in spatial point processes with clutter via model-based clustering. Journal of the American Statistical Association 93, 294–302 (1998)
Fraley, C., Raftery, A.E.: How many clusters? which clustering method? – answers via model-based cluster analysis. The Computer Journal 41, 578–588 (1998)
Zhang, B., Zhang, C., Yi, X.: Competitive EM algorithm for finite mixture models. Pattern Recognition 37, 131–144 (2004)
Figueiredo, M.A., Jain, A.K.: Unsupervised learning of finite mixture models. IEEE Trans. on PAMI 24, 381–396 (2002)
Bilmes, J.A.: A gentle tutorial of the EM algorithm and its application to parameter estimation for Gaussian mixture and hidden markov models. Technical Report ICSI TR-97-021, UC Berkeley (1997)
Dempster, A.P., Laird, N.M., Rubin, D.B.: Maximum-likelihood from incomplete data via the EM algorithm. Journal of Royal Statistic, Society B 39, 1–38 (1977)
Geman, S., Geman, D.: Stochastic relaxation, Gibbs distribution and the Bayesian restoration of images. IEEE Transactions on PAMI 6, 721–741 (1984)
Vincent, P., Bengio, Y.: Manifold Parzen windows. In: NIPS (2002)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2005 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Zhang, B., Zhang, C. (2005). Finite Mixture Models with Negative Components. In: Perner, P., Imiya, A. (eds) Machine Learning and Data Mining in Pattern Recognition. MLDM 2005. Lecture Notes in Computer Science(), vol 3587. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11510888_4
Download citation
DOI: https://doi.org/10.1007/11510888_4
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-26923-6
Online ISBN: 978-3-540-31891-0
eBook Packages: Computer ScienceComputer Science (R0)