Abstract
Popular nonlinear dimensionality reduction algorithms, e.g., SIE and Isomap suffer a difficulty in common: global neighborhood parameters often fail in tackling data sets with high variation in local manifold. To improve the availability of nonlinear dimensionality reduction algorithms in the field of machine learning, an adaptive neighbors selection scheme based on locally principal direction reconstruction is proposed in this paper. Our method involves two main computation steps. First, it selects an appropriate neighbors set for each data points such that all neighbors in a neighbors set form a d-dimensional linear subspace approximately and computes locally principal directions for each neighbors set respectively. Secondly, it fits each neighbor by means of locally principal directions of corresponding neighbors set and deletes the neighbors whose fitting error exceeds a predefined threshold. The simulation shows that our proposal could deal with data set with high variation in local manifold effectively. Moreover, comparing with other adaptive neighbors selection strategy, our method could circumvent false connectivity induced by noise or high local curvature.
Supported by Science-Technology Development Project of Tianjin(04310941R) and Applied Basic Research Project of Tianjin (i05YFJMJC11700).
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Barlow, H.: Unsupervised learning. Neural Computation 1, 295–311 (1989)
Marcus, G.: Programs of the Mind. Science 304, 1450-1451
Baum, E.: What Is Thought? MIT Press, Cambridge (2004)
Mardia, K.V., Kent, J.T., Bibby, J.M.: Multivariate Analysis. Academic Press, London (1979)
Tenenbaum, J.B., et al.: A Global Geometric Framework for Nonlinear Dimensionality Reduction. Science 290, 2319–2323 (2000)
Roweis, S.T., et al.: Nonlinear Dimensionality Reduction by Locally Linear Embedding. Science 290, 2323–2326 (2000)
De Silva, V., Tenenbaum, J.: Global versus local methods in nonlinear dimensionality reduction. In: de Silva, V., et al. (eds.) NIPS (2002)
Hou, Y., et al.: Robust Nonlinear Dimension Reduction: A Self-organized Approach. In: FSDK 2005 (2005)
Wang, J., Zhang, Z., Zha, H.: Adaptive Manifold Learning. In: NIPS 2004 (2004)
Belkin, M., Niyogi, P.: Laplacian Eigenmaps for Dimensionality Reduction and Data Representation. Neural Computation 15(6), 1373–1396 (2003)
Hou, Y., et al.: Adaptive manifold learning Based on Statistical Criterions, Tech report of Tianjin university
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2006 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Hou, Y., Yang, H., He, P. (2006). Locally Adaptive Nonlinear Dimensionality Reduction. In: Gelbukh, A., Reyes-Garcia, C.A. (eds) MICAI 2006: Advances in Artificial Intelligence. MICAI 2006. Lecture Notes in Computer Science(), vol 4293. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11925231_52
Download citation
DOI: https://doi.org/10.1007/11925231_52
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-49026-5
Online ISBN: 978-3-540-49058-6
eBook Packages: Computer ScienceComputer Science (R0)