Abstract
In data analysis or data mining, there are mainly two fundamental types of methodologies, called unsupervised and supervised learning algorithms. This chapter explores principal component analysis, independent component analysis, density estimation and clustering analysis.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Amari, S. (1998). Natural gradient works efficiently in learning. Neural Computation, 10, 251–276.
Amigó, E., Gonzalo, J., Artiles, J., & Verdejo, F. (2009). A comparison of extrinsic clustering evaluation metrics based on formal constraints. Information Retrieval, 12, 461–486.
Bezdek, J. C., Ehrlich, R., & Full, W. (1984). FCM: The fuzzy c-means clustering algorithm. Computers & Geoscience, 10, 191–2003.
Constantinopoulos, C., Titsias, M., & Likas, A. (2006). Bayesian feature and model selection for Gaussian mixture models. IEEE Transactions on Pattern Analysis and Machine Intelligence, 28, 1013–1018.
Cox, D. R. (1957). Note on grouping. Journal of the American Statistical Association, 52, 543–547.
Dempster, A. P., Laird, N. M., & Rubin, D. B. (1977). Maximum likelihood from incomplete data via the EM algorithm. Journal of the Royal Statistical Society. Series B, 39, 1–38.
Dunn, J. C. (1973). A fuzzy relative of the ISODATA process and its use in detecting compact well-separated clusters. Journal of Cybernetics, 3, 32–57.
Eguchi, S., Notsu, A., & Komori, O. (2017). Spontaneous learning for data distributions via minimum divergence. In F. Nielsen, F. Critchley, & C. Dodson (Eds.), Computational information geometry (pp. 79–99). Cham: Springer.
Fränti, P., & Sieranoja, S. (2018). K-means properties on six clustering benchmark datasets. Applied Intelligence, 48, 4743–4759.
Fränti, P., Rezaei, M., & Zhao, Q. (2014). Centroid index: Cluster level similarity measure. Pattern Recognition, 47, 3034–3045.
Ghosh, S., & Dubey, S. (2013). Comparative analysis of k-means and fuzzy c-means algorithms. International Journal of Advanced Computer Science and Applications, 4, 35–39.
Hammersley, J. M., & Morton, K. W. (1954). Poor man’s Monte Carlo. Journal of the Royal Statistical Society. Series B, 16, 23–38.
Hathaway, R. J., & Bezdek, J. C. (1995). Optimization of clustering criteria by reformulation. IEEE Transactions on Fuzzy Systems, 3, 241–245.
Henderson, D., Jacobson, S. H., & Johnson, A. (2003). The theory and practice of simulated annealing. In F. Glover & G. A. Kochenberger (Eds.), Handbook of metaheuristics. Boston: Springer.
Higuchi, I., & Eguchi, S. (1998). The influence function of principal component analysis by self-organizing rule. Neural Computation, 10, 1435–1444.
Higuchi, I., & Eguchi, S. (2004). Robust principal component analysis with adaptive selection for tuning parameters. Journal of Machine Learning Research, 5, 453–472.
Hosking, J. R., & Wallis, J. R. (1987). Parameter and quantile estimation for the generalized Pareto distribution. Technometrics, 29, 339–349.
Hotelling, H. (1933). Analysis of a complex of statistical variables into principal components. Journal of Educational Psychology, 24, 417–441.
Huang, S., Yeh, Y., & Eguchi, S. (2009). Robust Kernel principal component analysis. Neural Computation, 21, 3179–3213.
Hyvarinen, A. (1999). Gaussian moments for noisy independent component analysis. IEEE Signal Processing Letters, 6, 145–147.
Jain, A. K. (2010). Data clustering: 50 years beyond K-means. Pattern Recognition Letters, 31, 651–666.
Jolliffe, I. T. (2002). Principal component analysis. New York: Springer.
Jolliffe, I. T. (2021). A 50-year personal journey through time with principal component analysis. Journal of Multivariate Analysis, 104820.
Jolliffe, I. T., & Cadima, J. (2016). Principal component analysis: A review and recent developments. Philosophical Transactions of the Royal Society A, 374, 89–90.
Kamiya, H., & Eguchi, S. (2001). A class of robust principal component vectors. Journal of Multivariate Analysis, 77, 239–269.
Komori, O., & Eguchi, S. (2021). A unified formulation of k-means, fuzzy c-means and Gaussian mixture model by the Kolmogorov-Nagumo average. Entropy, 23, 518.
Lloyd, S. P. (1982). Least squares quantization in PCM. IEEE Transactions on Information Theory, 129–137.
MacQueen, J. (1967). Some methods of classification and analysis of multivariate observations. In L. M. L. Cam & J. Neyman (Eds.), Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability (pp. 281–297). Berkeley: University of California Press.
McNicholas, P. D., & Murphy, T. B. (2008). Parsimonious Gaussian mixture models. Statistics and Computing, 18, 285–296.
Minami, M., & Eguchi, S. (2002). Robust blind source separation by beta divergence. Neural Computation, 14, 1859–1886.
Mollah, M. N. H., Minami, M., & Eguchi, S. (2006). Exploring latent structure of mixture ICA models by the minimum beta-divergence method. Neural Computation, 18, 166–190.
Mollah, M. N. H., Sultana, N., Minami, M., & Eguchi, S. (2010). Robust extraction of local structures by the minimum beta-divergence method. Neural Networks, 23, 226–238.
Notsu, A., & Eguchi, S. (2016). Robust clustering method in the presence of scattered observations. Neural Computation, 28, 1141–1162.
Notsu, A., Komori, O., & Eguchi, S. (2014). Spontaneous clustering via minimum gamma-divergence. Neural Computation, 26, 421–448.
Pearson, K. (1901). LIII. On lines and planes of closest fit to systems of points in space. The London, Edinburgh, and Dublin Philosophical Magazine and Journal of Science,2, 559–572. https://doi.org/10.1080/14786440109462720.
Pickands, J. (1975). Statistical inference using extreme order statistics. The Annals of Statistics, 3, 119–131.
Rose, K., Gurewitz, E., & Fox, G. C. (1990). Statistical mechanics and phase transitions in clustering. Physical Review Letters, 65, 945–948.
Sofaer, H. R., Hoeting, J. A., & Jarnevich, C. S. (2019). The area under the precision-recall curve as a performance metric for rare binary events. Methods in Ecology and Evolution, 10, 565–577.
Steinhaus, H. (1957). Sur la division des corps matériels en parties. Bulletin L’Académie Polonaise des Science, 4, 801–804.
Tibshirani, R., Walther, G., & Hastie, T. (2001). Estimating the number of clusters in a data set via the gap statistic. Journal of Royal Statistic Society Series B, 63, 411–423.
Van Rijsbergen, C. (1974). Foundation of evaluation. Journal of Documentation, 30, 365–373.
Xu, R., & Wunsch, D. (2005). Survey of clustering algorithms. IEEE Transactions on Neural Networks, 16, 236–243.
Yu, J. (2005). General C-means clustering model. IEEE Transactions on Pattern Analysis and Machine Intelligence, 27, 1197–1211.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
Copyright information
© 2022 Springer Japan KK, part of Springer Nature
About this chapter
Cite this chapter
Eguchi, S., Komori, O. (2022). Unsupervised Learning Algorithms. In: Minimum Divergence Methods in Statistical Machine Learning. Springer, Tokyo. https://doi.org/10.1007/978-4-431-56922-0_5
Download citation
DOI: https://doi.org/10.1007/978-4-431-56922-0_5
Published:
Publisher Name: Springer, Tokyo
Print ISBN: 978-4-431-56920-6
Online ISBN: 978-4-431-56922-0
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)