Abstract
Recently, a three-stage version of K-Means has been introduced, at which not only clusters and their centers, but also feature weights are adjusted to minimize the summary p-th power of the Minkowski p-distance between entities and centroids of their clusters. The value of the Minkowski exponent p appears to be instrumental in the ability of the method to recover clusters hidden in data. This paper advances into the problem of finding the best p for a Minkowski metric-based version of K-Means, in each of the following two settings: semi-supervised and unsupervised. This paper presents experimental evidence that solutions found with the proposed approaches are sufficiently close to the optimum.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Arbelaitz, O., Gurrutxaga, I., Muguerza, J., Pérez, J.M., Perona, I.: An extensive comparative study of cluster validity indices. Pattern Recognit. 46, 243–256 (2012)
Bache, K., Lichman, M.: UCI machine learning repository. http://archive.ics.uci.edu/ml (2013)
Chan, E.Y., Ching, W.K., Ng, M.K., Huang, J.Z.: An optimization algorithm for clustering using weighted dissimilarity measures. Pattern Recognit. 37(5), 943–952 (2004)
Chiang, M.M.T., Mirkin, B.: Intelligent choice of the number of clusters in k-means clustering: an experimental study with different cluster spreads. J. Classif. 27(1), 3–40 (2010)
de Amorim, R.C., Fenner, T.: Weighting features for partition around medoids using the minkowski metric. In: Jaakko, H., Frank, K., Allan, T. (eds.) Advances in Intelligent Data Analysis. Lecture Notes in Computer Science, vol. 7619, pp. 35–44. Springer, Berlin (2012)
de Amorim, R.C., Komisarczuk, P.: On initializations for the minkowski weighted k-means. In: Jaakko, H., Frank, K., Allan, T. (eds.) Advances in Intelligent Data Analysis. Lecture Notes in Computer Science, vol. 7619, pp. 45–55. Springer, Berlin (2012)
de Amorim, R.C., Mirkin, B.: Minkowski metric, feature weighting and anomalous cluster initializing in k-means clustering. Pattern Recognit. 45(3), 1061–1075 (2012)
Frigui, H., Nasraoui, O.: Unsupervised learning of prototypes and attribute weights. Pattern Recognit. 37(3), 567–581 (2004)
Huang, J.Z., Ng, M.K., Rong, H., Li, Z.: Automated variable weighting in k-means type clustering. IEEE Trans. Pattern Anal. Mach. Intell. 27(5), 657–668 (2005)
Huang, J.Z., Xu, J., Ng, M., Ye, Y.: Weighting method for feature selection in k-means. In: Computational Methods of Feature Selection, pp. 193–209. Chapman & Hall, London (2008)
Makarenkov, V., Legendre, P.: Optimal variable weighting for ultrametric and additive trees and k-means partitioning: Methods and software. J. Classif. 18(2), 245–271 (2001)
Mirkin, B.: Clustering for Data Mining: A Data Recovery Approach, vol. 3. Chapman & Hall, London (2005)
Rousseeuw, P.J.: Silhouettes: a graphical aid to the interpretation and validation of cluster analysis. J. Comput. Appl. Math. 20, 53–65 (1987)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2014 Springer Science+Business Media New York
About this chapter
Cite this chapter
de Amorim, R.C., Mirkin, B. (2014). Selecting the Minkowski Exponent for Intelligent K-Means with Feature Weighting. In: Aleskerov, F., Goldengorin, B., Pardalos, P. (eds) Clusters, Orders, and Trees: Methods and Applications. Springer Optimization and Its Applications, vol 92. Springer, New York, NY. https://doi.org/10.1007/978-1-4939-0742-7_7
Download citation
DOI: https://doi.org/10.1007/978-1-4939-0742-7_7
Published:
Publisher Name: Springer, New York, NY
Print ISBN: 978-1-4939-0741-0
Online ISBN: 978-1-4939-0742-7
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)