Skip to main content
Log in

Statistical initialization of intrinsic K-means clustering on homogeneous manifolds

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

The K-means algorithm is widely applied for clustering, and its clustering effect is influenced by its initialization. However, most existing works focus on the initialization of K and centers in Euclidean spaces, but few works in the literature deal with the initialization of K-means clustering on Riemannian manifolds. In this paper, we propose a unified scheme for learning K and selecting the initial centers for intrinsic K-means clustering on homogeneous manifolds, which can also be generalized to other types of manifolds. First, geodesic verticality is presented based on the geometric properties abstracted from the definition of orthogonality in Euclidean spaces. Then, geodesic projection on Riemannian manifolds is proposed for learning K, which achieves nonlinear dimensionality reduction and improves the computing efficiency. Additionally, the Riemannian metric of \(\mathbb {S}^{n}\) is derived for the statistical initialization of the centers to improve the clustering accuracy. Finally, the intrinsic K-means algorithm for clustering on homogeneous manifolds based on the Karcher mean is given by applying the proposed manifold initialization, which improves the clustering effect. Simulations and experimental studies are conducted to show the effectiveness and accuracy of the proposed K-means scheme on manifolds.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14
Fig. 15
Fig. 16
Fig. 17
Fig. 18

Similar content being viewed by others

References

  1. Lu W (2020) Improved k-means clustering algorithm for big data mining under hadoop parallel framework. J Grid Comput 18(2):239–250

    Article  Google Scholar 

  2. Jaquier N, Rozo L, Caldwell DG, Calinon S (2021) Geometry-aware manipulability learning, tracking, and transfer. Int J Robot Res 40(2-3):624–650

    Article  Google Scholar 

  3. Arthur D, Vassilvitskii S (2006) K-means++: the advantages of careful seeding. Tech rep, Stanford

  4. Zhang W, Kong D, Wang S, Wang Z (2019) 3d human pose estimation from range images with depth difference and geodesic distance. J Vis Commun Image Represent 59:272–282

    Article  Google Scholar 

  5. Yan Z, Duckett T, Bellotto N (2020) Online learning for 3d lidar-based human detection: experimental analysis of point cloud clustering and classification methods. Auton Robot 44(2):147–164

    Article  Google Scholar 

  6. Borlea ID, Precup RE, Borlea AB, Iercan D (2021) A unified form of fuzzy c-means and k-means algorithms and its partitional implementation. Knowl-Based Syst 214:106731

    Article  Google Scholar 

  7. Fränti P, Sieranoja S (2018) K-means properties on six clustering benchmark datasets. Appl Intell 48(12):4743–4759

    Article  MATH  Google Scholar 

  8. Hamerly G, Elkan C (2003) Learning the k in k-means. Advances in neural information processing systems 16:281–288

    Google Scholar 

  9. Calinon S (2020) Gaussians on riemannian manifolds: applications for robot learning and adaptive control. IEEE Robot Autom Mag 27(2):33–45

    Article  Google Scholar 

  10. Hechmi S, Gallas A, Zagrouba E (2019) Multi-kernel sparse subspace clustering on the riemannian manifold of symmetric positive definite matrices. Pattern Recogn Lett 125:21–27

    Article  Google Scholar 

  11. Fathian K, Ramirez-Paredes JP, Doucette EA, Curtis JW, Gans NR (2018) Quest: a quaternion-based approach for camera motion estimation from minimal feature points. IEEE Robotics and Automation Letters 3(2):857–864

    Article  Google Scholar 

  12. Zeestraten MJ, Havoutis I, Silvério J, Calinon S, Caldwell DG (2017) An approach for imitation learning on riemannian manifolds. IEEE Robot Automn Lett 2(3):1240–1247

    Article  Google Scholar 

  13. Lang M, Hirche S (2017) Computationally efficient rigid-body gaussian process for motion dynamics. IEEE Robot Autom Lett 2(3):1601–1608

    Article  Google Scholar 

  14. Absil PA, Mahony R, Sepulchre R (2009) Optimization algorithms on matrix manifolds. Princeton University Press

  15. Pennec X, Fillard P, Ayache N (2006) A riemannian framework for tensor computing. Int J Comput Vis 66(1):41–66

    Article  MATH  Google Scholar 

  16. Lin Z, Yao F (2019) Intrinsic riemannian functional data analysis. Ann Stat 47(6):3533–3577

    Article  MathSciNet  MATH  Google Scholar 

  17. Saha J, Mukherjee J (2021) Cnak: cluster number assisted k-means. Pattern Recogn 110:107625

    Article  Google Scholar 

  18. Zhang T, Lin G (2021) Generalized k-means in glms with applications to the outbreak of covid-19 in the United States. Comput Stat Data Anal 159:107217

    Article  MathSciNet  MATH  Google Scholar 

  19. Wang F, Franco-Penya HH, Kelleher JD, Pugh J, Ross R (2017) An analysis of the application of simplified silhouette to the evaluation of k-means clustering validity. In: International conference on machine learning and data mining in pattern recognition. Springer, pp 291–305

  20. Zhang G, Zhang C, Zhang H (2018) Improved k-means algorithm based on density canopy. Knowl-Based Syst 145:289–297

    Article  Google Scholar 

  21. Nasser A, Hamad D, Nasr C (2006) K-means clustering algorithm in projected spaces. In: 2006 9Th international conference on information fusion. IEEE, pp 1–6

  22. Zhou J, Pedrycz W, Yue X, Gao C, Lai Z, Wan J (2021) Projected fuzzy c-means clustering with locality preservation. Pattern Recogn 113:107748

    Article  Google Scholar 

  23. Lasheng C, Yuqiang L (2017) Improved initial clustering center selection algorithm for k-means. In: 2017 Signal processing: algorithms, architectures, arrangements, and applications (SPA). IEEE, pp 275–279

  24. Fränti P, Sieranoja S (2019) How much can k-means be improved by using better initialization and repeats? Pattern Recogn 93:95–112

    Article  Google Scholar 

  25. Zhou P, Chen J, Fan M, Du L, Shen YD, Li X (2020) Unsupervised feature selection for balanced clustering. Knowl-Based Syst 193:105417

    Article  Google Scholar 

  26. Torrente A, Romo J (2021) Initializing k-means clustering by bootstrap and data depth. J Classif 38(2):232–256

    Article  MathSciNet  MATH  Google Scholar 

  27. Fränti P (2018) Efficiency of random swap clustering. J Big Data 5(1):1–29

    Article  MathSciNet  Google Scholar 

  28. Tîrnăucă C, Gómez-Pérez D, Balcázar JL, Montaña JL (2018) Global optimality in k-means clustering. Inf Sci 439:79–94

    Article  MathSciNet  MATH  Google Scholar 

  29. Rahman MA, Islam MZ (2018) Application of a density based clustering technique on biomedical datasets. Appl Soft Comput 73:623–634

    Article  Google Scholar 

  30. Limwattanapibool O, Arch-int S (2017) Determination of the appropriate parameters for k-means clustering using selection of region clusters based on density dbscan (srcd-dbscan). Expert Syst 34(3):e12204

    Article  Google Scholar 

  31. Giovanis DG, Shields MD (2020) Data-driven surrogates for high dimensional models using gaussian process regression on the grassmann manifold. Comput Methods Appl Mech Eng 370:113269

    Article  MathSciNet  MATH  Google Scholar 

  32. Chien SH, Wang JH, Cheng MY (2020) Performance comparisons of different observer-based force-sensorless approaches for impedance control of collaborative robot manipulators. In: 2020 IEEE conference on industrial cyberphysical systems (ICPS), vol 1. IEEE, pp 326–331

  33. Li H, Liu J, Yang Z, Liu RW, Wu K, Wan Y (2020) Adaptively constrained dynamic time warping for time series classification and clustering. Inf Sci 534:97–116

    Article  MathSciNet  MATH  Google Scholar 

Download references

Acknowledgements

This work was supported by the National Natural Science Foundation of China under Grant No. 52090054 and 52188102, and Natural Science Foundation of Hubei Province, China under Grant No. 2020CFA077.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Huan Zhao.

Ethics declarations

Conflict of Interests

The authors declare that they have no conflict of interest.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendix:: Convergence of intrinsic K-means

Appendix:: Convergence of intrinsic K-means

The convergence of the proposed intrinsic K-means algorithm is crucial. The calculation to minimize objective function (6) is an NP-hard problem. This can be proven as follows.

Loss function

$$ L(\mathbf{\mu},\mathbf{P},D) = \sum\limits_{n = 1}^{N} {\sum\limits_{k = 1}^{K} {{D_{nk}}{{|| {{{\text{Log}}_{{\mathbf{\mu}_{k}}}}{\mathbf{p}_{n}}} ||}^{2}}} } $$
(19)

where Dnk = 1 if pnDk; otherwise, Dnk = 0 and \({{|| {{{\text {Log}}_{{\mathbf {\mu }_{k}}}}{\mathbf {p}_{n}}} ||}} = {||{\text {Log}}_{\mathbf {e}} \mathbf {\mathcal {A}}_{\mathbf {e}}^{\mathbf {\mu }_{k}} \mathbf {p}_{n} ||_{\mathbf {\mu }_{k}}}\).

E-step

When we update D from Dt− 1 to Dt, the distance between points on the manifolds is determined by the Riemannian geodesic, and can be calculated by (5).

$$ {D_{nk}} = \left\{ {\begin{array}{*{20}{c}} {1, \text{if} k = \arg {{\min }_{j}}{{|| {{{\text{Log}}_{{\mathbf{\mu}_{j}}}}{\mathbf{p}_{n}}} ||}^{2}}} \\ {0, \text{otherwise} } \end{array}} \right. $$
(20)

where \({|| {\text {Log}}_{{\mathbf {\mu }_{j}}}{\mathbf {p}_{n}} ||} = {||{\text {Log}}_{\mathbf {e}} \mathbf {\mathcal {A}}_{\mathbf {e}}^{\mathbf {\mu }_{j}} \mathbf {p}_{n} ||_{\mathbf {\mu }_{j}}}\); thus, we can obtain

$$ L({\mathbf{\mu}^{(t - 1)}}, \mathbf{P}, {D^{(t)}}) \leqslant L({\mathbf{\mu}^{(t - 1)}}, \mathbf{P}, {D^{(t - 1)}}) $$
(21)

where \({D^{(t)}} = \arg {\min \limits _{D}}L({\mathbf {\mu }^{(t - 1)}}, \mathbf {P}, D)\).

M-step

When we update μ from μ(t− 1) to μ(t), the current mean is determined by Karcher mean (Algorithm 3).

$$ \mathbf{\mu}_{k}^{(t)} = mean(D_{k}^{(t)}) = \arg \mathop {\min }\limits_{\mathbf{\mu} \in \mathcal{M}} {\mathbf{E}}\left[ {{{|| {{{\text{Log}}_{\mathbf{\mu}} }{\mathbf{p}^{(k)}}} ||}^{2}}} \right] $$
(22)

This is a decreasing process; thus, we can obtain

$$ L({\mathbf{\mu}^{(t)}},\mathbf{P},{D^{(t)}}) \leqslant L({\mathbf{\mu}^{(t - 1)}},\mathbf{P},{D^{(t)}}) $$
(23)

where \({\mathbf {\mu }^{(t)}} = \arg {\min \limits _{\mathbf {\mu }} }L(\mathbf {\mu } ,\mathbf {P},{D^{(t)}})\). Each iteration of the algorithm decreases the otherwise positive quantization error until the error reaches a fixed point.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Tan, C., Zhao, H. & Ding, H. Statistical initialization of intrinsic K-means clustering on homogeneous manifolds. Appl Intell 53, 4959–4978 (2023). https://doi.org/10.1007/s10489-022-03698-8

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10489-022-03698-8

Keywords

Navigation