Abstract
Minimum spanning trees (MSTs) provide a convenient representation of datasets in numerous pattern recognition activities. Moreover, they are relatively fast to compute. In this paper, we quantify the extent to which they are meaningful in low-dimensional partitional data clustering tasks. By identifying the upper bounds for the agreement between the best (oracle) algorithm and the expert labels from a large battery of benchmark data, we discover that MST methods can be very competitive. Next, we review, study, extend, and generalise a few existing, state-of-the-art MST-based partitioning schemes. This leads to some new noteworthy approaches. Overall, the Genie and the information-theoretic methods often outperform the non-MST algorithms such as K-means, Gaussian mixtures, spectral clustering, Birch, density-based, and classical hierarchical agglomerative procedures. Nevertheless, we identify that there is still some room for improvement, and thus the development of novel algorithms is encouraged.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Clustering methods aim at finding some meaningful partitions of a given dataset in a purely unsupervised manner. They have proven useful in abundant practical applications; e.g., in medical, environmental, and earth sciences or signal processing (Guo et al., 2023; Hwang et al., 2023; Zhao et al., 2023; Zhou et al., 2023). Up to this date, many partitional clustering approaches have been proposed; for an overview, see the works by Wierzchoń and Kłopotek (2018), Blum et al. (2020), or Jaeger and Banks (2023). Methods to assess their usefulness include internal (Arbelaitz et al., 2013; Gagolewski et al., 2021; Halkidi et al., 2001; Jaskowiak et al., 2022; Maulik & Bandyopadhyay, 2002; Milligan & Cooper, 1985; Xu et al., 2020) and external cluster validity measures (Horta & Campello, 2015; Rezaei & Fränti, 2016; van der Hoef & Warrens, 2019; Wagner & Wagner, 2006; Warrens & van der Hoef, 2022) that were applied on various kinds of benchmark data (Dua & Graff, 2021; Fränti & Sieranoja, 2018; Gagolewski, 2022; Graves & Pedrycz, 2010; Thrun & Ultsch, 2020).
Given a dataset \(\textbf{X}=\{\textbf{x}_1,\dots , \textbf{x}_n\}\) with n points in \(\mathbb {R}^d\), the space of all its possible k-partitions, \(\mathcal {X}_k\), is very large. Namely, the number of divisions of \(\textbf{X}\) into \(k\ge 2\) nonempty, mutually disjoint clusters is equal to the Stirling number of the second kind, \(\big \{\begin{array}{l}n \\ k\end{array}\big \}=O(k^n)\).
In practice, to make the identification of clusters more tractable, many algorithms tend to construct simpler representations (samples) of the search space. For instance, in the well-known K-means algorithm by Lloyd (1957), we iteratively seek k (continuous) cluster centroids so that a point’s cluster belongingness can be determined through the proximity thereto. In hierarchical agglomerative algorithms, we start with n singletons, and then keep merging pairs of clusters that optimise some utility measure, e.g., the minimal or average intra-cluster distances, until we obtain k point groups (Murtagh, 1983; Müllner, 2011; Szekely & Rizzo, 2005). In divisive schemes, on the other hand, we start with one cluster consisting of all the points, and keep splitting it into smaller and smaller chunks based on some criteria.
From the perspective of clustering tractability, different spanning trees offer a very attractive representation of data. Any spanning tree representing a dataset with n points has \(n-1\) edges. If we remove \(k-1\) of them, we will obtain k connected components which we can treat as clusters; compare Fig. 1. This reduces the size of the sample of the search space to explore greatly: down to \({{n-1}\atopwithdelims (){k-1}}=O(n^{k-1})\), and usually \(k \ll n\). Some heuristics, such as greedy approaches, allow for further simplifications.
In particular, the minimum spanning tree (MST; the shortest dendrite) with respect to the Euclidean metric minimises the sum of pairwise distances. More formally, given an undirected weighted graph \(G = (V, E, W)\) with \(V=\{1,\dots ,n\}\), \(E=\{ \{u,v\}, u< v \}\), and \(W(\{u,v\})=\Vert \textbf{x}_u-\textbf{x}_v\Vert \), the minimum spanning tree \(T = \textrm{MST}(G) = (V, E', W')\), \(E' \subset E\), \(W' = W \vert _{E'}\) is a connected tree spanning V with \(E'\) minimising \(\sum _{\{u,v\}\in E'} W(\{u,v\})\). Note that in this paper, we assume that an MST is always unique. This can be assured by adding, e.g., a tiny amount of noise to the points’ coordinates in the case where there are tied distances.
MSTs are fast to compute. For general metrics, they require \(O(n^2)\) time; see the classic algorithms by Borůvka (1926), Jarník (1930)Footnote 1, and Kruskal (1956). In small-dimensional Euclidean spaces, further speed-ups are achievable. For instance, March et al. (2010) prove that the time complexity is \(\Omega (n\log n)\) for \(d=2\). An approximate MST can be computed as well (Naidan et al., 2019; Zhong et al., 2015).
Applications of MST-based clustering algorithms are plentiful: they are useful, e.g., in studying functional relationships of genes or recognising patterns in images (Xu et al., 2002; Yin & Liu, 2009). Unlike in the K-means algorithm which is connected to Voronoi’s diagrams, in our setting, the clusters do not necessarily have to be convex. MST-based clustering methods allow for detecting well-separated, relatively dense clusters of arbitrary shapes, e.g., spirals, connected line segments, or blobs; see Fig. 2 for an example. This is true at least in low-dimensional spaces, albeit not necessarily in general ones due to the so-called curse of dimensionality (Blum et al., 2020). Also, note that some statistical properties of MSTs were studied by Di Gesu and Sacco (1983) and Jackson and Read (2010a, 2010b).
In this work, we aim to review, unify, and extend a large number of existing approaches to clustering based on MSTs that yield a specific number of clusters, k. Studying an extensive set of benchmark data of low-dimensionality (so that the results can be assessed visually) and moderate sizes (\(n < 10{,}000\)), we determine which of them performs best. Furthermore, we quantify how well particular MST-based methods perform in general, and determine whether they are competitive relative to other popular clustering procedures.
We have structured the current article as follows. Section 2 reviews existing MST-based methods and introduces a few of their generalisations, which include the divisive and agglomerative schemes optimising different cluster validity measures (with or without additional constraints). In Sect. 3, we answer the question of whether MSTs can provide us with a meaningful representation of the cluster structure in a large battery of synthetic benchmark datasets. Then, we pinpoint the best-performing algorithms, and compare them with state-of-the-art non-MST approaches. Section 4 concludes the paper and suggests some topics for further research.
2 Methods
Table 1 lists all the methods that we examine in this study. Let us first describe the MST-based approaches in detail.
2.1 Divisive Algorithms over MSTs
Perhaps the most widely known MST-based method is the classic single linkage scheme (Wrocław Taxonomy, dendrite method, nearest neighbour clustering). It was proposed by Polish mathematicians Florek, Łukasiewicz, Perkal, Steinhaus, and Zubrzycki (1951). Gower and Ross (1969) noted that it can be computed using a divisive scheme: remove the \(k-1\) edges of the maximal lengths from the MST, and consider the resulting connected components as separate clusters.
Another divisive algorithm over MSTs was studied by Caliński and Harabasz (1974). They optimised the total within-cluster sum of squaresFootnote 2 (see Eq. 1 below). To avoid multiplying entities beyond necessity, let us introduce it in a more general language.
Let \(F:\mathcal {X}_l \rightarrow \mathbb {R}\) be some objective function that we would like to maximise over the set of possible partitionings of any cardinality l (not just k, which we treat as fixed). We will refer to it as a cluster validity measure. Moreover, given a subgraph \(G'=(V,E'')\) of \(G=(V,E)\) representing \(\textbf{X}\) such that \(G'\) has l connected components, let us denote by \(C(V, E'')=(X_1, \dots , X_l)\in \mathcal {X}_l\) the l-partition corresponding to these components.
In the single linkage scheme, the objective function is such that we maximise the sum of weights of the omitted MST edges. Furthermore, in the setting of the Caliński and Harabasz (1974) paper, we maximise (note the minus):
where \(\varvec{\mu }_{i}\) is the centroid (componentwise arithmetic mean) of the i-th cluster. Naturally, many other objective functions can be studied. For instance, Müller et al. (2012) considered the information-theoretic criterion based on entropyFootnote 3 which takes into account cluster sizes and average within-cluster MST edges’ weights:
where \(L_{i}\) denotes the sum of the weights of edges in the subtree of the MST representing the i-th cluster and \(n_i\) denotes its size. This leads to an algorithm called ITM. Its Python implementation is available at https://github.com/amueller/information-theoretic-mst.
We can consider many internal cluster validity indices as objective functions. Each of them leads to a separate, standalone clustering algorithm (many of which have not yet been inspected in the literature). They are denoted by “MST/D_...” in Table 1. We will study the measures recently reviewed by Gagolewski et al. (2021):
-
the indices by Ball and Hall (1965), Davies and Bouldin (1979, Def. 5), and Caliński and Harabasz (1974, Eq. 3) the latter being equivalent to the WCSS;
-
two versions of the Silhouette index by Rousseeuw (1987);
-
the generalisations of the Dunn (1974) index proposed by Bezdek and Pal (1998) (GDunn_dX_dY) and Gagolewski et al. (2021) (DuNN_M_X_Y);
-
the within-cluster near-neighbour count (WCNN_M) by the same authors.
Let us stress that Gagolewski et al. (2021) considered the optimisation of these measures in the space of all possible partitions, whereas here, we restrict ourselves to the space induced by MSTs.
2.2 Agglomerative Algorithms over MSTs
The single linkage was rediscovered by Sneath (1957), who introduced it as a general agglomerative scheme. Its resemblance to the famous Kruskal (1956) MST algorithm (and hence that an MST suffices to compute it) was pointed out by, amongst others, Gower and Ross (1969). Starting from n singletons, we keep merging the clusters containing the points joined by the edges of the MST in ascending order of weights. For a given MST with edges sorted increasingly, the amortised run-time is \(O(n-k)\) based on the disjoint-sets (union-find) data structure (Cormen et al., 2009, Chapter 21).
Given a cluster validity measure F, we can generalise this agglomerative approach as follows.
In the single linkage case, the objective is to maximise the sum of the weights of the unconsumed MST edges (edges in \(E'\setminus (E''\cup \{\{u,v\}\})\)), or, equivalently, find the minimal \(\{u,v\} \in E'\setminus E''\).
Unfortunately, the agglomerative procedure is oftentimes slow to compute; many cluster validity measures usually run in \(\Omega (n^2)\) time and, as typically k is much smaller than n, O(n) iterations of the routine are required overall. Furthermore, the objective functions might not be well-defined for singleton and small clusters; this is problematic for they are the starting point of Algorithm 2. Hence, due to an already large number of procedures in our study, in the sequel, we will only consider the agglomerative maximising of the aforementioned information criterion (2), leading to the algorithm which we denote by IcA (information criterion – agglomerative strategy) in Table 1.
2.3 Variations on the Agglomerative Scheme
Genie proposed by Gagolewski et al. (2016) is an example variation on the agglomerative single linkage theme, where the total edge lengths are optimised in a greedy manner, but under the constraint that if the Gini index of the cluster sizes grows above a given threshold g, only the smallest clusters take part in the merging. Thanks to this, we can prevent the outliers from being classified as singleton clusters.
Formally, let \((c_1,\dots ,c_l)\) be a sequence such that \(c_i\) denotes the cardinality of the i-th cluster in a given l-partition. The Gini index is defined as \(\textrm{G}(c_1,\dots ,c_l)=\sum _{i=1}^l (l-2i+1) c_{(i)}/ (n-1) \sum _{i=1}^l c_i\), where \(c_{(i)}\) denotes the i-th greatest value in the sequence. This index a measure of inequality of the cluster sizes: it takes the value of 0 if all components are equal, and yields 1 if all elements but one are equal to 0.
We will rely on the implementation of Genie included in the genieclust package for Python and R (Gagolewski, 2021). Given a precomputed MST, the procedure runs in \(O(n\sqrt{n})\) time.
The algorithm depends on the threshold parameter g. The original paper recommends the use of \(g=0.3\), but for some datasets, other settings may work better. In this study, we will only compare the results generated by \(g\in \{0.1, 0.3, 0.5, 0.7\}\), as we observed that the clusterings obtained for the thresholds in-between these pivots do not differ substantially.
Cena (2018) in her PhD thesis, noted that the manual selection of the right g threshold can be avoided by suggesting an agglomerative scheme optimising a cluster validity measure which uses a particular warm start. Namely, what we denote by Genie+Ic (\(k+l\)) in Table 1, is a variation of Algorithm 2 maximising the information criterion (2), but one that starts at an intersection of clusters returned by the Genie algorithm with different parameters (as usually one of the thresholds leads to a clustering of high quality). In this paper, for greater flexibility, we allow the initial partitions to be possibly more fine-grained.
We shall only consider \(H=\{0.1, 0.3, 0.5, 0.7\}\) and \(l\in \{0,5,10\}\), as we observe that other choices led to similar results. We have implemented this algorithm in the genieclust (Gagolewski, 2021) package.
Of course, any cluster validity index F may be taken instead of \(\textrm{IC}\) in Algorithm 4. However, our preliminary research suggests that amongst many of the measures considered, the information criterion works best. Due to space constraints, these results will not be included here.
2.4 Other MST-Based Methods
Other MST-based methods that we implemented for the purpose of this study include:
-
HEMST, which deletes edges from the MST to achieve the best attainable reduction in the standard deviation of edges’ weights (Grygorash et al., 2006);
-
CTCEHC, which constructs a preliminary partition based on the vertex degrees and then merges clusters based on the geodesic distance between the cluster centroids (Ma et al., 2021).
There are a few other MST-based methods in the literature, but usually they do not result in a given-in-advance number of clusters, k, a requirement which we impose in the next section for benchmarking purposes. For instance, Zahn (1971) determines the MST and deletes “inconsistent” edges (with weights significantly larger than the average weight of the nearby edges), but their number cannot be easily controlled.
We do not include the methods whose domain is not solely comprised of the information from MSTs (González-Barrios & Quiroz, 2003; Karypis et al., 1999; Mishra & Mohanty, 2019; Zhong et al., 2011, 2010). They construct MSTs based on transformed distances (Campello et al., 2015; Chaudhuri & Dasgupta, 2010), which use MSTs for very different purposes, such as auxiliary density estimation or some refinement thereof (Peter, 2013; Wang et al., 2009).
2.5 Reference Non-MST Methods
In the next section, we will compare the aforementioned algorithms against many popular non-MST approaches. They are denoted by an asterisk in Table 1. We consider two large open-source aggregator packages which provide consistent interfaces to many state-of-the-art methods. From scikit-learn in Python, see Pedregosa et al. (2011), we call the following:
-
K-means following the Lloyd (1957) approach with n_init=100 restarts from random initial candidate solutions, which is a heuristic that seeks cluster centres that minimise the within-cluster sum of squares;
-
the expectation-maximisation algorithm for estimating the parameters of a Gaussian mixture with settings n_init=100 and covariance_type=“full”, i.e., each mixture component has its own general covariance matrix (e.g., Dempster et al. 1977);
-
Spectral, which applies K-means (n_init=10) over a projection of the normalised Laplacian; amongst the 20 different parameter settings tested (affinity in {“rbf”, “laplacian”, “poly”, “sigmoid”} and \(\gamma \) in {0.25, 0.5, 1.0, 2.5, 5.0}), the best turned out to be affinity=“laplacian”, i.e., based on the kernel \(K(x, y) = \exp ( -\gamma \Vert x-y \Vert _1)\) with \(\gamma =5\) (e.g., Donath and Hoffman, 1973);
-
Birch proposed by Zhang et al. (1996), which incrementally builds the so-called clustering feature tree representing a simplified version of a given dataset (we tested 24 parameter sets: threshold in {0.005, 0.01, 0.025, 0.05, 0.1, 0.25, 0.5, 1.0} and branching_factor in {10, 50, 100}), and then clusters it using the Ward linkage algorithm; the best identified parameters were @threshold=0.01@ and branching_factor=50;
whereas from FCPS in R, see Thrun and Stier (2021), we call:
-
PAM (Partition around medoids), which is a robustified version of K-means that relies on the concept of cluster medoids rather than centroids (Rousseeuw & Kaufman, 1990);
-
Clara (Clustering LARge Applications), which runs the foregoing PAM on multiple random dataset samples and then chooses the best medoid set (Rousseeuw & Kaufman, 1990);
-
Diana (DIvisive ANAlysis clustering), which applies a divisive strategy to split clusters of the greatest diameters (Rousseeuw & Kaufman, 1990);
-
Adaptive density peaks, which uses nonparametric multivariate kernels to identify areas of higher local point density and the corresponding cluster centres (Rodriguez & Laio, 2014; Wang & Xu, 2015);
-
Hardcl (On-line update algorithm – Hard competitive learning) by Ripley (2007) and Softcl (Neural gas – Soft competitive learning) by Martinetz et al. (1993), both being a on-line version of K-means where centroids are iteratively moved towards randomly chosen points.
We also consider a few notable agglomerative hierarchical methods not based on MSTs:
-
The classical agglomerative hierarchical clustering approaches utilising the average, complete, Ward, centroid, median, and McQuitty linkages implemented in the fastcluster package for Python by Müllner (2011);
-
HDBSCAN by Campello et al. (2015) implemented in the hdbscan package for Python by McInnes et al. (2017), which uses a robustified version of the single linkage algorithm with respect to the so-called mutual reachability distance governed by different minPts parameter settings;
-
Minimax, which is another agglomerative algorithm whose linkage function depends on the radius of a ball enclosing all points in the two clusters about to be merged (protoclust package for R); see the papers by Murtagh (1983) and Bien and Tibshirani (2011);
-
MinEnergy (energy package for R), which is a variation on the Ward linkage proposed by Szekely and Rizzo (2005).
Furthermore, we include the clusterings that maximise 52 different cluster validity measures from the paper by Gagolewski et al. (2021).
We would like to emphasise that three of the above methods rely on feature engineering (selection of noteworthy features, scaling of columns, noise point or outlier removal, etc.), and were included as a point of reference for the algorithms that rely on “raw” distances. Spectral and Adaptive density peaks are kernel-based approaches which work in modified feature spaces, compare Wierzchoń and Kłopotek (2018) or Blum et al. (2020), and Gaussian Mixture can be thought of as a method that applies custom scaling of features in each cluster and thus does not rely on the original pointwise distance structure. Note that, overall, how to perform an appropriate feature engineering is a separate problem that should be carefully considered prior to an algorithm’s run.
3 Experiments
3.1 Clustering Datasets, Reference Labels, and Assessing the Similarity Thereto
We have gathered many synthetic datasets from numerous sources, including those by renowned experts in the field, see the works by Fränti and Sieranoja (2018), Fränti and Virmajoki (2006), Sieranoja and Fränti (2019), Rezaei and Fränti (2016), Ultsch (2005), Thrun and Ultsch (2020), Thrun and Stier (2021), Graves and Pedrycz (2010), Dua and Graff (2021), Karypis et al. (1999), McInnes et al. (2017), Bezdek et al. (1999), and Jain and Law (2005), as well as instances that we have proposed ourselves. All datasets are included in the benchmark suite for clustering algorithms version 1.1.0 which is available for download from https://github.com/gagolews/clustering-data-v1/releases/tag/v1.1.0 (Gagolewski, 2022). Each of them comes with one or more reference label vectors.
For tractability reasons, we only take into account the datasets with \(n < 10{,}000\). Moreover, we restrict ourselves to two- and three-dimensional instances (\(d\le 3\)) so that the meaningfulness of the reference labels and the clustering results can be inspected visually. In total, we have 61 datasets; see Table 2 for the complete list.
In the “FCPS” and “SIPU” batteries, 25 partitions come from the original dataset creators, i.e., external experts: Thrun and Ultsch (2020), Fränti and Sieranoja (2018), etc. As van Mechelen et al. (2023) and Ullmann et al. (2022) advocate for minimising conflicts of interest in clustering evaluation, in the sequel, the subset consisting of these 25 “third-party” label sets will also be studied separately.
Other reference partitions were created by the current authors: in Table 2, we mark them clearly with an asterisk. We should note that even though the “Graves” battery includes datasets proposed by Graves and Pedrycz (2010), the original source did not define some of the reference partitions clearly enough (the figures therein are in greyscale). As most clusters therein are well-separable, we filled this gap to the best of our abilities. Similarly, the “chameleon_*” and “hdbscan” datasets in the “Other” battery were equipped with labels based on the points’ density as we presumed that it was the intention of their original authors, Karypis et al. (1999) and McInnes et al. (2017).
Most importantly, the readers can independently inspect the sensibleness of the label vectors using the interactive data explorer tool available at https://clustering-benchmarks.gagolewski.com/. Many of the reference partitions are based on human Euclidean visual intuition, but some of them are also generated from different probability distribution mixtures. Overall, the benchmark datasets are quite diverse, which is important from the perspective of the results’ generalisability and the fact that there exist different reasonable concepts of what a cluster is (Hennig, 2015; van Mechelen et al., 2023). Our battery includes well-separated clusters, overlapping point groups, clusters of various densities and shapes (Gaussian-like blobs, thin but long point chains, circles, etc.), noise points, and outliers.
The clustering algorithms were run in a purely unsupervised manner: they are only fed with the data matrix \(\textbf{X}\in \mathbb {R}^{n\times d}\) and k as defined by the reference label vector, never with the particular true labels. As our aim is to study the behaviour of the algorithms on benchmark datasets of the form as proposed by their respective authors, no kind of data preprocessing (e.g., standardisation of variables, removal of noise points) was applied. Thus, all the algorithms under scrutiny were treated in the very same manner. In particular, all MST-based algorithms, K-means, and classical agglomerative methods (amongst others) rely on exactly the same information extracted from the raw Euclidean distance matrix. All of the computed partitions can be downloaded from and previewed at https://github.com/gagolews/clustering-results-v1.
As a measure of clustering quality, we consider the well-known adjusted Rand index (AR) by Hubert and Arabie (1985), which is given by:
where the confusion matrix \(\textbf{C}\) is such that \(c_{i, j}\) denotes the number of points in the i-th reference cluster that a given algorithm assigned to the j-th cluster, with \(c_{i,\cdot }=\sum _{j=1}^k c_{i,j}\), and \(c_{\cdot ,j}=\sum _{i=1}^k c_{i,j}\). AR counts the number of concordant pairs between two clusterings and is adjusted for chance: its expected value is 0 if two independent partitions generated from the same marginal distribution are given. The index yields the value of 1 only if two partitions are identical, and a negative value for partitions “worse than random”. For more properties of AR, see, e.g., the work by Warrens and van der Hoef (2022).
Some datasets come with many reference label vectors: the total number of unique partitions is 81. This is in line with the recommendations by Gagolewski (2022), see also Dasgupta and Ng (2009) and von Luxburg et al. (2012), who argued that there can be many possible ways to cluster a dataset and thus an algorithm should be rewarded for finding any of the reference partitions. Thus, for each clustering algorithm and dataset, the maximum of the adjusted Rand indices with respect to all reference label vectors was considered (e.g., for Graves/fuzzy, we take the maximum of five similarity scores).
It is also worth noting that 19 partitions include some points marked as noise: they do not belong to any actual cluster, but are supposed to make an algorithm’s job more difficult. They are ignored when computing the confusion matrix after the clustering has been performed.
If an algorithm failed to converge or generated an error, we set the corresponding AR to 0. This was the case only in six instances (twice for spectral and once for Softcl, MST/D_WCNN_25, MST/D_DuNN_25_Mean_Mean, and MST/D_DuNN_25_Min_Max).
3.2 Are MST-Based Methods Promising?
Recall that the number of possible partitions of an MST with n edges into k subtrees is equal to \((n-1)(n-2)\cdots (n-k+1)\). For all datasets with \(k\le 4\), we identified the true maximum of AR using the brute-force approach, i.e., considering all the possible partitions generated by the cancelling of edges from the MST. The remaining cases were too time-consuming to examine exhaustively. Therefore, we applied a tabu-like steepest ascent search strategy with at least 10 random restarts. Although there is no theoretical guarantee that the identified maxima are global, the results are promising as they agreed with the exact solutions for \(k\le 4\) and were no worse than the observable maximal AR scores for the MST-based methods.
Let us thus consider the “Max MST” rows in Table 3, which denote the aforementioned theoretically achievable maximal AR index values (a hypothetical “oracle” MST-based algorithm). We see that overall, MST-based methods are potentially very competitive, despite their simplicity. Only in \(6/61\simeq 10\%\) instances (\(2/25=8\%\) if we restrict ourselves to third-party labels), we obtained the AR index less than 0.8, namely: SIPU/s3, s4, WUT/z1, graph, twosplashes, and olympic (see Fig. 2 for the depiction of the last two datasets). Here, the minimum spanning trees with respect to the Euclidean distance between unpreprocessed points do not constitute valid representations of the feature space, as the clusters are too overlapping. Also, \(10/61\simeq 16\%\) (\(5/25\simeq 20\%\)) cases end up with AR < 0.95.
Nevertheless, we must note that some datasets are problematic (AR < 0.8) also for non-MST methods that we described in Sect. 3.4 (compare the “Max Obs. Non-MST” rows in Table 3). These are: SIPU/s3, s4, pathbased, Other/chameleon_t8_8k, WUT/cross, twosplashes, and Graves/parabolic.
Overall (compare “Max All”), \(53/61\simeq 87\%\) (\(20/25\simeq 75\%\)) instances are discovered correctly (AR \(\ge 0.95\)) by at least one clustering approach; more precisely, all but SIPU/s2, s3, s4, WUT/twosplashes, graph, mk3, olympic, and Graves/parabolic. These eight failures might mean that:
-
better clustering algorithms should be developed or taken into account,
-
some further data preprocessing must be applied in order to reveal the cluster structure; this is true for the WUT/twosplashes dataset which benefits from the standardisation of features,
-
there is something wrong with the reference label vectors, e.g., an alternative sensible partition exists but was not formalised, the clusters are overlapping too heavily in which case the partitional clustering algorithms should be replaced by soft/fuzzy ones.
3.3 Which MST-Based Algorithms Work Best?
That the MST-based methods are generally promising does not mean that we are in possession of an algorithm that gets the most out of the information conveyed by the minimum spanning trees, nor that a single strategy is always best. We should thus inspect which approaches and/or objective functions are more useful than others.
Figure 3 depicts, for each MST-based method, the adjusted Rand indices relative to the aforementioned “Max MST”, i.e., how well each algorithm compares to the best theoretically achievable solution.
The agglomerative Genie algorithm is the top performer. The approaches optimising the information criterion (Genie+Ic, ITM) also give high average relative AR. Optimising the recently-proposed near-neighbour-based index, DuNN_25_Min_Max, yields high median relative scores too.
As far as other “standalone” algorithms are concerned, HEMST and Single linkage exhibit inferior performance, and CTCEHC is comparable with the divisive Caliński–Harabasz criterion optimiser.
Quite strikingly, some well-established internal cluster validity measures promote clusterings of very poor agreeableness with the reference labels (Davies–Bouldin, SilhouetteW, some generalised Dunn indices). Choosing the wrong objective function to optimise over MST can lead to very poor results. This puts their meaningfulness into question. This is in line with the observation made by Gagolewski et al. (2021), where a similar study over the space of all possible partitionings was performed.
3.4 How MST-Based Methods Compare Against Other Clustering Approaches?
Let us compare the best MST-based approaches with other popular clustering algorithms mentioned in Sect. 2.5. Table 4 gives the summary of the AR indices over all 61 benchmark datasets and their subset consisting of 25 third-party labels only. In both cases, the best MST-based methods are high-performers: Genie_G0.3, Genie+Ic (k+0), and MST/D_DuNN_25_Min_Max identify approximately half of the reference partitions correctly (AR \(\ge 0.95\)).
The spectral and Gaussian mixture methods (which are not based on MSTs) have some potential leverage as they apply a form of feature engineering automatically. Despite this, they did not outperform their competitors, but overall they work very well too. However, their performance drops when we use the extended dataset repertoire that includes cases of noisy clusters and those of many different shapes. In particular, even if we exclude the WUT datasets created by a subset of the current authors, the “Graves” and “Other” batteries (which were proposed by other researches) feature 1 “difficult” (AR \(< 0.8\)) and 13 “easy” (AR \(\ge 0.95\)) cases for the Genie_G0.3 algorithm, and 8 “difficult” but only 5 “easy” instances for the spectral one. Overall, the average AR indices on the 40 first datasets from Table 2 are equal to 0.9 (increase from 0.85) and 0.79 (decrease from 0.83) for Genie_G0.3 and spectral, respectively.
On the other hand, the fact that the average and centroid linkages, K-means, and PAM methods were able to identify 8–10 partitions from the “SIPU” and “FCPS” subsets may indicate that those batteries consist of many datasets with isotropic Gaussian-like blob-shaped clusters.
Certain more recent methods: MinEnergy, CTCEHC, and ITM perform reasonably well. We should also take note of the quite disappointing performance of the HDBSCAN algorithm, which is essentially the (robustified) single linkage method applied on a version of the distance matrix that lessens the influence of potential outliers (where the minPts parameter plays the role of a smoothing factor). HDBSCAN is, theoretically, capable of marking some points as noise, but it did not do so on our benchmark data.
4 Conclusion
Naturally, no algorithm can perform best in all the scenarios. Hence, in practice, it is worth to consider the outputs of many methods at the same time. We demonstrated that the minimum spanning tree-based methods are potentially very competitive clustering approaches. MSTs are suitable for representing dense clusters of arbitrary shapes, and can be relatively robust to outliers in the outer parts of the feature space.
Many MST-based methods are quite simple and easy to compute: once the minimum spanning tree is determined (which takes up to \(O(n^2)\) time, but approximate methods exist too; e.g., Naidan et al. 2019), we can potentially get a whole hierarchy of clusters of any cardinality. This property makes them suitable for solving extreme clustering tasks, where the number of clusters k is large (Kobren et al., 2017). For instance, the Genie algorithm, which is the top performer in this study, needs \(O(n\sqrt{n})\) time, given a prebuilt MST, to generate partitions of all cardinalities. On the other hand, the well-known K-means algorithm is fast for small fixed ks only as each iteration thereof requires O(ndk) time.
Although MSTs can be distorted by introducing many noise points in-between the well-separable clusters, they can most likely be improved by appropriate feature engineering, e.g., scaling of data columns, noise point removal, modifying the distance matrix, automatic data projections onto simpler spaces, etc. (Blum et al., 2020; Campello et al., 2015; D’Urso & Vitale, 2022; Temple, 2023; Yin & Liu, 2009). Combining automated data preprocessing with clustering leads to a distinct class of methods: they should be studied separately, so we leave them as a topic for future research.
Overall, no single MST-based method surpasses all others, but the new divisive and agglomerative approaches we have proposed in this paper perform well on certain dataset types. However, compared to the theoretically achievable maxima, we note that there is still some room for improvement (see Table 3). Thus, the development of new algorithms is encouraged. In particular, it might be promising to explore the many possible combinations of parameters and objective functions we have left out due to the obvious space constraints in this paper.
Future work should involve the testing of clustering methods based on near-neighbour graphs and more complex MST-inspired data structures (Fränti et al., 2006; González-Barrios & Quiroz, 2003; Karypis et al., 1999; Zhong et al., 2011, 2010). Also, the application of the MST-based algorithms could be examined in the problem of community detection in graphs (Gerald et al., 2023).
It would be also interesting to inspect the stability of the results when different random subsets of benchmark data are selected or study the problem of overlapping clusters (Campagner et al., 2023). Moreover, let us again emphasise that have studied only low-dimensional datasets (\(d\le 3\)) so that the results could be assessed visually. A separate analysis, requiring most likely a different methodology, should be performed on medium- and large-dimensional data.
Finally, let us recall that we have only focused on partitional clustering methods that guarantee to return a fixed-in-advance number of clusters k. In the future, it would be interesting to allow for the relaxation of these constraints, e.g., study cluster hierarchies and soft clustering approaches.
Data Availability
All benchmark data are publicly available at https://clustering-benchmarks.gagolewski.com/ (Gagolewski, 2022). The project homepage features a visual dataset explorer. We used version 1.1.0 of the benchmark dataset battery, which can be fetched from https://github.com/gagolews/clustering-data-v1/releases/tag/v1.1.0. All computed partitions can be downloaded from and previewed at https://github.com/gagolews/clustering-results-v1. All cluster validity measures are implemented in the genieclust package for Python and R (Gagolewski, 2021). Our implementation of the HEMST and CTCEHC algorithms is available at https://github.com/lukaszbrzozowski/msts.
Notes
References
Arbelaitz, O., Gurrutxaga, I., Muguerza, J., Pérez, J. M., & Perona, I. (2013). An extensive comparative study of cluster validity indices. Pattern Recognition, 46(1), 243–256. https://doi.org/10.1016/j.patcog.2012.07.021
Ball, G., & Hall, D. (1965). ISODATA: A novel method of data analysis and pattern classification (Tech. Rep. No. AD699616). Stanford Research Institute.
Bezdek, J., Keller, J., Krishnapuram, R., Kuncheva, L., & Pal, N. (1999). Will the real Iris data please stand up? IEEE Transactions on Fuzzy Systems, 7(3), 368–369. https://doi.org/10.1109/91.771092
Bezdek, J., & Pal, N. (1998). Some new indexes of cluster validity. IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics), 28(3), 301–315. https://doi.org/10.1109/3477.678624
Bien, J., & Tibshirani, R. (2011). Hierarchical clustering with prototypes via Minimax linkage. The Journal of the American Statistical Association, 106(495), 1075–1084.
Blum, A., Hopcroft, J., & Kannan, R. (2020). Foundations of data science. Cambridge University Press. Retrieved from https://www.cs.cornell.edu/jeh/book.pdf
Borůvka, O. (1926). O jistém problému minimálním. Práce Moravské Přírodovědecké Společnosti v Brně, 3, 37–58.
Caliński, T., & Harabasz, J. (1974). A dendrite method for cluster analysis. Communications in Statistics, 3(1), 1–27. https://doi.org/10.1080/03610927408827101
Campagner, A., Ciucci, D., & Denoeux, T. (2023). A general framework for evaluating and comparing soft clusterings. Information Sciences, 623, 70–93. https://doi.org/10.1016/j.ins.2022.11.114
Campello, R.J.G.B., Moulavi, D., Zimek, A., & Sander, J. (2015). Hierarchical density estimates for data clustering, visualization, and outlier detection. ACM Transactions on Knowledge Discovery from Data, 10(1), 5:1–5:51. https://doi.org/10.1145/2733381
Cena, A. (2018). Adaptive hierarchical clustering algorithms based on data aggregation methods (Unpublished doctoral dissertation). Polish Academy of Sciences: Systems Research Institute. (In Polish).
Chaudhuri, K., & Dasgupta, S. (2010). Rates of convergence for the cluster tree. Advances in neural information processing systems (pp. 343–351).
Cormen, T., Leiserson, C., Rivest, R., & Stein, C. (2009). Introduction to algorithms. MIT Press and McGraw-Hill.
Dasgupta, S., & Ng, V. (2009). Single data, multiple clusterings. Proceedings NIPS Workshop Clustering: Science or Art? Towards Principled Approaches.
Davies, D.L., & Bouldin, D.W. (1979). A cluster separation measure. IEEE Transactions on Pattern Analysis and Machine Intelligence, PAMI–1(2), 224–227. https://doi.org/10.1109/TPAMI.1979.4766909
Dempster, A., Laird, N., & Rubin, D. (1977). Maximum likelihood from incomplete data via the EM algorithm. Journal of the Royal Statistical Society, Series B, 39(1), 1–22. https://doi.org/10.1111/j.2517-6161.1977.tb01600.x
Di Gesu, V., & Sacco, B. (1983). Some statistical properties of the minimum spanning forest. Pattern Recognition, 16(5), 525–531.
Donath, W., & Hoffman, A. (1973). Lower bounds for the partitioning of graphs. IBM Journal of Research and Development, 17(5), 420–425. https://doi.org/10.1147/rd.175.0420
Dua, D., & Graff, C. (2021). UCI Machine Learning Repository. Irvine, CA. http://archive.ics.uci.edu/ml
Dunn, J. (1974). A fuzzy relative of the ISODATA process and its use in detecting compact well-separated clusters. Journal of Cybernetics, 3(3), 32–57. https://doi.org/10.1080/01969727308546046
D’Urso, P., & Vitale, V. (2022). A Kemeny distance-based robust fuzzy clustering for preference data. Journal of Classification, 39, 600–647. https://doi.org/10.1007/s00357-022-09420-0
Edwards, A. W. F., & Cavalli-Sforza, L. L. (1965). A method for cluster analysis. Biometrics, 21(2), 362–375. https://doi.org/10.2307/2528096
Eggels, A., & Crommelin, D. (2019). Quantifying data dependencies with Rényi mutual information and minimum spanning trees. Entropy, 21(2). https://doi.org/10.3390/e21020100
Florek, K., Łukasiewicz, J., Perkal, J., Steinhaus, H., & Zubrzycki, S. (1951). Sur la liaison et la division des points d’un ensemble fini. Colloquium Mathematicum, 2, 282–285.
Fränti, P., & Sieranoja, S. (2018). K-means properties on six clustering benchmark datasets. Applied Intelligence, 48(12), 4743–4759.
Fränti, P., & Virmajoki, O. (2006). Iterative shrinking method for clustering problems. Pattern Recognition, 39(5), 761–765.
Fränti, P., Virmajoki, O., & Hautamäki, V. (2006). Fast agglomerative clustering using a k-nearest neighbor graph. IEEE Transactions on Pattern Analysis and Machine Intelligence, 28(11).
Gagolewski, M. (2021). genieclust: Fast and robust hierarchical clustering. SoftwareX, 15, 100722. https://doi.org/10.1016/j.softx.2021.100722
Gagolewski, M. (2022). A framework for benchmarking clustering algorithms. SoftwareX, 20, 101270. Retrieved from https://clustering-benchmarks.gagolewski.com/https://doi.org/10.1016/j.softx.2022.101270
Gagolewski, M., Bartoszuk, M., & Cena, A. (2016). Genie: A new, fast, and outlier-resistant hierarchical clustering algorithm. Information Sciences, 363, 8–23.
Gagolewski, M., Bartoszuk, M., & Cena, A. (2021). Are cluster validity measures (in)valid? Information Sciences, 581, 620–636. https://doi.org/10.1016/j.ins.2021.10.004
Gerald, T., Zaatiti, H., Hajri, H., et al. (2023). A hyperbolic approach for learning communities on graphs. Data Mining and Knowledge Discovery. https://doi.org/10.1007/s10618-022-00902-8
González-Barrios, J. M., & Quiroz, A. J. (2003). A clustering procedure based on the comparison between the k nearest neighbors graph and the minimal spanning tree. Statistics & Probability Letters, 62, 23–34. https://doi.org/10.1016/S0167-7152(02)00421-2
Gower, J.C., & Ross, G.J.S. (1969). Minimum spanning trees and single linkage cluster analysis. Journal of the Royal Statistical Society. Series C (Applied Statistics), 18(1), 54–64.
Graham, R., & Hell, P. (1985). On the history of the minimum spanning tree problem. Annals of the History of Computing, 7(1), 43–57.
Graves, D., & Pedrycz, W. (2010). Kernel-based fuzzy clustering: A comparative experimental study. Fuzzy Sets and Systems, 161, 522–543.
Grygorash, O., Zhou, Y., & Jorgensen, Z. (2006). Minimum spanning tree based clustering algorithms. Proceedings ICTAI’06 (pp. 1–9).
Guo, X., Yang, Z., Li, C., Xiong, H., & Ma, C. (2023). Combining the classic vulnerability index and affinity propagation clustering algorithm to assess the intrinsic aquifer vulnerability of coastal aquifers on an integrated scale. Environmental Research, 217, 114877. https://doi.org/10.1016/j.envres.2022.114877
Halkidi, M., Batistakis, Y., & Vazirgiannis, M. (2001). On clustering validation techniques. Journal of Intelligent Information Systems, 107–145,. https://doi.org/10.1023/A:1012801612483
Hennig, C. (2015). What are the true clusters? Pattern Recognition Letters, 64, 53–62. https://doi.org/10.1016/j.patrec.2015.04.009
Hero III, A.O., & Michel, O. (1998). Robust entropy estimation strategies based on edge weighted random graphs. In: A. Mohammad-Djafari (Ed.), Bayesian inference for inverse problems (vol. 3459, pp. 250 – 261). SPIE. https://doi.org/10.1117/12.323804
Horta, D., & Campello, R. (2015). Comparing hard and overlapping clusterings. Journal of Machine Learning Research, 16(93), 2949–2997.
Hubert, L., & Arabie, P. (1985). Comparing partitions. Journal of Classification, 2, 193–218.
Hwang, Y.-C., Ahn, H.-Y., Jun, J. E., Jeong, I.-K., Ahn, K. J., & Chung, H. Y. (2023). Subtypes of type 2 diabetes and their association with outcomes in Korean adults - A cluster analysis of community-based prospective cohort. Metabolism, 141, 155514. https://doi.org/10.1016/j.metabol.2023.155514
Jackson, T., & Read, N. (2010a). Theory of minimum spanning trees. II. Exact graphical methods and perturbation expansion at the percolation threshold. Physical Review E, 81, 021131.
Jackson, T., & Read, N. (2010b). Theory of minimum spanning trees. I. Meanfield theory and strongly disordered spin-glass model. Physical Review E, 81, 021130.
Jaeger, A., & Banks, D. (2023). Cluster analysis: A modern statistical review. Wiley Interdisciplinary Reviews: Computational Statistics, 15(3), e1597. https://doi.org/10.1002/wics.1597
Jain, A., & Law, M. (2005). Data clustering: A user’s dilemma. Lecture Notes in Computer Science, 3776, 1–10.
Jarník, V. (1930). O jistém problému minimálním (z dopisu panu O. Borůvkovi). Práce Moravské Přírodovědeckě Společnosti v Brně, 6, 57–63.
Jaskowiak, P., Costa, I., & Campello, R. (2022). The area under the ROC curve as a measure of clustering quality. Data Mining and Knowledge Discovery, 36, 1219–1245. https://doi.org/10.1007/s10618-022-00829-0
Karypis, G., Han, E., & Kumar, V. (1999). CHAMELEON: Hierarchical clustering using dynamic modeling. Computer, 32(8), 68–75. https://doi.org/10.1109/2.781637
Kobren, A., Monath, N., Krishnamurthy, A., & McCallum, A. (2017). A hierarchical algorithm for extreme clustering. Proceedings 23rd ACM SIGKDD’17 (pp. 255–264). https://doi.org/10.1145/3097983.3098079
Kruskal, J. B. (1956). On the shortest spanning subtree of a graph and the traveling salesman problem. Proceedings of the American Mathematical Society, 7, 48–50.
Lloyd, S. (1957). Least squares quantization in PCM. IEEE Transactions on Information Theory, 28, 128–137. (Originally a 1957 Bell Telephone Laboratories Research Report; republished in 1982) https://doi.org/10.1109/TIT.1982.1056489
Ma, Y., Lin, H., Wang, Y., Huang, H., & He, X. (2021). A multi-stage hierarchical clustering algorithm based on centroid of tree and cut edge constraint. Information Sciences, 557, 194–219. https://doi.org/10.1016/j.ins.2020.12.016
March, W.B., Ram, P., & Gray, A.G. (2010). Fast Euclidean minimum spanning tree: Algorithm, analysis, and applications. Proceedings of the 16th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (pp. 603–612). ACM.
Martinetz, T. M., Berkovich, S. G., & Schulten, K. J. (1993). ‘Neural-gas’ network for vector quantization and its application to time-series prediction. IEEE Transactions on Neural Networks, 4(4), 558–569.
Maulik, U., & Bandyopadhyay, S. (2002). Performance evaluation of some clustering algorithms and validity indices. IEEE Transactions on Pattern Analysis and Machine Intelligence, 24(12), 1650–1654. https://doi.org/10.1109/TPAMI.2002.1114856
McInnes, L., Healy, J., & Astels, S. (2017). hdbscan: Hierarchical density based clustering. The Journal of Open Source Software, 2(11), 205. https://doi.org/10.21105/joss.00205
Milligan, G. W., & Cooper, M. C. (1985). An examination of procedures for determining the number of clusters in a data set. Psychometrika, 50(2), 159–179.
Mishra, G., & Mohanty, S. K. (2019). A fast hybrid clustering technique based on local nearest neighbor using minimum spanning tree. Expert Systems with Applications, 132, 28–43. https://doi.org/10.1016/j.eswa.2019.04.048
Murtagh, F. (1983). A survey of recent advances in hierarchical clustering algorithms. The Computer Journal, 26(4), 354–359.
Müller, A., Nowozin, S., & Lampert, C. (2012). Information theoretic clustering using minimum spanning trees. Proceedings German Conference on Pattern Recognition. https://github.com/amueller/information-theoretic-mst
Müllner, D. (2011). Modern hierarchical, agglomerative clustering algorithms. arXiv:1109.2378
Naidan, B., Boytsov, L., Malkov, Y., & Novak, D. (2019). Non-metric space library (NMSLIB) manual, version 2.0 [Computer software manual]. Retrieved from https://github.com/nmslib/nmslib/blob/master/manual/latex/manual.pdf
Pál, D., Póczos, B., & Szepesvári, C. (2010). Estimation of rényi entropy and mutual information based on generalized nearest-neighbor graphs. Advances in Neural Information Processing Systems, 23.
Pedregosa, F., et al. (2011). scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12, 2825–2830.
Peter, S. (2013). Local density-based hierarchical clustering using minimum spanning tree. Journal of Discrete Mathematical Sciences and Cryptography, 16,. https://doi.org/10.1080/09720529.2013.778471
Prim, R. C. (1957). Shortest connection networks and some generalizations. Bell System Technical Journal, 36(6), 1389–1401. https://doi.org/10.1002/j.1538-7305.1957.tb01515.x
Rezaei, M., & Fränti, P. (2016). Set matching measures for external cluster validity. IEEE Transactions on Knowledge and Data Engineering, 28(8), 2173–2186. https://doi.org/10.1109/TKDE.2016.2551240
Ripley, B. D. (2007). Pattern recognition and neural networks. Cambridge University Press.
Rodriguez, A., & Laio, A. (2014). Clustering by fast search and find of density peaks. Science, 344(6191), 1492–1496. https://doi.org/10.1126/science.124207
Rousseeuw, P. J. (1987). Silhouettes: A graphical aid to the interpretation and validation of cluster analysis. Journal of Computational and Applied Mathematics, 20, 53–65. https://doi.org/10.1016/0377-0427(87)90125-7
Rousseeuw, P. J., & Kaufman, L. (1990). Finding groups in data. Wiley. https://doi.org/10.1002/9780470316801
Sieranoja, S., & Fränti, P. (2019). Fast and general density peaks clustering. Pattern Recognition Letters, 128, 551–558. https://doi.org/10.1016/j.patrec.2019.10.019
Sneath, P. (1957). The application of computers to taxonomy. Journal of General Microbiology, 17(1), 201–226. https://doi.org/10.1099/00221287-17-1-201
Szekely, G., & Rizzo, M. (2005). Hierarchical clustering via joint betweenwithin distances: Extending Ward’s minimum variance method. Journal of Classification, 22, 151–183. https://doi.org/10.1007/s00357-005-0012-9
Temple, J. (2023). Characteristics of distance matrices based on Euclidean, Manhattan and Hausdorff coefficients. Journal of Classification. https://doi.org/10.1007/s00357-023-09435-1
Thrun, M., & Stier, Q. (2021). Fundamental clustering algorithms suite. SoftwareX, 13, 100642. https://doi.org/10.1016/j.softx.2020.100642
Thrun, M., & Ultsch, A. (2020). Clustering benchmark datasets exploiting the fundamental clustering problems. Data in Brief, 30, 105501. https://doi.org/10.1016/j.dib.2020.105501
Ullmann, T., Hennig, C., & Boulesteix, A.-L. (2022). Validation of cluster analysis results on validation data: A systematic framework. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, 12(3), e1444. https://doi.org/10.1002/widm.1444
Ultsch, A. (2005). Clustering with SOM: U*C. Workshop on self-organizing maps (pp. 75–82). WSOM 2005.
van der Hoef, H., & Warrens, M. (2019). Understanding information theoretic measures for comparing clusterings. Behaviormetrika, 46, 353–370. https://doi.org/10.1007/s41237-018-0075-7
van Mechelen, I., Boulesteix, A.-L., Dangl, R., et al. (2023). A white paper on good research practices in benchmarking: The case of cluster analysis. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, e1511. https://doi.org/10.1002/widm.1511
von Luxburg, U., Williamson, R., & Guyon, I. (2012). Clustering: Science or art? In: I. Guyon et al. (Eds.), Proceedings ICML Workshop on Unsupervised and Transfer Learning (vol. 27, pp. 65–79).
Wagner, S., & Wagner, D. (2006). Comparing clusterings – An overview(Tech. Rep. No. 2006-04). Faculty of Informatics, Universität Karlsruhe (TH).
Wang, X., Wang, X., & Wilkes, D. M. (2009). A divide-and-conquer approach for minimum spanning tree-based clustering. IEEE Transactions on Knowledge and Data Engineering, 21(7), 945–958.
Wang, X., & Xu, Y. (2015). Fast clustering using adaptive density peak detection. Statistical Methods in Medical Research, 26(6). https://doi.org/10.1177/0962280215609948
Ward, J. H., Jr. (1963). Hierarchical grouping to optimize an objective function. Journal of the American Statistical Association, 58(301), 236–244. https://doi.org/10.1080/01621459.1963.10500845
Warrens, M., & van der Hoef, H. (2022). Understanding the adjusted Rand index and other partition comparison indices based on counting object pairs. Journal of Classification, 39, 387–509. https://doi.org/10.1007/s00357-022-09413-z
Wierzchoń, S., & Kłopotek, M. (2018). Modern algorithms of cluster analysis. Springer.
Xu, Q., Zhang, Q., Liu, J., & Luo, B. (2020). Efficient synthetical clustering validity indexes for hierarchical clustering. Expert Systems with Applications, 151, 113367. https://doi.org/10.1016/j.eswa.2020.113367
Xu, Y., Olman, V., & Xu, D. (2002). Clustering gene expression data using a graph-theoretic approach: An application of minimum spanning trees. Bioinformatics, 18(2), 536–545.
Yin, F., & Liu, C.-L. (2009). Handwritten Chinese text line segmentation by clustering with distance metric learning. Pattern Recognition, 42(12), 3146–3157. https://doi.org/10.1016/j.patcog.2008.12.013
Zahn, C. (1971). Graph-theoretical methods for detecting and describing gestalt clusters. IEEE Transactions on Computers, C-20(1), 68–86.
Zhang, T., Ramakrishnan, R., & Livny, M. (1996). BIRCH: An efficient data clustering method for large databases. Proceedings ACM SIGMOD International Conference on Management of data – SIGMOND ’96 (pp. 103–114).
Zhao, W., Ma, J., Liu, Q., & et al. (2023). Comparison and application of SOFM, fuzzy c-means and k-means clustering algorithms for natural soil environment regionalization in China. Environmental Research, 216, 114519. https://doi.org/10.1016/j.envres.2022.114519
Zhong, C., Malinen, M., Miao, D., & Fränti, P. (2015). A fast minimum spanning tree algorithm based on k-means. Information Sciences, 205, 1–17. https://doi.org/10.1016/j.ins.2014.10.012
Zhong, C., Miao, D., & Fränti, P. (2011). Minimum spanning tree based splitand- merge: A hierarchical clustering method. Information Sciences, 181, 3397–3410. https://doi.org/10.1016/j.ins.2011.04.013
Zhong, C., Miao, D., & Wang, R. (2010). A graph-theoretical clustering method based on two rounds of minimum spanning trees. Pattern Recognition, 43(3), 752–766. https://doi.org/10.1016/j.patcog.2009.07.010
Zhou, H., Bai, J., Wang, Y., Ren, J., Yang, X., & Jiao, L. (2023). Deep radio signal clustering with interpretability analysis based on saliency map. Digital Communications and Networks. https://doi.org/10.1016/j.dcan.2023.01.010
Acknowledgements
The authors would like to thank the Editor and Reviewers for providing their helpful comments that led to the improvement of earlier versions of this paper.
Funding
This research was partially supported by the Australian Research Council Discovery Project ARC DP210100227 (MG).
Author information
Authors and Affiliations
Contributions
MG: conceptualisation, methodology, data curation, software, visualisation, investigation, formal analysis, writing — original draft, writing — revision; AC: methodology, data curation, investigation; MB: software, data curation, investigation; ŁB: software, investigation
Corresponding author
Ethics declarations
Conflict of Interest
The authors declare no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Gagolewski, M., Cena, A., Bartoszuk, M. et al. Clustering with Minimum Spanning Trees: How Good Can It Be?. J Classif (2024). https://doi.org/10.1007/s00357-024-09483-1
Accepted:
Published:
DOI: https://doi.org/10.1007/s00357-024-09483-1