Model Selection Strategies for Determining the Optimal Number of Overlapping Clusters in Additive Overlapping Partitional Clustering

In various scientific fields, researchers make use of partitioning methods (e.g., K-means) to disclose the structural mechanisms underlying object by variable data. In some instances, however, a grouping of objects into clusters that are allowed to overlap (i.e., assigning objects to multiple clusters) might lead to a better representation of the underlying clustering structure. To obtain an overlapping object clustering from object by variable data, Mirkin’s ADditive PROfile CLUStering (ADPROCLUS) model may be used. A major challenge when performing ADPROCLUS is to determine the optimal number of overlapping clusters underlying the data, which pertains to a model selection problem. Up to now, however, this problem has not been systematically investigated and almost no guidelines can be found in the literature regarding appropriate model selection strategies for ADPROCLUS. Therefore, in this paper, several existing model selection strategies for K-means (a.o., CHull, the Caliński-Harabasz, Krzanowski-Lai, Average Silhouette Width and Dunn Index and information-theoretic measures like AIC and BIC) and two cross-validation based strategies are tailored towards an ADPROCLUS context and are compared to each other in an extensive simulation study. The results demonstrate that CHull outperforms all other model selection strategies and this especially when the negative log-likelihood, which is associated with a minimal stochastic extension of ADPROCLUS, is used as (mis)fit measure. The analysis of a post hoc AIC-based model selection strategy revealed that better performance may be obtained when a different—more appropriate—definition of model complexity for ADPROCLUS is used.


Introduction
In various domains of research, clustering models are used to summarize and reveal the structural mechanisms underlying data regarding a set of objects (e.g., object by variable data or object by object similarity data).Empirical and theoretical arguments may support the hypothesis that these underlying mechanisms can be captured by a grouping of the objects into homogenous clusters.A psychological researcher, for example, might investigate a patient population and collect data about the extent to which these patients express a number of clinical symptoms, like sleep disturbances, irritability, concentration problems, psychomotor retardation, or weight loss.In this example, the underlying mechanisms may be disclosed by grouping the patients into clusters such that patients from the same cluster show similar symptom profiles; the underlying mechanisms then pertain to the syndrome(s), like major depressive disorder or generalized anxiety disorder, that each patient group suffers from.As a second example, take network data, which are commonly encountered in computer, biological or social sciences (e.g., the Worldwide Web, power grids, food webs, metabolic networks, acquaintance networks and collaboration networks; Girvan & Newman, 2002;Kanaya et al., 2014;Scott, 2000), in which for each object pair information is available on the strength to which objects are connected (e.g., similarity data).In this type of data, natural divisions and community structures among the objects may be revealed by grouping strongly (inter)connected objects.
Three types of cluster structures (for a graphical representation of these three types, see Fig. 1 of Wilderjans et al., 2011) can be distinguished that may underlie such data (a detailed review of the three clustering types can be found in Van Mechelen et al., 2004).A first type of clustering is partitioning, which separates the objects into a set of exhaustive and mutually exclusive clusters, which implies that clusters are not allowed to overlap (i.e., no object can belong to more than a single cluster).Well-known partitioning methods include K-means (Forgy, 1965;Hartigan & Wong, 1979;Jancey, 1966;Macqueen, 1967;Steinheus, 1956), latent clustering (Lazarsfeld & Henry, 1968) and model-based clustering (Fraley & Raftery, 2002;Mclachlan & Chang, 2004).Although model-based clustering methods yield probabilistic or fuzzy clusterings (i.e., for each data object, the methods estimate the probability of belonging to each cluster), which could be interpreted as overlapping clusterings, these methods assume an underlying object partition.A basic assumption underlying the mixture model in a model-based cluster analysis is that each data object is generated from only a single mixture component (i.e., cluster), with the posterior probabilities indicating the probability that a specific data object is generated from each particular mixture component.Therefore, the assumptions of model-based clustering methods imply a partitioning of the data and do not assume that clusters overlap in the sense that objects can belong to more than one cluster.Hence, we categorize model-based clustering methods under the first type of cluster structures.
A second type of clustering pertains to hierarchical clustering methods (Duda et al., 2001;Gordon, 1987Gordon, , 1996;;Sneath & Sokal, 1973).These methods need a (dis)similarity object by object matrix as input.When originally having an object by variable data matrix, this matrix can be transformed into a (dis)similarity matrix by, for instance, computing correlations or Euclidean distances between the data objects.Hierarchical clustering methods yield a set of nested clusters in which each cluster can be obtained by merging objects and/or smaller clusters.The nestedness of the clusters implies that, for each pair of clusters, the intersection of two clusters consists of the smaller of both clusters or equals the empty cluster.Note that in case of the former, this indicates that object clusters overlap, which implies that some objects belong to multiple clusters.Most commonly used hierarchical clustering methods are agglomerative (i.e., repeatedly merging objects/clusters that are most similar) and divisive (i.e., recursively splitting clusters into smaller ones) methods (Manning et al., 2008).
Lastly, overlapping clustering methods can be distinguished as a third type of clustering.In this type of clustering, as is true for hierarchical clustering, an object may be a member of multiple clusters at the same time.The main difference between both types of clustering is that hierarchical methods only allow for cluster overlap in terms of nested clusters (i.e., two clusters are only allowed to overlap in the sense that one cluster is a subset of the other one), whereas in overlapping clustering methods the cluster overlap is not restricted in any way (see Fig. 1 of Wilderjans et al., 2011).The advantage of overlapping clustering, compared to the other two clustering types, is the possibility to account for more complex underlying cluster structures which may lead to more accurate representations of the sometimes intricate reality.In the patient by symptom data set, for example, in which clusters represent subjects with the same clinical syndrome, a patient might suffer from more than one syndrome at the same time (i.e., syndrome co-morbidity; Khanmohammadi et al., 2017).Such a patient should logically be assigned to multiple clusters.In the case of community selection in network data, an object might be strongly connected to two or more communities and it would therefore be sensible to allow the object to be a member of both clusters (Ding et al., 2016).
Besides interesting from a theoretical point of view (see above), the necessity of using overlapping clusters has also been acknowledged in empirical research (Banerjee et al., 2005).First, overlapping clustering is often successfully used in community detection problems (Tang & Liu, 2009;Wang et al., 2010;Fellows et al., 2011;Bonchi et al., 2013).Take as an example social networks in which subjects could be a member of multiple communities (Azaouzi et al., 2019), biological protein networks where proteins are part of various protein complexes simultaneously (Palla et al., 2008), or genetic networks, in which genes influence multiple cellular functions and as such belong to several metabolic pathways (Segal et al., 2003;Battle et al., 2004;Hastie et al., 2000).Second, in video classification, overlapping clustering is a necessary requirement to meaningfully group multi-genre movies based on the defining themes (e.g., a movie may simultaneously belong to the genres of action, horror and science fiction; Snoek et al., 2006).Third, in emotion detection, overlapping clustering methods are used to group text documents or music pieces based on the emotions they elicit (Wieczorkowska et al., 2006).Fourth, text clustering makes use of overlapping clustering methods to cluster text documents based on the topics discussed in them (Gil-Garc İa & Pons-Porrata, 2010;PėRez-Suȧrez et al., 2013b) or to learn patterns of cross-posting to several newsgroups (Lang, 1995).Finally, in the context of clustering in graphs, Khandekar et al. (2012) found that overlapping clusterings outperform non-overlapping ones in terms of optimizing the conductance over the clusters.In all of these examples, non-overlapping clusterings are not able to fully capture the rich cluster structure underlying the data.For instance, very similar multi-genre movies that, for example, contain both science fiction and horror themes may arbitrarily get allocated to separate clusters.Indeed, some of these movies may be placed in a science fiction cluster and separated from similar movies in which the science fiction theme is less pronounced compared to the horror theme.Indeed, any partitional algorithm will arbitrarily split an overlapping multi-genre movie cluster into two or more separate genre clusters, herewith obscuring the meaningful overlapping cluster structure.In this paper, we will only pay attention to methods for overlapping clustering.
In the literature, several methods were proposed to obtain overlapping clusters (for an overview, see Ben N'Cir et al., 2015;Baadel et al., 2016;Xie et al., 2013).In his review, Ben N' Cir et al. (2015) notes that most existing overlapping clustering methods can be conceived as extensions of more classical clustering methods, which according to the authors can be categorized into four classes of methods (i.e., method families): hierarchical, graphbased, generative and partitional methods.Besides these four main classes of methods, other methods were recently proposed that tackle the problem of overlapping clustering in alternative ways.For example, extensions of correlation clustering (Bonchi et al., 2013) and topological maps (Cleuziou, 2013), the latter being an extension of Self-Organizing-Maps (SOM; Kohonen, 1995).The four main classes of overlapping clustering methods will be briefly discussed here.First, hierarchical clustering methods aim to decrease the discrepancies between the original dissimilarities between the data objects and the dissimilarities predicted by the hierarchical structure.An example of an overlapping variant of the hierarchical clustering methods is the weak-hierarchies methods of Bertrand and Janowitz (2003).These methods are, in general, too restrictive regarding the overlap that is allowed.Second, graph-theory based methods, an example of which is Overlapping Clustering based on Relevance (OClustR; PėRez-Suȧrez et al., 2013a), are mostly used in the context of community detection in complex networks (Zhang et al., 2007;Davis & Carley, 2008;Wang & Fleury, 2011;Fellows et al., 2011).Although relevant in their applications, a clear disadvantage of these methods is their large computational complexity.Third, generative methods search for the unknown-overlapping-clusters in the data by fitting a mixture model.Mixture models for overlapping clusters (Banerjee et al., 2005;Heller & Ghahramani, 2007;Fu & Banerjee, 2008;Zhong & Ghosh, 2003), in which it is assumed that a data object is generated by simultaneously activating multiple mixture components, are flexible in the choice of probability distribution for each cluster.As a disadvantage, however, mixture models for overlapping clustering do not allow the user to easily exert control over the type and amount of overlap between clusters.
A fourth class of clustering methods are partitional methods that belong to the most commonly used methods for clustering.Within partitional methods, Ben N' Cir et al. (2015) makes a distinction between uncertain memberships and hard membership models.In uncertain memberships partitional models, a fuzzy, possibilistic or evidential framework is adopted to represent the clusters' memberships of each data object, which implies that each data object can belong to a larger or smaller extent to a (multiple) cluster(s).Hard membership partitional methods, on the contrary, rely on a binary function to model clusters' memberships (i.e., hard clustering), which implies that objects belong or do not belong to one (or more) cluster(s).Examples of uncertain memberships methods for overlapping clustering are fuzzy C-means (FCM; Lingras and West, 2004;Zhang et al., 2007;Höppner et al., 1999;Hruschka et al., 2009;Bezdek, 1981), possibilistic C-means (PCM; Krishnapuram & Keller, 1993;Krishnapuram & Keller, 1996) and possibilistic fuzzy C-means (PFCM; Pal et al., 2005), which modify the cluster results from a standard non-overlapping method into overlapping clusters.Other examples are Evidential c-means (ECM; Masson & Denoeux, 2008) and Belief c-means (BCM; Liu et al., 2012), which tailor the objective criterion in order to model cluster overlap.For hard membership partitional methods, a further distinction is made by Ben N'Cir et al. (2015) between additive and geometrical based methods.Additive methods model cluster overlap as the sum of representative objects of the clusters in question.As a consequence, these methods aim at minimizing the sum of squared residuals between the scores of each observation and the sum of clusters' representatives scores to which the observation in question belongs.Examples of additive overlapping clustering methods are Principal Cluster Analysis (PCL; Mirkin, 1987Mirkin, , 1990) ) and a related version of PCL that uses the Alternating Least Square algorithm (ALS; Depril et al., 2008), Lowdimensional Additive Overlapping Clustering (Depril et al., 2012) and Bi-clustering ALS (Wilderjans et al., 2013).Geometrical methods, on the contrary, conceive cluster overlap in terms of the barycenter on the related cluster representatives.As such, these methods minimize the sum of distances between each observation and the average-instead of the sum-of clusters' representatives to which the observation belongs to.Examples of geometrical methods for overlapping clustering are Overlapping K-means (OKM; Cleuziou, 2008), Parameterized R-OKM (Ben N'Cir et al., 2013), Overlapping k-Medoid (OKMED; Cleuziou, 2009), Weighted Overlapping K-means (WOKM; Huang et al., 2005;Cleuziou, 2009), Overlapping Partitioning Cluster (OPC; Chen & Hu, 2006), Multi-Cluster Overlapping K-means Extension (MCOKE; Baadel et al., 2015) and Kernel Overlapping K-means (KOKM; Ben N'Cir et al., 2010;Ben N'Cir & Essoussi, 2012).
In the remainder of this paper, we will only focus on partitional methods for overlapping clustering that imply a hard membership model (i.e., binary memberships) and that conceive cluster overlap as additive.Moreover, we will limit our attention to methods that can be applied to object by variable data.To obtain an overlapping (additive) object (hard) clustering based on object by variable data, Mirkin (1987Mirkin ( , 1990) ) developed the ADditive PROfile CLUStering (ADPROCLUS) method, which operates on the object by variable data directly.Note that the original ADCLUS model of Shepard and Arabie (1979) can only be used for analyzing an object by object (dis)similarity matrix.To improve the Principal Cluster Analysis algorithm of Mirkin (1987Mirkin ( , 1990)), Depril et al. (2008) proposed an Alternating Least Squares (ALS) algorithm to estimate an ADPROCLUS model from a given data set, and Wilderjans et al. (2011) provided easy-to-use software to perform ADPROCLUS.
An important question when performing any cluster analysis concerns the identification of the optimal number of clusters underlying the data, which boils down to a problem of model selection.To tackle this problem, one strategy consists of, first, computing a number of cluster solutions with an increasing number of clusters.Next, the optimal number of clusters is determined by means of a model selection procedure that searches for an optimal compromise between model fit and clustering complexity, with the latter being related to the number of clusters.For partitioning and hierarchical clustering methods, a broad range of such model selection procedures has been proposed (e.g., Akaike, 1974;Caliński and Harabasz, 1974;Hannan & Quinn, 1979;Schwarz, 1978;Sugar & James, 2003;Tibshirani et al., 2001b;Wang, 2010;Steinley & Brusco, 2011) and these approaches have been compared to each other in extensive simulation studies (e.g., Steinley & Brusco, 2011;Milligan & Cooper, 1985).
For overlapping clustering, however, almost no strategies for model selection have been proposed.Moreover, up to now, no systematic study has been conducted in which various model selection strategies for determining the optimal number of overlapping clusters underlying a given object by variable data set are compared.To address this gap in the literature, in this paper, various commonly used model selection strategies for other purposes than ADPROCLUS are adapted to or specifically tailored towards ADPROCLUS, and these strategies are evaluated in an extensive simulation study.The model selection strategies included in this study are as follows: Akaike's Information Criterion (AIC, Akaike, 1974), corrected AIC (AICc, Hurvich & Tsai, 1989), Bayesian Information Criterion (BIC, Schwarz, 1978), Hannan-Quinn Measure (HQM, Hannan & Quinn, 1979), CHull (Ceulemans et al., 2011;Wilderjans et al., 2012), Steinley and Brusco's Lower Bound Technique (LBT;Steinley & Brusco, 2011), the Caliński-Harabasz Index (CH;Caliński & Harabasz, 1974), the Dunn Index (DI;Dunn, 1973Dunn, , 1974)), the Krzanowski-Lai Index (KL; Krzanowski & Lai, 1988) and the Average Silhouette Width (ASW; Rousseeuw, 1987).Further, we also tailored two cross-validation based procedures for model selection in clustering to an ADPROCLUS context.
The remainder of this paper is organized in four main sections: In Section 2, we will present Mirkin's ADPROCLUS model (Mirkin 1987(Mirkin , 1990)), the associated loss function and the Alternating Least Squares (ALS) algorithm of Depril et al. (2008) to fit ADPRO-CLUS to given object by variable data.Next, the model selection strategies that will be compared in the simulation study are introduced in Section 3. In this section, we will also discuss the derivation of a likelihood function for a minimal stochastic extension of the ADPROCLUS model, which is needed for the calculation of information criteria for ADPROCLUS solutions.In Section 4, the design, procedure and results of an extensive simulation study to evaluate the performance of the model selection strategies to determine the correct number of overlapping clusters in ADPROCLUS will be presented and discussed.Finally, some concluding remarks are presented in Section 5.

Model
The additive profile clustering (ADPROCLUS) model, as introduced by Mirkin (1987Mirkin ( , 1990)), approximates an I × J object by variable data matrix X by an I × J model matrix M that can be decomposed into an I × K binary cluster membership matrix A and a K × J real-valued cluster profile matrix P, with K indicating the number of overlapping clusters.In particular, M is decomposed as: The entries a ik in matrix A represent whether object i belongs to cluster k (a ik = 1) or not (a ik = 0), with each object being allowed to belong to one cluster, to multiple clusters or to no cluster at all.The entries p kj of matrix P denote the prototypical value of variable j in cluster k.The k-th row of P therefore contains the variable profile for the k-th object cluster (k = 1, . . ., K).The predicted/reconstructed data entries m ij based on the ADPROCLUS model can be calculated as: ( 2 ) As a consequence, the estimated variable profile for a particular object i can be obtained by summing the cluster-specific variable profiles (i.e., rows of P)-hence the name additive profile clustering -associated with the clusters to which object i belongs (for an illustration of the ADPROCLUS model, see Wilderjans et al., 2011).

Loss Function
In practice, as data always contain (considerable amounts of) noise, no model matrix M that can be decomposed into A and P with K clusters (and K < min(I,J)) can be found that reconstructs X perfectly.As a consequence, discrepancies between X and M are allowed: with E being an I × J matrix of residuals denoting the discrepancies between X and M.
The aim of an ADPROCLUS analysis is therefore, given a fixed number of clusters K, to estimate a binary membership matrix A and a real-valued profile matrix P such that model matrix M = AP reconstructs data matrix X as close as possible in least squares sense (i.e, sum of squared residuals).In particular, the following least squares loss function needs to be minimized: lf (A, P ) = X − AP 2 F , (4) where .F indicates the Frobenius matrix norm (i.e., the Euclidean norm applied to the entries of the matrix).The loss function in (4) computes the sum of squared residuals between X and M, which becomes clear when the loss function is written element-wise: with e ij being the elements of the residual matrix E. Depril et al. (2008) proposed three algorithms for fitting the ADPROCLUS model to data and showed that these algorithms outperform the initial Principal Cluster Analysis (PCL) algorithm, which is also denoted as SEFIT, of Mirkin (1987Mirkin ( , 1990)).In this paper, we will use the ALS lf 2 -algorithm of Depril et al. (2008), which is an alternating least squares (ALS) algorithm for the minimization of loss function (4).This algorithm was adopted as, overall, this algorithm performs better or at least as good than the other algorithms presented in Depril et al. (2008).Moreover, this algorithm is straightforward to program and is not as time-consuming as the other algorithms from Depril et al. (2008).In this algorithm, starting from an initial random or (pseudo-)rational estimate of A (see Section 2.3.2), the following two steps are alternated until convergence: re-estimating P conditional upon A and updating A conditional upon P. The algorithm is considered converged when updating A and P does only yield a negligible decrease in the value of the loss function (e.g., a relative decrease smaller than some pre-determined tolerance value, like .000001).

Alternating Least Squares Algorithm
As re-estimating P with A kept fixed boils down to a multivariate multiple regression problem, the update of P is given by the following closed-form expression: where (A A) −1 denotes the matrix inverse of A A. Note that the Moore-Penrose pseudoinverse (Moore, 1920) needs to be used for inverting A A when A is not of full column rank (for more details, see Depril et al., 2008).A can be re-estimated conditionally upon P row by row as the loss function is separable as follows Chaturvedi and Carroll (1994): where, when P is kept fixed, each term only depends on a different row of A. As a consequence, loss function (4) can be decreased globally by decreasing each term in (7) separately.To update a row of A for a given P, all possible binary membership patterns (i.e., 2 K in total) are evaluated and the binary pattern yielding the lowest loss value is retained.This procedure is applied such that each row of A is updated once.

Initialization Strategies for Membership Matrix A
To obtain an initial estimate of membership matrix A, which is used to start the ALS algorithm, three strategies can be used: a random, a rational or pseudo-rational strategy (for more information on these types of initial starting configurations, see Ceulemans et al., 2007).A random initial A is generated by drawing its entries independently from a Bernoulli distribution with parameter π = .5(i.e., a 50% chance for a 0 or a 1).A rational estimate of the initial A can be determined in two ways.First, by using the overlapping clustering that is obtained by running the SEFIT-algorithm of Mirkin (1987Mirkin ( , 1990)); we will call this the SEFIT-rational start.Second, by taking a random sample of K objects (i.e., rows) from data matrix X, collecting these samples in a matrix P and computing the optimal A conditionally upon this P; we will call this the A (P) -rational start.Finally, a pseudo-rational initial A can be obtained by slightly perturbating one of both rationally determined initial A's.In particular, a small percentage, like 20%, of the entries of a rationally obtained initial A are switched from zero to one or vice versa.

Multi-start Procedure and Data Preprocessing
As noted by Depril et al. (2008), loss function (4) has multiple local optimal solutions, implying that the solution obtained with the previously discussed ALS algorithm may strongly depend on the initial membership matrix A chosen.To minimize the risk of the algorithm retaining a suboptimal solution, a multi-start procedure is recommended (a similar multi-start procedure is also recommended for K-means, see Steinley, 2003).In such a procedure, the algorithm is run multiple times, each time starting with a different initial A.
The best solution, in terms of minimizing loss function (4), across all runs of the algorithm is retained as the final solution.Regarding the choice of initial A's, Depril et al. (2008) suggest to use a mix of multiple random and (pseudo-)rational starts.The clustering solutions obtained by the ADPROCLUS algorithm are strongly influenced by preprocessing steps, like multiplicative (e.g., normalization) and additive (e.g., centering) transformations of the variables.Based on the arguments discussed in Wilderjans et al. (2011), it is advisable to preprocess the raw data before performing ADPROCLUS.Two possible forms of preprocessing are as follows: First, divide the variables by their range or standard deviation to account for large between-variable differences in variance (Milligan & Cooper, 1988).Second, when required, convert raw scores into deviation scores from a mean, a reference level or a normative score (for more information, see Wilderjans et al., 2011).

Model Selection Strategies
In this section, we describe the various model selection strategies that will be compared in their capacity to determine the correct number of overlapping clusters for ADPRO-CLUS.First, six existing strategies are described that are (often) used in the context of partitioning and that quite easily can be adapted for use in ADPROCLUS.Next, four commonly encountered model selection strategies for partitioning are presented that cannot in a straightforward way be adapted for use in ADPROCLUS as they rely on the extent to which between-cluster distances are large relative to the within-cluster distances: the Caliński-Harabasz (CH;Caliński & Harabasz, 1974), the Dunn Index (DI;Dunn, 1973;Dunn, 1974), the Krzanowski-Lai Index (KL; Krzanowski & Lai, 1988) and the Average Silhouette Width (ASW; Rousseeuw, 1987).As such, these four strategies only make sense when clusters are sought that are internally cohesive and externally isolated (Cormack, 1971), which is not the case for overlapping clusters.Therefore, we tailored these methods in such a way that they can also be applied in the context of overlapping clustering.Finally, two (cross-) validation based strategies are introduced that were specifically tailored to an ADPROCLUS context.The total number of estimated ADPROCLUS parameters (i.e., IK binary memberships a ik and JK real-valued profile values p jk ) is taken as complexity measure fp for an ADPROCLUS solution with K clusters: Note that an extra parameter is added, which represents the residual variance σ 2 e that has to be estimated when fitting the minimal stochastic extension of ADPROCLUS (see Section 3.1.1).

Existing Model Selection Strategies Adapted to ADPROCLUS
Some of the strategies that will be presented, like strategies based on information criteria, depend on the fit of the ADPROCLUS model in terms of a (log-)likelihood value.As the ADPROCLUS model is a deterministic model in which no distributional assumptions are made regarding (the model parameters and) the residuals e ij , no such likelihood value can be calculated.Therefore, a minimal stochastic extension of the ADPROCLUS model is proposed that allows the construction of a likelihood function for the ADPROCLUS model.It will turn out that the cluster solution optimizing this likelihood function equals the solution optimizing the original least squares loss function (4), which justifies the use of this log-likelihood value as a measure of model (mis)fit for ADPROCLUS.

Likelihood Function
Combining (2) and (3) results in A minimal stochastic extension of the ADPROCLUS model can be constructed by assuming that the residuals e ij are independently drawn from N(0, σ 2 e ); this implies that the data entries x ij are assumed to be distributed as with σ 2 e = VAR(e ij ) being the residual error variance.Assuming all x ij 's being independent from each other (and identically distributed), results in the negative log-likelihood -l(X|M) of the whole data set X being equal to: where n = IJ refers to the number of terms in the likelihood function, which equals the number of data points x ij .Acknowledging that the error variance σ 2 e can be estimated as the log-likelihood can be simplified as It appears that the negative log-likelihood ( 13) can be optimized by minimizing its last term only (i.e., all other terms are constants), which fully depends on As a consequence, minimizing the ADPROCLUS least squares loss function ( 4) is equivalent to minimizing the negative log-likelihood function ( 11) associated with the proposed minimal stochastic extension of ADPROCLUS.In other words, the optimal ADPROCLUS solution minimizing the original least squares loss function (4) equals the solution optimizing the (negative) log-likelihood function (11).

CHull
The CHull method (Ceulemans et al., 2011;Wilderjans et al., 2012) is an automated procedure for determining the elbow in a scree-like plot (Cattell, 1966).Such a plot is often used for model selection purposes, like determining the number of components in Principal Components Analysis (PCA) or the number of clusters in a cluster analysis.In a scree plot, the (mis)fit f of various obtained solutions, which, depending on the type of analysis, may be eigenvalues or the amount of explained variance, is plotted against a measure fp of the complexity of these solutions, like the number of components or clusters.In the case of ADPROCLUS, the loss value (4) of the optimal solution may serve as a model misfit measure and the number of estimated parameters (8) as a model complexity value.To determine the optimal number of overlapping clusters, one searches for an elbow point in the scree plot in which the loss value (4) is displayed against the fp value (8) for solutions with increasing numbers of overlapping clusters k = 1, . . ., K.
CHull can be used to identify the elbow point in the scree plot in an automated way instead of visually, which may be quite subjective.In the CHull procedure, first, the solutions are determined that are located on the boundary of the convex hull of the plotted points.Depending on whether the fit measure pertains to model fit, like percentage explained variance, or model misfit, like sum of squared residuals, the upper or lower boundary of the convex hull, respectively, is sought for.Next, the solution on the convex hull is selected that has an optimal balance between model fit and complexity (for an overview of the six-step CHull procedure, see Wilderjans et al., 2012).Note that CHull can never select the least and most complex model.
In our simulation study, we will use CHull with fp = ((I + J) × K) + 1 (see ( 8)) as complexity measure.For the fit measure, we will evaluate two CHull strategies: one using the value of the least squares loss function (4) and one based on the value of the negative log-likelihood function ( 11), which are in the following referred to as CHull LSQ and CHull NLL, respectively.

Akaike Information Criterion (AIC)
Information criteria are commonly used to tackle model selection problems in a variety of statistical models, such as mixture analysis (Solka et al., 1998;Steele & Raftery, 2009), exploratory factor analysis (Preacher et al., 2013), generalized linear (mixed) models (Mclachlan & Peel, 2000;Müller & Stadtmüller, 2005) and K-means (Steinley & Brusco, 2011).One of the most commonly used information criterion is the Akaike Information Criterion (AIC, Akaike, 1974), in which twice the negative log-likelihood -l of the model is penalized by two times the number of estimated parameters fp in the model: The optimal model is the model with the smallest AIC value.

Corrected AIC (AICc)
As observed by Hurvich and Tsai (1989), AIC does not perform well in applications where the sample size is small (for a thorough discussion regarding the bias in AIC, see Bozdogan, 2000) as in these situations AIC has the tendency to select a too complex model (for a similar observation in the context of mixture of factor analyzers, see Bulteel et al., 2013).To correct for this bias, Hurvich and Tsai (1989) developed the corrected Akaike Information Criterion (AICc) and demonstrated its advantage in small-sample applications.Essentially, the AICc extends the AIC by penalizing additional parameters more heavily: where n indicates the sample size.Note that for ADPROCLUS, we take n = IJ as the data matrix X contains IJ values (and, as a consequence, IJ terms separately contribute to the likelihood function).Note that when the sample size becomes much larger than fp, the correction term approaches 0 and therefore becomes negligible.As a consequence, the AICc should generally be preferred over AIC (Burnham & Anderson, 2004) as it offers serious advantages in small-sample applications and no substantial disadvantages in large-sample applications.

Bayesian Information Criterion (BIC)
A central assumption to AIC is that each observation in the data set provides new and independent information about the underlying model (Bozdogan, 1987).However, Schwarz (1978) argued that this assumption may be unrealistic for large data sets and therefore proposed the Bayesian Information Criterion (BIC), which takes the sample size into account, as an alternative to AIC: A significant advantage of the BIC over AIC is that its probability to select the true model is increasing as the sample size gets larger, as long, of course, the true model is among the candidate models.In situations where the number of estimated parameters does not depend on the sample size, the probability of the BIC to select the true model approaches 1 (for more information, see Vrieze, 2012).

Hannan-Quinn Measure (HQM)
Similar to AICc and BIC, the Hannan-Quinn Measure (HQM, Hannan and Quinn, 1979) is proposed to alleviate AIC's tendency of selecting a too complex model.Although HQM was originally introduced in the context of autoregressive modelling, later it was adapted to other likelihood-based models (see, Claeskens & Hjort, 2008).The HQM can be calculated as In the simulation study, all four information criteria will be included with (-l) being the negative log-likelihood value (11), fp equaling ((I + J) × K) + 1 and n = IJ.

Lower Bound Technique (LBT)
Steinley and Brusco (2011) proposed the Lower Bound of SSE Technique (LBT) for determining the optimal number of clusters in K-means clustering.Note that the ADPROCLUS and the K-means clustering model are mathematically very similar.In particular, there are only two differences between both clustering models.A first difference is that in K-means the binary membership matrix is restricted to take the form of a partition matrix, which implies that each row of the membership matrix is constrained to sum to one, whereas no such constraint is imposed on the binary membership matrix of the ADPROCLUS model.
A second difference between K-means and the ADPROCLUS model is that in K-means the profile matrix contains the cluster-specific variable means, whereas this is not the case for ADPROCLUS due to the cluster overlap.Because of this mathematical similarity between the K-means and the ADPROCLUS model, the LBT can be applied without modification to an overlapping clustering context.Generally, the LBT is a normalized index that is defined as where SST refers to the sum of the squared entries of X after centering the variables by subtracting the variable mean from each data point, and SSE K equals the loss value for a solution with K clusters (for ADPROCLUS: loss function value (4)).SSE K min indicates the lower bound for a solution with K clusters and is calculated as: with λ (XX ) denoting the eigenvalues of XX , ordered from largest to smallest (λ ).As such, LBT compares the observed SSE K -value for a K-cluster solution to a minimal value SSE K min that is associated with that particular number of clusters K.The optimal solution is the solution for which |LBT K | is minimal.In an extensive simulation study on K-means, the LBT outperformed traditional model selection strategies, such as the CH Index, the BIC, and Wilk's (Steinley & Brusco, 2011).

The Cali ński-Harabasz (CH) Index
The Caliński-Harabasz Index (CH; Caliński & Harabasz, 1974), which is also known as the variance ratio criterion, was originally proposed to determine the number of clusters in the context of a non-overlapping clustering problem.The CH Index is computed as: where k is the number of (non-overlapping) clusters, N the total number of data points, SS the overall within-cluster variance (i.e., total within-cluster sum of squared differences) and SS (k)  B the overall between-cluster variance (i.e., total between-cluster variance in the cluster centroids) for the best solution with k clusters.As a small SS W -which always will become smaller when k increases-indicates that clusters are homogeneous and that cluster centroids are nicely spread out, which implies a large SS B , the largest value for the CH Index points at the optimal number of clusters k that should be retained.
In the context of overlapping clusters, however, the CH Index is not a good measure to determine the optimal number of-overlapping-clusters as the CH Index searches for a clustering with homogeneous clusters that are clearly spread out, which does not make sense for overlapping clusters.As a solution, in order to adapt the CH Index to the context of overlapping clustering, we make use of the fact that each clustering with k overlapping clusters-at least in the case of a hard clustering (i.e., binary memberships)-can always be transformed into a clustering with maximally 2 k non-overlapping clusters (e.g., for a clustering with k = 2 overlapping clusters, one out of 2 k = 2 2 = 4 possible membership patterns-00, 10, 01 or 11-is allocated to each object, which implies a clustering of the objects into-maximally-four non-overlapping clusters).Now, the CH Index can be computed by considering the non-overlapping clustering that is derived from the overlapping clustering, herewith, however, using a value of 2 k for the number of clusters k (e.g., for 2, 3, and 4 clusters, we used a value of k of 4, 8 and 16, respectively) in formula (20) and calculating SS (k)  W and SS

(k)
B based on the non-overlapping clusters and associated cluster centroids.

The Dunn Index (DI)
The Dunn Index (DI; Dunn, 1973Dunn, , 1974) is a cluster validity index that aims at identifying the value for k that yields a solution with compact and well separated clusters, where the variance between cluster members (i.e., the within-cluster variance) is small and the cluster means are sufficiently far apart from each other.The Dunn Index can be computed as: with C i , C j and C l denoting cluster i, j and l, δ(C i , C j ) = min x∈C i ,y∈C j d(x, y) , (C l ) = max x,y∈C l d(x, y) , d(x, y) the Euclidean distance between objects x and y and k the considered number of clusters.The optimal number of clusters k is identified by choosing the solution with k clusters for which the Dunn Index ( 21) is maximal.As the Dunn Index-for the same reason as for the CH Index-can only be used in the context of non-overlapping clusters, we adapted the Dunn Index in a similar way as the CH Index by computing DI for the non-overlapping clustering with 2 k clusters that can be derived from the overlapping clustering with k clusters (see Section 3.1.8).

The Krzanowski-Lai Index (KL)
Similar to the CH Index, the Krzanowski-Lai Index (KL; Krzanowski & Lai, 1988) uses the within-cluster sum of squares as a criterion for determining the optimal number of clusters k.The KL Index for a solution with k clusters is defined as where DIFF k denotes a scaled difference between the within-cluster sum of squares (SS W ) of the two sequential clustering solutions with k − 1 and k clusters: where J is the number of variables.The value of k maximizing the KL Index ( 22) is regarded as the optimal number of clusters.Note that the KL Index in formula ( 22) shows some similarity with the st-ratio used in CHull.In particular, the st-ratio equals DIFF k DIFF k+1 , with , f k being a (mis)fit value for an ADPROCLUS solution with k clusters (e.g., SS W or the value of the least squares loss function or the negative loglikelihood function) and c k denoting the complexity of such a solution (e.g., k or fp).CHull is a flexible method as the user can decide which (mis)fit and complexity measure to use.It should be noted that due to the need for sequential clustering solutions, the KL index is not defined for the smallest and largest numbers of clusters k considered.To adapt the KL Index to additive overlapping clusterings with ADPROCLUS, we compute the KL Index for the non-overlapping clustering with 2 k clusters that can be derived from the overlapping clustering with k clusters (see Section 3.1.8),with 2 k used as value for k in (23).

The Average Silhouette Width (ASW)
The Average Silhouette Width, or Silhouette Index (ASW; Rousseeuw, 1987) was introduced as a graphical method for model selection in partitional clustering, specifically for K-means and K-medians clustering.In this vein, the ASW was designed to be useful in the case of compact, clearly separated and roughly spherical clusters.The ASW for a solution with k clusters is defined as where the Silhouette value of each object s i (i = 1, . . ., N) is defined as and where is a measure of the average dissimilarity of object i to all other objects in the same cluster C i .|C i | is the number of data points assigned to cluster C i and d(i, j ) is the (Euclidean) distance between data points i and j in cluster C i .The smaller the value of a i , the better we can regard the assignment of object i to the cluster C i (i.e., object i is quite close to the other objects of C i ).Similarly, calculates the smallest average distance between object i and all objects in any other cluster (of which object i is not a member).We can regard b i as a measure of how far object i is from the (objects in the) closest neighboring cluster.The optimal number of clusters k is chosen as the k for which the ASW in ( 24) is maximized.Note that the computation of a i (i.e., distance to objects in the same cluster) and b i (i.e., distance to objects in other clusters) only makes sense when an object partition with more or less compact clusters is obtained.We therefore, in order to be able to use ASW for model selection for ADPROCLUS, consider the partition with 2 k non-overlapping clusters that can be derived from the solution with k overlapping clusters (see Section 3.1.8)when computing the ASW statistic.

Tailored Model Selection Strategies
Cross-validation has been widely used in the literature to determine an optimal prediction model (Stone, 1974;Browne, 2000).Tibshirani et al. (2001a) adapted the idea of crossvalidation to a cluster analysis context and, based on this, Wang (2010) proposed a modified cross-validation procedure for determining the correct number of clusters in K-means clustering that is based on a clustering instability criterion.Wang (2010) defined clustering instability as with E being the expected value and Z n and Z * n indicating a pair of independent samples of size n; (Z n ; K) and (Z * n ; K) denote the cluster solution with K clusters that is obtained by applying the clustering algorithm under study to Z n and Z * n , respectively; d represents a distance metric between cluster solutions that is interpreted as a measure for clustering instability (i.e., when clusterings become more different, their distance becomes larger).
Inspired by Wang (2010), we created a clustering instability measure for ADPROCLUS and used this measure in two different (cross-)validation scenario's, with the first being computationally less expensive than the second one.

Simplified (Cross-)validation
We propose the following four-step procedure to measure cluster instability.First, an I × J data matrix X is split into three parts, X 1 , X 2 and X 3 with m, m and I − 2m observations, respectively.A sensible choice for m would be an integer value close to I 3 .Second, ADPROCLUS is applied to X 1 and X 2 , resulting in cluster membership matrices A 1 and A 2 and profile matrices P 1 and P 2 .Third, cluster membership matrices A (P 1 ) 3 and A (P 2 ) 3 for the observations in X 3 are computed that are conditionally optimal upon the profile matrices P 1 and P 2 , respectively (see Section 2.3 for information on how to compute the conditional optimal A given P).Finally, the clustering instability measure s is calculated as follows: with A (P 1 ) 3 P 1 and A the difference in (mis)fit for X 3 when using the profile matrices P 1 and P 2 .Note that a stable clustering is indicated by a small s-value.

Complex Cross-validation
In the "simplified" (cross-)validation procedure, only one part of the data is used as training set (i.e., X 1 and X 2 ) and the other part as validation set (i.e., X 3 ), without interchanging (i.e., crossing-over) the roles of both data parts.As such, the specific choice of which part of the data figures as training and which part as validation set influences the estimate of the crossvalidation error.To lower the variance in this estimate, we also developed a cross-validation procedure for ADPROCLUS that resembles fivefold cross-validation.In this procedure, data set X is first randomly split into five parts, called folds, with I 5 observations each; these five sets are labeled as X 1 3 , . . ., X 5 3 .Subsequently, for each X n 3 (n = 1, . . ., 5), the remaining 80% of observations are equally divided into two training sets, X n 1 and X n 2 , for which ADPROCLUS solutions with varying K are obtained.The clustering instability criterion ( 30) is calculated for each K and each of the five folds of the data and a mean score s av (K) across the five folds is obtained for each K.The optimal number of clusters K * is then defined as the K for which s av (K) is minimal.Note that this strategy can be used with any user-defined number of data folds n.In practice, often n = 5 or n = 10 is used.As the data set in this strategy is split only once into n folds, the particular split that is used may strongly influence the selection of the optimal K as the s av (K)-value for each K may be dependent on this particular split.To reduce the uncertainty in the s av (K)-values, a procedure may be implemented in which the "complex" cross-validation procedure is applied multiple times to X, each time with a different split of the objects in X into n folds.The final s av (K)-value is then obtained by averaging over the s av (K)-values obtained per split.In this paper, however, we will only study the cross-validation procedure with a single split.

Simulation Study
In this section, the results of a simulation study are presented to compare the effectiveness of thirteen strategies for determining the optimal number of overlapping clusters in ADPRO-CLUS.To determine their model selection performance, the model selection strategies are evaluated in terms of the accuracy and precision to which they identify the optimal number of clusters underlying the data (Steinley, 2003).Besides overall (i.e., across all generated data sets) differences between the proposed model selection strategies, we are also interested in how both performance aspects depend on the following data characteristics: the size of the data, the amount of noise in the data, the number of overlapping clusters underlying the data and the amount and type of cluster overlap.

Design and Procedure
Design To generate data sets with varying data characteristics, the following factors were systematically manipulated: • The size, I × J, of data matrix X, at two levels: 200 × 15 and 400 × 15; note that the number of variables J is fixed to 15; • The number of overlapping clusters, K, at two levels: 3 and 5; • The amount of cluster overlap, which is defined as the percentage of objects belonging to more than a single cluster, at three levels: 0%, 35% and 75%.• The degree of non-occurring overlapping membership patterns, which is defined as the number of distinct binary membership patterns indicating cluster overlap (i.e., multiple 1's in the membership pattern) that do not occur in A, at three levels: none, medium and high.Table 1 specifies the number of distinct non-occurring overlapping membership patterns for each condition; • The noise level, , which is defined as the percentage of the total variance in X that is accounted for by E, at three levels: 0.1, 0.4 and 0.7.Note that this implies that 10%, 40% or 70%, respectively, of the data is noise.
These data characteristics-and their manipulated levels-were chosen in order to generate data sets that are representative for typical data from psychology and the social sciences.In particular, the considered number of cases, variables and clusters, along with the adopted noise levels (except maybe for 70% of noise, which is included to seriously challenge the model selection strategies under study) are commonly encountered in social sciences data.The other two factors (i.e., the amount of cluster overlap and the degree of non-occurring overlapping membership patterns), which are hard to determine for empirical data, were chosen in order to investigate the performance of the model selection strategies considered under widely varying cluster overlap patterns.
Procedure Each data set X is obtained through independent generation of a cluster membership matrix A, a profile matrix P and noise E, and subsequently calculating X = AP+E.The values in the real-valued cluster profile matrix P are independently drawn from a normal distribution N(0, 10).The entries of E are independently drawn from a normal distribution N(0, σ 2 ), where the variance σ 2 is chosen such that the data contain the desired proportion of noise .To generate the binary cluster membership matrix A, first, for each possible membership pattern, it is determined exactly how often it should occur in the rows of A according to the levels of the data characteristics number of overlapping clusters, amount of cluster overlap and degree of non-occurring overlapping membership patterns, with the all-zero pattern always occurring for one out of 20 objects.To control the desired amount of cluster overlap, it is calculated how many membership patterns should show overlap (i.e., 0%, 35% or 75%) and, as a consequence, how many patterns should not show overlap (i.e., 95%, 60% or 20%, respectively).For the patterns showing no overlap, it is taken care of that each such pattern occurs the same amount of times (e.g., 60%  3 = 20% of the cases for the conditions with k = 3 clusters and an overlap of 35%) in the rows of A. Subsequently, to account for the desired degree of non-occurring overlapping membership patterns, from all possible membership patterns indicating cluster overlap, a certain number, as specified in Table 1, of patterns is independently sampled and removed from the set of overlapping membership patterns occurring in the rows of A. Note that for K clusters, the number of membership patterns showing overlap equals 2 K − (K + 1) as there are 2 K possible patterns in total of which there are K + 1 showing no overlap (i.e., the all-zero pattern and K patterns, one for each cluster, with a single one).It is guaranteed that each remaining overlapping membership pattern appears the same number of times in the rows of A. Finally, a matrix A is filled according to how often each membership pattern should be encountered (as calculated above) and the rows of A are permuted randomly.
A full crossing of all levels of all manipulated factors is not possible as in the 0%-clusteroverlap condition the degree of non-occurring overlapping membership patterns cannot be varied.Therefore, data sets without any cluster overlap (0% condition; referred to as NOL data sets) and data sets with cluster overlap (35% and 75% conditions; referred to as OL data sets) will be analyzed and discussed separately.For NOL data sets, the design yields a total of 2 (number of clusters) × 2 (data size) × 1 (amount of cluster overlap) × 1 (degree of non-occurring overlapping membership patterns) × 3 (noise level) = 12 conditions.For OL data sets, the design yields 2 (number of clusters) × 2 (data size) × 2 (amount of cluster overlap) × 3 (degree of non-occurring overlapping membership patterns) × 3 (noise level) = 72 conditions.For each condition, 10 replicate data sets are generated, resulting in a total of 840 simulated data sets.
For each generated data set, ADPROCLUS models with K = 1 up to K = 8 clusters are obtained.For each analysis, a multi-start procedure is performed consisting of the following (see Section 2.3): one SEFIT-rational start, nine pseudo-rational starts based on the SEFITrational start, five A (P) -rational starts and 15 random starts.Further, to lower the risk of the algorithm retaining a local optimal solution, also an alternative rational start is included that makes use of the fact that the cluster analysis is performed sequentially with an increasing number of clusters K.In particular, an alternative rational start for A is obtained by taking the optimal solution A (K−1) encountered for K − 1 clusters and adding one extra column filled with zeros and ones drawn independently from a Bernoulli distribution with probability 0.50; we call this the A (K−1) -rational start.Note that for K = 1, this type of rational start boils down to a random start.Subsequently, also nine pseudo-rational starts based on this alternative rational start were included.Finally, ten pseudo-rational starts were included that were obtained by perturbating (with each entry of A having a chance of 0.20 to be changed, see Section 2.3) the best solution encountered across the first 40 starts (see above); we call this the best-of-40-pseudo-rational start.Note that, in total, the multi-start procedure was run with 50 starts.Both cross-validation based procedures (see Section 3.2) were applied to each data set (also with 50 multi-starts).Regarding the split of the data into training and validation sets for performing the simplified (cross-)validation, we set the number of objects m in the training sets X 1 and X 2 equal to m = 2I 5 (i.e., each training set contains 40% of the observations and the validation set has 20% of the observations).For the complex crossvalidaton, we used fivefold cross-validation.The simulation study was programmed in R version 3.2.3 and performed on a supercomputer consisting of Intel Xeon E5-2660 CPUs with a clock frequency of 2.6 GHz.

Evaluation Criteria and Analysis Strategy
Evaluation Criteria Two measures were taken into account when assessing the performance of the model selection strategies: accuracy and precision (Steinley, 2003).The accuracy of a strategy pertains to whether or not the strategy in question correctly identifies the true number of clusters underlying the data.The precision of a strategy equals the absolute difference between the by the strategy estimated number of clusters K est and the true number of clusters K true ; for a particular data set d, the precision d equals d = |K est d −K true d |.Note that this measure is an adapted version (i.e., absolute instead of squared differences) of the precision measure proposed in Steinley and Brusco (2011).
Analysis Strategy A mixed-design (repeated measures-) analysis of variance (ANOVA) is conducted to investigate how the precision of a model selection strategy varies as a function of the manipulated data characteristics and the adopted model selection strategy.This mixed-ANOVA is performed for OL and NOL data sets separately, with the manipulated data characteristics acting as between-subject variables and the model selection strategy as a within-subject variable.As effect size measure for the mixed-design ANOVA, we use generalized eta squared (η 2 G ), as proposed by Bakeman (2005).Only significant main and interaction effects with a generalized eta squared larger than or equal to 0.15 (η 2 G ≥ 0.15), indicating a medium effect (Cohen, 1988;Bakeman, 2005), are reported.Note that no ANOVA is performed on the accuracy measure as this is a binary outcome variable.To investigate how accuracy varies in function of the data characteristics and the model selection strategies, a table with the mean accuracy for each level of each manipulated factor will be displayed for each strategy.

Results
First, the overall performance of the model selection strategies in terms of accuracy and precision is discussed and the pattern of over-and underestimation of the model selection strategies is investigated.Next, the influence of data characteristics on accuracy and precision is examined.Subsequently, the best performing model selection strategies are compared to each other in more detail.Finally, the computation time of the ADPROCLUS algorithm is discussed.

Accuracy , Precision and Degree of Over-and Underestimation of the Model Selection Strategies
Accuracy In Table 2, the mean accuracy of all model selection strategies is displayed, computed globally across all simulated data sets and as a function of the noise level .From this table it appears that, in data sets with overlap (OL), CHull using the (negative) log-likelihood as a measure of (mis-)fit (CHull NLL) performed best, selecting the correct model 460 out of 720 times (63.8%).CHull LSQ, using the loss function as (mis-)fit measure, performed slightly worse, with 434 out of 720 correctly identified models (60.3%).The procedure with the second best accuracy in OL data sets was Akaike Information Criterion (AIC; 48.2%), followed by the corrected AIC (AICc; 37.8%), both of which performed notably worse than CHull.The Dunn index (DI; 31.5%), the Lower Bound Technique (LBT; 30.8%), the Hannan-Quinn Measure (HQM; 30.6%) and complex cross-validation (CVc; 30.4%) all retrieved the correct number of clusters in about one third of all OL cases.Performance of the Average Silhouette Width (ASW; 19.9%), simplified cross-validation (CVs; 19.4%), the Krzanowski-Lai Index (KL; 16.3%), the Caliński-Harabasz Index (CH; 13.2%) and the Bayesian Information Criterion (BIC; 11.3%) was disappointing.A similar performance pattern was observed for data sets without overlap (NOL; see the values between brackets in Table 2): CHull NLL (75.8%) showed the highest accuracy, followed by CHull LSQ (70.0%).AIC (57.5%) and AICc (49.2%) had the second and third highest accuracy levels, respectively, followed by HQM (39.2%),CVc (32.5%),LBT (24.2%),DI (20.8%) and KL (20.0%).Again, CH, BIC and CVs performed very poorly, also with low noise, with only 6.7%, 16.7% and 19.2% correctly retrieved models, respectively.Finally, the ASW completely fails for data sets without overlap (1.7%).
The accuracy is displayed for the true number of clusters ( Correct ; in bold).underestimation and overestimation pertain to the precision for data sets that respectively under-and overestimate the true number of overlapping clusters Precision An examination of the precision of the model selection strategies, which is presented in Table 2 with smaller -values indicating the strategy being more precise, further confirmed CHull and AIC being the best and second best performing strategies, respectively, with all other strategies performing substantially worse.In OL data sets, CHull NLL ( = 0.67) and CHull LSQ ( = 0.73) performed best in terms of precision, followed by AIC ( = 1.30) which performed worse by a factor of 2 compared to CHull NLL.The fourth to seventh best strategies were complex ( = 1.56) and simplified ( = 1.67)CV, KL ( = 1.69), and AICc ( = 1.70), respectively.HQM ( = 2.02), LBT ( = 2.02), ASW ( = 2.10), DI ( = 2.26), CH ( = 2.37) and BIC ( = 2.76) performed worse by multiple orders of magnitude compared to CHull, with the average distance from the estimated to the true number of clusters being larger than two for all these five strategies (i.e., selecting a solution that has two clusters more or less than the optimal solution with the true number of clusters).A similar pattern was encountered for NOL data sets in which AIC ( = 0.95) performed worse than CHull NLL ( = 0.38) and CHull LSQ ( = 0.48) by a factor of 2 to 2.5 orders of magnitude.Similar to OL data sets, complex CV ( = 1.39),AICc ( = 1.40), simplified CV ( = 1.65) and KL ( = 1.68) come in fourth to seventh place, respectively, followed by HQM ( = 1.77),DI ( = 1.98),CH ( = 2.11), LBT ( = 2.13), ASW ( = 2.40) and BIC ( = 2.64).

Degree of Over-and Underestimation
To study the degree to which each model selection strategy over-or under-estimates the true number of clusters in ADPROCLUS, we computed the -measure, taking for each strategy the data sets in which respectively overestimation ( overestimation ) and underestimation ( underestimation ) occurred separately.The mean underestimation and overestimation for each model selection strategy, computed globally and per noise level, is presented in Table 2.In this table, one can see that, when not identifying the true number of clusters, the CH Index, LBT, ASW and all information-theoretic approaches (i.e., AIC, AICc, BIC, HQM) have a strong tendency to underestimate the number of clusters in ADPROCLUS.In particular, for these methods, underestimation occurs for all data sets (in both OL and NOL data sets) in which the estimated number of clusters does not equal the true number of clusters.As such, overestimation was never observed for these strategies.For AIC, this may be a rather surprising result as this strategy is known-as demonstrated in several studies-for its tendency to select too complex models (i.e., overestimation) in regression, partitioning methods and mixtures of factor analyzers (Steinley & Brusco, 2011;Vrieze, 2012;Bulteel et al., 2013).These studies, however, were performed under different conditions than the current study, which may-partly-explain this rather surprising result.Nevertheless, in general, which is in line with previous studies, AIC selects a more complex model (i.e., a smaller amount of underestimation) than AICc and BIC.For the Dunn index, overestimation occurred to a stronger extent than underestimation, whereas the opposite is true for the cross-validation based strategies.For the CHull strategies and the KL Index, over-and underestimation was encountered to the same extent.In general, under-and overestimation becomes more severe with increasing noise levels, except for the simplified cross-validation (CVs) strategy and the KL Index in which the effect of the noise level is less univocal.For the information-theoretic strategies (and also for LBT and CH), underestimation was (almost) maximal for the largest noise level, implying that for these data sets, the information-theoretic measures (almost) always erroneously select the most simple model with a single cluster.This pattern of results seems to suggest that for ADPRO-CLUS information-theoretic strategies punish the model fit too hard by the complexity of the model, indicating that the definition of model complexity fp used may not be optimal for these strategies (for a further discussion of this issue, see Section 4.4).

Effect of Data Characteristics on Accuracy and Precision
Accuracy The percentage of data sets for which the true number of clusters is identified (i.e., the mean accuracy Correct across data sets), overall and per model selection strategy, is displayed in Table 3 as a function of the levels of the manipulated data characteristics.From this table, it can be concluded that, overall, accuracy was more severely impaired for  data sets that contain more noise (i.e., = 67.5%,21.6% and 7.0% for = 0.1, 0.4 and 0.7, respectively), for data sets with more underlying clusters (i.e., = 35.7%for threeand 28.3% for five-cluster data sets), for data with a larger amount of cluster overlap (i.e., = 27.3% for large versus 36.3% and 33.3% for medium and no overlap, respectively) and for data with a larger degree of non-occurring overlapping membership patterns (i.e., for a high and medium degree = 28.1% and 32.8%, respectively, versus 34.1% for no overlap).For the number of objects in the data, only a small effect was observed (i.e., = 31.2%versus 32.8% for 200 and 400 objects, respectively).
Further, the choice of model selection strategy had a strong effect (η 2 G = 0.40 and 0.43 for OL and NOL data sets, respectively) on model selection performance, suggesting large performance differences between strategies.In particular, overall, both CHull versions are clearly more precise than AIC and the cross-validation based strategies, with BIC performing the worst (see Table 2).The main effect of the model selection strategies, however, is qualified by a model selection strategy by number of clusters interaction (η 2 G = 0.28 and 0.26 in OL and NOL data sets, respectively) and a model selection strategy by noise level interaction (η 2 G = 0.33 and 0.24 in OL and NOL data sets, respectively), both of which are visualized in Fig. 1 (for OL data sets).Regarding the interaction with the number of clusters (see the upper left panel in Fig. 1), in general, for both OL and NOL (not shown in figure) data sets, the performance deterioration with increasing numbers of underlying clusters is stronger for strategies that perform overall worse than for strategies that perform overall better.An exception to this are the cross-validation based strategies for which precision is nearly influenced by the number of clusters and AIC for which precision increased with increasing number of clusters.With respect to the interaction between model selection strategies and noise level (see the upper right panel in Fig. 1), in both OL and NOL (not shown in figure) data sets, the decrease in precision differs between the best and worst performing model selection strategies.In particular, for the overall best performing strategies (i.e., CHull NLL, CHull LSQ, and AIC), the deterioration in precision becomes stronger when the noise increases (i.e., a larger deterioration when going from = 0.4 to 0.7 compared to going from 0.1 to 0.4); for the overall worse performing strategies (i.e., BIC, ASW, CH, DI, HQM and the cross-validation based strategies), the strongest deterioration is observed for going from 0.1 to 0.4 instead of from 0.4 to 0.7.For KL, the precision is almost not influenced by the amount of noise in the data.Both interaction effects are qualified by a three-way model selection strategy by number of clusters by noise level interaction (η 2 G = 0.16 and 0.19 in OL and NOL data sets, respectively).In the lower panels of Fig. 1, in which this three-way interaction is displayed, one can see that for all model selection strategies-except the DI Index-the deterioration in precision with increasing noise level is stronger for data sets with five underlying clusters (bottom right panel) than for data sets with three underlying clusters (bottom left panel).Again, for KL, the precision is almost constant across noise levels.

Comparison between the Best Performing Model Selection Strategies: CHull NLL and AIC
A cross-tabulation of the accuracy of CHull NLL and AIC, which is depicted in Table 4, shows that for 48.9% of the data sets both strategies correctly identify the true underlying number of clusters, whereas both strategies fail to do so for 33.8% of the data sets.The 284 data sets for which both methods fail mainly belong to the very hard conditions that are characterized by a large noise level of = 0.7 (N = 212), a large amount of cluster overlap (N = 160) and five clusters underlying the data (N = 160).When looking at the data sets where AIC retrieves the correct model (i.e., for 49.5% of the data sets), CHull almost always also identifies the correct model (i.e., 99% of the cases); on the contrary, AIC only selects the correct model in 74.6% of the data sets for which CHull retrieves the correct model (i.e., for 65.6% of the data sets).When investigating the cross-tabulation across noise levels, CHull appears to consistently identify the correct model for the data sets for which AIC is correct (99.6%, 97% and 100% of the cases for = 0.1, 0.4 and 0.7, respectively), while in data sets where CHull retrieves the correct model, AIC is only correct at low noise levels and AIC's performance deteriorates drastically with increasing noise (100%, 63.7% and 2.9% of the cases for = 0.1, 0.4 and 0.7, respectively).It can be concluded that AIC does not have a substantial advantage over CHull and that the advantage of CHull over AIC becomes more prominent when the data contain more noise.In particular, the 140 data sets for which CHull-but not AIC-retrieved the correct model belong mainly to more difficult

Computation Time of the ADPROCLUS Algorithm
Table 5 displays the computation time for an average multi-start of the ADPROCLUS algorithm as a function of the number of clusters K and the number of objects I , the two factors that have the largest impact on the computation time.It appears that the computation time increased drastically both with increasing number of clusters K and increasing number of objects I .On average, the computation time for a solution with K clusters was about two to three times longer than the computation time for a solution with K − 1 clusters.Likewise, computation of ADPROCLUS solutions for data sets with I = 400 objects took, on average, 3.90 times longer compared to data sets with I = 200 objects.

Model Complexity: Post Hoc Analysis
As suggested by previous research on model selection in other statistical models (e.g., Hamerly & Elkan, 2004;Steinley & Brusco, 2011;Preacher et al., 2013), model complexity in this study was defined as the number of estimated model parameters: fp = ((I+J)×K)+1 (see Section 3).Defining complexity in this way, however, is based on the assumption that each parameter, regardless of its type (i.e., binary membership versus real-valued profile), should be weighed equally when determining the complexity of a model.For ADPRO-CLUS, one may doubt whether it makes sense to consider a binary membership parameter (from A) as important for the definition of model complexity as a real-valued profile parameter (from P).This implies that our definition of model complexity for ADPROCLUS may not be optimal.
The strong tendency of information-theoretic strategies to underestimate the number of clusters, as demonstrated in our simulation study, suggests that the associated test statistics penalize complex models too hard.In particular, the information criteria used in this study all select the model that has the lowest value on a test statistic that takes the following general form: where the first term decreases and the second term increases for more complex models, and λ is a weight that determines the penalty for increasing model complexity.As such, the number of clusters in ADPROCLUS may be underestimated when complex models are penalized too much.One reason for this over-penalization may be that the complexity of a model is not well defined (i.e., fp does not represents the true complexity of the model).
Applied to ADPROCLUS, it seems that the definition of fp in (8) yields a too large estimate of the complexity of the model.Relevant in this regard is the work of Lee ( 2001) on model complexity of additive clustering models.Bear in mind that additive clustering is similar-but not identical-to additive profile clustering.In particular, additive clustering models, like the ADCLUS model (Shepard & Arabie, 1979), operate on object by object (dis)similarity data instead of on object by variable data.Moreover, additive clustering tries to explain the observed (dis)similarities in terms of a set of overlapping clusters and associated weights that are summed across all clusters that are shared by the members of an object pair.Additive overlapping clustering, however, tries to reconstruct object by variable data by means of an overlapping object clustering and a set of associated cluster profiles that are summed across the clusters an object belongs to.In an attempt to derive a BIC statistic for additive clustering models, Lee (2001) used the number of clusters K as an estimate for model complexity (i.e., fp = K).Using this very conservative estimate of model complexity in the AIC and BIC measure (see ( 14) and ( 16)) for ADPROCLUS resulted in an overestimation of the number of clusters for all 840 data sets.In particular, for each data set, the most complex model (i.e., K = 8) was erroneously identified as the correct model by both AIC and BIC.It can be concluded that a sensible estimate for model complexity in an ADPROCLUS context should be substantially larger than the number of clusters K and substantially smaller than the estimate given in (8).
A post hoc analysis was conducted to investigate how the performance of AIC changes as a function of the definition of model complexity.Note that the AIC test statistic ( 14) can be rewritten as AIC = 2(-l) + 2(w × fp), (32) where w = 1 for AIC and w = n n−fp−1 (AICc), w = log(n) 2 (BIC) and w = log(log(n)) (HQM) for the other information-theoretic measures.To investigate the effect of using other estimates for model complexity (i.e., w × fp) on the model selection performance of information-theoretic strategies in ADPROCLUS, the parameter w was manipulated to increase in steps of 0.025 over the interval between 0.5 and 1 (i.e., 0.500, 0.525, 0.550, . . ., 1) and this post hoc AIC-based strategy with all different w-values was applied to all 840 simulated data sets.In Fig. 2, mean accuracy (upper panel) and precision (lower panel) is displayed for each value of w, with the dashed black line indicating the performance of CHull NLL, which was the best performing model selection strategy in the simulation study.Both accuracy and precision varied dramatically for different values of w.Accuracy ( = 77%) and precision ( = 0.53) were both maximal for w = 0.625.Note that when using the optimal w, the post hoc AIC-based strategy clearly outperforms both CHull versions in terms of accuracy and precision.Even for w-values close to the optimal w, the post hoc strategy outperforms CHull NLL.When investigating the proportion of correctly retrieved models by the post hoc strategy for w = 0.625 as a function of noise level, we see that for data sets with noise level = 0.1 ( = 100%) and = 0.4 ( = 90.0%), the post hoc AIC-based strategy could reliably estimate the number of clusters, whereas performance decreased to 31.4% for data sets with a noise level of = 0.7.
It is not clear to which extent the optimal value of w and associated optimal definition of model complexity for ADPROCLUS as fp = 0.625(IK + JK + 1) is applicable to data sets with other data characteristics than the data sets generated in the simulation study.In particular, the optimal w-value may strongly depend on the number of objects I and/or on the number of variables J that was used for the simulated data.Further research on this issue may show how generalizable the obtained results are and may result in a modification of the complexity value fp that appropriately captures the true complexity of the ADPROCLUS model.However, the results demonstrate the usefulness of information criteria to tackle the non-trivial model selection problem in ADPROCLUS, given an appropriate definition of model complexity is used.

Omega: Alternative Instability Measure for the Cross-validation Based Strategies
Both cross-validation based strategies, CVs and CVc, performed very poor in the simulation study.One reason for this may be the particular choice of instability measure s in (30).Instead of looking at the misfit for X 3 when using P 1 and P 2 as profile matrices, we can alternatively look at the dissimilarity between the obtained clusterings for X 3 based on P 1 and P 2 .To this end, the Omega index (Collins & Dent, 1988), which generalizes Hubert and Arabie's Adjusted Rand Index (ARI; Hubert & Arabie, 1985) to the case of overlapping clustering, may be used.In particular, the Omega index between A (P 1 ) 3 and A (P 2 ) 3 may be adopted as a stability measure.Applied to the 840 data sets at hand, the accuracy of simplified CV using the Omega index improved to 29.4% (compared to 19.4% using the original test statistic (30)) and accuracy of complex CV improved to 34.8% (compared to 30.4% using (30)).Although the Omega index seems to somewhat improve the performance of the proposed cross-validation strategies, the performance of the CV strategies is not at the level of computationally less demanding methods like CHull and AIC.

Generalizability of the Results to Data Sets with a Different Number of Variables
In the simulation study, the number of variables in the data set was kept constant at J = 15.In order to show that the performance results of the different model selection strategies that were reported above can also be generalized to data sets with a different number of variables, we ran a second simulation study in which we fixed the number of variables at J = 8.As the ANOVA results (see Section 4.3.2) pointed out that the number of objects had a negligible effect on the performance of the different model selection strategies, in order to reduce the computation time, we fixed the number of objects to 200.As such, the size, I × J, of data matrix X was kept constant at 200×8.All other factors and manipulated factor levels stayed the same as described in Section 4.1, resulting in the generation of 420 (60 NOL and 360 OL) data sets.For each generated data set, we fitted ADPROCLUS models with K = 1 up to K = 6 clusters.Again, in order to reduce computation time and because increasing the number of clusters from K = 6 to K = 8 for the simulation study reported above did not change the pattern of the performance results, we decided to run ADPROCLUS only with K = 1 up to K = 6 clusters.For each analysis, we applied a multi-start procedure with 50 starts (also for the analyses for both cross-validation based procedures with again using fivefold cross-validation for the complex cross-validation procedure).

Discussion
The problem of model selection in ADPROCLUS, specifically, and in overlapping clustering, in general, has not yet been tackled in a satisfactory way nor has it been studied extensively.Therefore, in this paper, being one of the first to do so, the capacity of different model selection strategies to determine the correct number of overlapping clusters in ADPROCLUS is investigated in a systematic way by means of a large-scale simulation study.The results demonstrated that for the more easy data sets (i.e., low amounts of noise and a low number of underlying clusters), model selection strategies based on CHull and AIC reliably selected the model with the correct number of clusters.The performance of CHull proved to be more stable than AIC in more difficult, but also more realistic, data conditions (i.e., more noise in the data, a larger number of underlying clusters that show more overlap, and the non-occurrence of certain membership patterns that indicate overlap).The amount of noise in the data had the largest effect on the overall performance and data sets with a large noise level posed a major difficulty for all model selection strategies.In the more realistic data conditions, AIC, as well as the three other information criteria (i.e., AICc, BIC and HQM), tended to underestimate the number of clusters, suggesting that for more complex models (i.e., with more clusters) these criteria penalize the model fit too strongly.In general, when not identifying the correct number of clusters, CHull tends to under/over-estimates the number of clusters to a lesser degree than AIC and the other model selection strategies.The results for the Caliński-Harabasz and Dunn Index, the LBT measure (Steinley & Brusco, 2011), the KL and ASW Index and the proposed cross-validation (CV) based strategies were disappointing, implying that these methods should not be used for model selection in ADPROCLUS.It can be concluded that CHull (Ceulemans et al., 2011;Wilderjans et al., 2012), especially the version in which the likelihood ( 11) is used as fit measure, is the best strategy for determining the optimal ADPROCLUS model out of a range of models with an increasing number of clusters.
In the remainder of this section, several issues related to determining the number of clusters in ADPROCLUS are discussed that deserve further investigation.First, the presence of-large amounts of-noise in the data that may contain some cluster structure is discussed.Second, some thoughts about the local minima problem in ADPROCLUS are presented.Third, some considerations on the scalability of ADPROCLUS algorithms is presented and, finally, determining whether data need to be modelled with an overlapping versus non-overlapping cluster structure is discussed.
Noise in the Data That Has a Cluster Structure For all design cells in the simulation study with large amounts of noise (i.e., 70% of the data is noise), it is observed that for none of the model selection strategies the performance is at a satisfactory level.In particular, the best-although not very encouraging-results are obtained for both CHull methods (i.e., an accuracy around 20%) and both cross-validation methods (i.e., between 10% and 15%), whereas for the other strategies the optimal number of cluster K is (almost) never identified (i.e., equals or is close to zero).A possible reason for this poor performance of the model selection strategies when the data are very noisy is that the noise may contain some-overlapping-cluster structure which misguides the algorithm and, as a consequence, obstructs the algorithm from uncovering the true cluster structure and true number of clusters underlying the data.This may especially be the case when the amount of noise in the data is large and when clusters overlap to a large extent as in these situations the true clusters are hard to disclose and easily can be obscured by cluster structure present in the noise.An indication in this direction are the results of a study of Brusco (2004) in which, in the context of clustering binary data, the effect of structured noise on the recovery of the true cluster structure is investigated.In particular, this study shows that, when generating data with two different true underlying cluster structures, the clustering algorithm tends to only uncover one of both true structures.Moreover, the recovery of this true cluster structure is hampered by the other true cluster structure, which in this case could be considered as structured noise.Indeed, the structure in the noise is picked up by the algorithm, herewith obscuring the simulated true cluster structure, which can be expected to have a serious deteriorating effect on the performance of any model selection strategy.This may especially be the case when the true structure is rather weak compared to the-strong-structure present in the noise (Ceulemans & Kiers, 2009).
Local Minima in ADPROCLUS All model selection strategies proposed in this paper rely in some way on the fit of the ADPROCLUS model (i.e., sum of squared residuals or negative log-likelihood) to the data.As such, when comparing model selection strategies, it is of utmost importance that for each ADPROCLUS model included in the model comparison (e.g., a set of models with increasing numbers of clusters), the global optimal solution is found and its (optimal) fit value is used in the calculations for the model selection strategies; otherwise, the local minima problem in ADPROCLUS may negatively impact-as it may get intertwined with-the ADPROCLUS model selection problem.The ADPROCLUS loss function, however, is known to suffer from a severe local minima problem, implying that the algorithm sometimes retains a local optimal solution that fits the data worse than the global optimal solution (Depril et al., 2008).Moreover, when the number of clusters of the ADPROCLUS model and the amount of noise in the data increases, it may be expected that the local minima problem enlarges as the former implies a larger solution space and the latter an increase in the number of local optimal solutions in which the algorithm can get stuck.To mitigate the local minima problem, often an extensive multi-start procedure is used in which the algorithm is run multiple times and the best solution encountered across all runs is retained.However, the number of multi-starts used is often kept the same for each K, implying that for smaller K, where the problem of local minima is expected to be less of an issue, the algorithm may the local minima problem to a larger extent than for a larger K.It is not clear yet how this problem of local minima and the dependence of its severity on K affects the proposed model selection strategies.It is also not yet thoroughly investigated how the number of starts in a multi-start procedure should be tuned to account for the differences in severity of the local minima problem for increasing values of K. To deal with the larger local minima problem caused by larger amounts of noise in the data, an interesting proposal has been launched by Steinley (2006).The author observes that, when the data contain noise and clusters are hard to disentangle from each other (because, for example, clusters overlap to a larger extent), the global optimum solution is not necessarily the solution that captures the true underlying cluster structure the best.As a consequence, in such a situation, one should not only look at the global optimum solution but also take local optimal solutions into account.As a way out, Steinley (2006) proposes to profile the local optima in order to aid the model selection process.Further research regarding these issues is highly recommended and may also have important ramifications for other problems suffering from local minima, like K-means, model-based clustering and biclustering.

Limited Size of the Data Sets Used and Scalability of the ADPROCLUS Algorithm
The data sets used in this simulation study were limited to only include 200 or 400 objects and a small number of variables (i.e., 8 or 15).Data sets in scientific practice, however, easily have a (much) larger number of objects and variables.It is not clear whether the (slight) trend of model selection strategies performing better when the data size increases, as observed in the simulation, also holds when the number of objects and variables becomes (really) large.Fitting ADPROCLUS models to large data sets, however, may be problematic for two reasons.First, the increasing severity of the local minima problem when the data size and thus the number of parameters to be estimated (i.e., solution space) increases (see earlier comment).And secondly, the enormous increase in computational burden for the ADPROCLUS algorithm, which may become prohibitive for large numbers of objects I and/or variables J, even when parallel computing infrastructure can be used.Note that the sometimes large computation time for fitting the ADPROCLUS model-even when applied to data of a moderate size-and this especially for larger number of clusters, limits to some extent the real world application of the proposed model selection strategies as these strategies require the fitting of ADPROCLUS models with increasing number of clusters.Regarding the latter, as a way out, it may be worthwhile to develop a faster and more scalable algorithm for ADPRO-CLUS.A good starting point here may be the work of France and Abbassi (2011) in which additive clustering algorithms are boosted by combining them with clusterwise optimization and multi-label learning strategies.The rationale of this approach consists of training an overlapping clustering model, like ADPROCLUS, on a manageable subset of the data and using the resulting cluster solution subsequently in a multi-label learning method to also find the cluster memberships of the objects that were omitted in first instance.Increasing the algorithmic speed of the ADPROCLUS algorithm will improve the usability of ADPROCLUS and the proposed model selection strategies in empirical practice.

Selecting between Overlapping and Non-overlapping Underlying Cluster Structures
Before identifying the optimal number of clusters in ADPROCLUS, it first should be determined whether an overlapping clustering structure is underlying the data at hand, warranting the use of ADPROCLUS as such.When clusters do not overlap, and thus the true cluster structure is a partition, K-means or another partitioning method (e.g., model-based clustering) should be preferred over ADPROCLUS.Just performing ADPROCLUS on the data and checking whether or not objects are assigned to multiple clusters is not a good strategy to determine the nature of the cluster structure underlying a data set at hand as the data contain noise which may blur the underlying cluster structure.In particular, in our simulation study, an ADPROCLUS model often yielded a solution with overlapping clusters, even when the true underlying cluster structure was a partition.In particular, for data sets without cluster overlap, ADPROCLUS assigned, on average, 59.5% of the objects to multiple clusters; in general, this tendency became larger when the data contained more noise (i.e., 31.3%, 79.4% and 67.8% for = 0.1, 0.4 and 0.7, respectively).
An alternative, and probably better, option to determine the nature of the underlying cluster structure consists of constructing a test to determine whether clusters overlap.One way to obtain this is to derive a similar minimal stochastic extension for K-means as was proposed for ADPROCLUS.In particular, such an extension of K-means would lead to the same formulas ( 9)-( 13) as presented in Section 3.1.1,with only m ij now being the predicted data value under K-means instead of under ADPROCLUS.As the stochastic extension of the K-means model is nested within the similar extension of the ADPROCLUS model (i.e., the restriction in K-means that rows of A should sum to one is omitted in ADPROCLUS), a likelihood ratio (LR) test may be constructed that tests the hypothesis whether the more simple K-means model fits the data equally well as the more complex ADPROCLUS model.When this hypothesis is rejected, it can be concluded that the underlying cluster structure indeed shows overlap.In such a LR test, the change in likelihood between both models, when applied to the same data set with the same number of clusters K, is weighted against the difference in model complexity.In other words, it is determined whether the extra freedom-and complexity-in the ADPROCLUS model (i.e., allowing clusters to overlap) is compensated by a substantial increase in model fit/likelihood.
Applying this LR test (using as model complexity for ADPROCLUS (8) and for K-means fp = I + JK + 1) to the data sets from our simulation study resulted in 49.0% of the OL data sets being correctly identified as data sets with an underlying overlapping cluster structure, whereas only 14.2% of NOL data sets were erroneously identified as data sets containing overlapping clusters.It should be noted that for NOL data sets, the true underlying model did not fully comply with the K-means model (i.e., the underlying cluster structure being a partition) as always 5% of the objects belonged to no cluster at all (i.e., have an all-zero membership pattern).For data sets with a larger noise level , the LR test, in general, less often indicated an overlapping cluster structure, resulting in the LR test being less good in detecting an overlapping cluster structure when the noise in the data increased.In particular, when = 0.1, for OL data sets, 97.1% of them were correctly identified as containing an overlapping cluster structure.For = 0.4 and = 0.7, this percentage shrank to 50% and 0%, respectively.Moreover, all NOL data sets that were erroneously marked as showing cluster overlap belonged to the condition with almost no noise ( = 0.1).Some caution regarding the use of such a LR test is needed as its performance critically depends on an adequate definition of model complexity for the models compared (i.e., the distribution of the test statistic depends on the difference in model complexity between the models compared).However, both for ADPROCLUS and K-means, the definition of model complexity is still an issue that is not yet resolved in a satisfactory way.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made.The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material.If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder.To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Fig. 1
Fig. 1 Two-way interactions (upper panels) and three-way interaction (lower panels) for the mixed-ANOVA on OL data sets: (upper left) Precision as a function of the true number of clusters K and the model selection strategy used; (upper right) Precision as a function of the noise level and the model selection strategy used; (lower panels) Precision as a function of the noise level , the model selection strategy used and the true number of clusters K for K = 3 (left panel) and K = 5 (right panel) larger noise level of = 0.4 (N = 74) and = 0.7 (N = 66) and 400 observations (N = 83).

Fig. 2
Fig. 2 Accuracy (upper panel) and precision (lower panel), averaged across all data sets, of selecting the optimal number of clusters in ADPROCLUS for different values of the weighting parameter w in a post hoc analysis of AIC performance

Table 1
The number of distinct non-occurring overlapping membership patterns as a function of the amount of underlying clusters K and the degree of non-occurring overlapping membership patterns

Table 2
Mean accuracy (expressed as a percentage) and precision of thirteen model selection strategies for selecting the optimal number of overlapping clusters in ADPROCLUS for OL (n = 720) and, between brackets, NOL data sets (n = 120), computed globally and as a function of the noise level overestimation

Table 3
Mean accuracy (displayed as a percentage and rounded to one decimal place) computed for each model selection strategy separately and across all model selection strategies, as a function of the various levels of the manipulated data characteristics

Table 4
Cross-tabulation of the number of correctly retrieved models for CHull NLL and AIC, computed overall and per noise level

Table 5
Mean computation time (in minutes, rounded to two decimal places) and, between brackets, standard deviations, for a single multi-start of the ADPROCLUS algorithm, averaged across all simulated data sets, as a function of the number of objects I and the number of clusters K