References
BRADLEY, P.E. (2017), “Finding Ultrametricity in Data Using Topology”, Journal of Classification, 34, 76–84.
DANG, U.J., PUNZO, A., MCNICHOLAS, P.D., INGRASSIA, S., and BROWNE, R.P. (2017), “Multivariate Response and Parsimony for Gaussian Cluster Weighted-Models”, Journal of Classification, 34, 4–34.
DE SOETE, G. (1984), “Ultrametric tree representations of incomplete dissimilarity data”, Journal of Classification, 1, 235–242.
DE SOETE, G. (1988), “OVWTRE: A Program for Optimal Variable Weighting for Ultrametric and Additive Tree Fitting”, Journal of Classification, 5, 101–104.
DE SOETE, G., CARROLL, J.D., and DESARBO, W.S. (1987), “Least Squares Algorithms for Constructing Constrained Ultrametric and Additive Tree Representations of Symmetric Proximity Data”, Journal of Classification, 4, 155–173.
FRIEDMAN, J.H., and MEULMAN, J.J. (2004), “Clustering Objects on Subsets of Attributes”, Journal of the Royal Statistical Society, B, 66, 815–849.
HENNIG, C., and LIAO, T.F. (2013), “How to Find an Appropriate Clustering for Mixed-Type Variables with Application to Socio-Economic Stratification”, Journal of the Royal Statistical Society, C, 62, 309–369.
JOSSE, J., CHAVENT, M., LIQUET, B., and HUSSON, F. (2012), “Handling Missing Values With Regularized Iterative Multiple Correspondence Analysis”, Journal of Classification, 29, 91–116.
LAPOINTE, F.-J., and LEGENDRE, P. (1991), “The Generation of Random Ultrametric Matrices Representing Dendrograms”, Journal of Classification, 8, 177–210.
MAKARENKOV, V., and LEGENDRE, P. (2001), “Optimal Variable Weighting for Ultrametric and Additive Trees and K-means Partitioning: Methods and Software”, Journal of Classification, 18, 245–271.
MILLIGAN, G.W. (1980), “An Examination of the Effect of Six Types of Error Perturbation on Fifteen Clustering Algorithms”, Psychometrika, 45, 325–342.
MURTAGH, F. (2004). “On Ultrametricity, Data Coding, and Computation”, Journal of Classification, 21, 167–184.
PUNZO, A., and MCNICHOLAS, P.D. (2017), “Robust Clustering in Regression Analysis via the Contaminated Gaussian Cluster-Weighted Model”, Journal of Classification, 34, 249–293.
SICILIANO, R., D’AMBROSIO, A., ARIA, M., and AMODIO, S. (2016), “Analysis of Web Visit Histories, Part I: Distance-Based Visualization of Sequence Rules”, Journal of Classification, 33, 298–324.
SMITH, T.J. (2001), “Constructing Ultrametric and Additive Trees Based on the L1-Norm”, Journal of Classification, 18, 185–207.
SRIRAM, N. (1990), “Clique Optimization: A Method to Construct Parsimonious Ultrametric Trees From Similarity Data”, Journal of Classification, 7, 33–52.
SRIRAM, M., and LEWIS, S. (1993), “Constructing Optimal Ultrametrics”, Journal of Classification, 10, 241–268.
STEINLEY, D. (2003), “Local Optima in K-means Clustering: What You Don't Know May Hurt You”, Psychological Methods, 9, 386–396.
STEINLEY, D. (2006), “Profiling Local Optima in K-means Clustering: Developing a Diagnostic Technique”, Psychological Methods, 11, 178–192.
STEINLEY, D., and BRUSCO, M.J. (2008), “Selection of Variables in Cluster Analysis: An Empirical Comparison of Eight Procedures”, Psychometrika, 73, 46–62.
STEINLEY, D., and BRUSCO, M.J. (2011), “Evaluating the Performance of Model-Based Clustering: Recommendations and Cautions”, Psychological Methods, 16, 63–79.
SULIS, I., and PORCU, M. (2017). “Handling Missing Data in Item Response Theory. Assessing the Accuracy of a Multiple Imputation Procedure Based on Latent Class Analysis”, Journal of Classification, 34, 327–359.
VRAC, M., BILLARD, L., DIDAY, E., and CHEDIN, A. (2012). “Copula Analysis of Mixture Models”, Computational Statistics, 27, 427–457.
ZUBAREV, A. P. (2014). “On Stochastic Generation of Ultrametrics in High-Dimensional Euclidean Spaces”, p-Adic Numbers, Ultrametric Analysis, and Applications, 6, 155–165.
Rights and permissions
About this article
Cite this article
Editorial. J Classif 34, 361–365 (2017). https://doi.org/10.1007/s00357-017-9245-7
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00357-017-9245-7