Abstract
We consider the robust version of the classic k-center clustering problem, where we wish to remove up to z points (outliers), so as to be able to cluster the remaining points in k clusters with minimum maximum radius. We study such a problem under the fully dynamic adversarial model, where points can be inserted or deleted arbitrarily. In this setting, the main goal is to design algorithms that maintain a high quality solution at any point in time, while requiring a “small” amortized cost, i.e. a “small” number of operations per insertion or deletion, on average. In our work, we provide the first constant bi-criteria approximation algorithm for such a problem with its amortized cost being independent of both z and the size of the current input.
T-H. Hubert Chan was partially supported by the Hong Kong RGC under the grants 17201220 and 17202121. The work of Mauro Sozio has been carried out in the frame of a cooperation between Huawei Technologies France SASU and Telecom Paris (Grant no. YBN2018125164).
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
- 1.
We note here that the known coresets for k-center with outliers do not imply an algorithm with bounded amortized update time for such a problem, in general metric.
- 2.
However, for practical implementation, we could possibly re-use the underlying intermediate data structures.
References
Bateni, M., Esfandiari, H., Jayaram, R., Mirrokni, V.S.: Optimal fully dynamic k-centers clustering. CoRR abs/2112.07050 (2021)
de Berg, M., Monemizadeh, M., Zhong, Y.: k-center clustering with outliers in the sliding-window model. In: ESA, vol. 204, pp. 1–13. Schloss Dagstuhl (2021)
Bhaskara, A., Vadgama, S., Xu, H.: Greedy sampling for approximate clustering in the presence of outliers. In: NeurIPS, pp. 11146–11155 (2019)
Ceccarello, M., Pietracaprina, A., Pucci, G.: Solving k-center clustering (with outliers) in MapReduce and streaming, almost as accurately as sequentially. Proc. VLDB Endow. 12(7), 766–778 (2019)
Chan, T.H., Guerqin, A., Sozio, M.: Fully dynamic k-center clustering. In: WWW, pp. 579–587 (2018)
Chan, T.H., Guerquin, A., Hu, S., Sozio, M.: Fully dynamic k-center clustering with improved memory efficiency. IEEE Trans. Knowl. Data Eng. 34(7), 3255–3266 (2022)
Charikar, M., Khuller, S., Mount, D.M., Narasimhan, G.: Algorithms for facility location problems with outliers. In: SODA, pp. 642–651 (2001)
Charikar, M., O’Callaghan, L., Panigrahy, R.: Better streaming algorithms for clustering problems. In: STOC, pp. 30–39 (2003)
Cohen-Addad, V., Hjuler, N.O.D., Parotsidis, N., Saulpic, D., Schwiegelshohn, C.: Fully dynamic consistent facility location. In: NeurIPS, pp. 3250–3260 (2019)
Cohen-Addad, V., Schwiegelshohn, C., Sohler, C.: Diameter and k-center in sliding windows. In: ICALP, pp. 1–12 (2016)
Ding, H., Yu, H., Wang, Z.: Greedy strategy works for k-center clustering with outliers and coreset construction. In: ESA (2019)
Goranci, G., Henzinger, M., Leniowski, D., Svozil, A.: Fully dynamic k-center clustering in doubling metrics. arXiv preprint arXiv:1908.03948 (2019)
Harris, D.G., Pensyl, T.W., Srinivasan, A., Trinh, K.: A lottery model for center-type problems with outliers. ACM Trans. Algorithms 15(3), 36:1–36:25 (2019)
Henzinger, M., Kale, S.: Fully-dynamic coresets. In: 2020 ESA LIPIcs, vol. 173, pp. 57:1–57:21 (2020)
Hou, B.J., Zhang, L., Zhou, Z.H.: Learning with feature evolvable streams. In: Advances in Neural Information Processing Systems, pp. 1417–1427 (2017)
Huang, L., Jiang, S., Li, J., Wu, X.: Epsilon-coresets for clustering (with outliers) in doubling metrics. In: FOCS, pp. 814–825 (2018)
Malkomes, G., Kusner, M.J., Chen, W., Weinberger, K.Q., Moseley, B.: Fast distributed k-center clustering with outliers on massive data. In: Advances in Neural Information Processing Systems, pp. 1063–1071 (2015)
Matthew McCutchen, R., Khuller, S.: Streaming algorithms for k-center clustering with outliers and with anonymity. In: Goel, A., Jansen, K., Rolim, J.D.P., Rubinfeld, R. (eds.) APPROX/RANDOM -2008. LNCS, vol. 5171, pp. 165–178. Springer, Heidelberg (2008). https://doi.org/10.1007/978-3-540-85363-3_14
Putina, A., Sozio, M., Rossi, D., Navarro, J.M.: Random histogram forest for unsupervised anomaly detection. In: ICDM, pp. 1226–1231. IEEE (2020)
Schmidt, M., Sohler, C.: Fully dynamic hierarchical diameter k-clustering and k-center. arXiv preprint arXiv:1908.02645 (2019)
Wang, H., Fan, W., Yu, P.S., Han, J.: Mining concept-drifting data streams using ensemble classifiers. In: SIGKDD, pp. 226–235 (2003)
Zhang, P., Li, J., Wang, P., Gao, B.J., Zhu, X., Guo, L.: Enabling fast prediction for ensemble models on data streams. In: SIGKDD, pp. 177–185 (2011)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Chan, TH.H., Lattanzi, S., Sozio, M., Wang, B. (2022). Fully Dynamic k-Center Clustering with Outliers. In: Zhang, Y., Miao, D., Möhring, R. (eds) Computing and Combinatorics. COCOON 2022. Lecture Notes in Computer Science, vol 13595. Springer, Cham. https://doi.org/10.1007/978-3-031-22105-7_14
Download citation
DOI: https://doi.org/10.1007/978-3-031-22105-7_14
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-22104-0
Online ISBN: 978-3-031-22105-7
eBook Packages: Computer ScienceComputer Science (R0)