1 Introduction

More is more is one of the central tenets associated with Big Data. More data means more information from which we hope to gain a better understanding of an underlying truth. The artificial intelligence and machine learning communities have focused on modeling and learning increasingly elaborate statistical models for obtaining new knowledge from the data. However, in the era of Big Data, scalability has become essential for any learning algorithm. In light of this, research has begun to focus on aggregation tools that provide trade-offs between information and space requirement. Note, that these are also extremely valuable for practical applications. The design of a data aggregation can be decoupled from the design of the actual learning algorithm. This is commonly referred to as the sketch-and-solve paradigm. The main idea is to reduce the size of the data first, to have a sketch whose size has only little or even no dependency on the initial sizeFootnote 1: “Big Data turns into tiny data!” Then a learning algorithm is applied to the little sketch only. Its complexity measures like running time, memory, and communication are thus significantly less dependent on the size of the initial data. In most cases, the learning algorithm remains unchanged. For a given data aggregation, the main challenge is to bound the trade-off between its size and its loss of information.

What constitutes relevant information is, of course, application dependent. Coresets are an algorithmic framework towards quantifying such trade-offs for various objective functions of interest. Generally speaking, we are given a data set A, a set of candidate solutions C, and some optimization function f. Our aim is to find a significantly smaller data set S such that for all candidate solutions \(c\in C\), f(Sc) approximates f(Ac).

Coresets turn out be extremely useful in the context of both, approximation and streaming algorithms. Given that a problem admits a coreset whose size is independent of the input, a polynomial time approximation scheme (PTAS) can often be derived via brute force enumeration. While such a PTAS will not necessarily be of practical relevance, this example illustrates how coresets can make computationally expensive algorithms more viable. In a streaming setting, we are given the further restriction that we do not have a random access to the data, but must process the data points one by one. It turns out that if a problem admits a coreset, there exists a black box reduction to streaming algorithms commonly referred to as the merge-and-reduce principle, cf. [49]. It is thus not necessary to develop a special streaming algorithm for that problem.

Coresets have been studied for several base problems that arise in many applications of artificial intelligence and machine learning. Some examples include clustering [4, 17, 42, 50, 75], classification [56, 59, 89], regression [34, 40, 41, 54], and the smallest enclosing ball problem [6, 15, 16, 48]; we refer to [84] for a recent extensive literature overview. If one of these base problems arises in some application, a coreset construction can often be used in a black box manner to improve the performance of the learning algorithm. This survey, in contrast, is intended to give the reader an overview over existing methods to design and analyze new coreset constructions for individual learning tasks of interest. We believe such an methodological introduction to coresets, and their benefits in the design of fast approximation- and streaming algorithms can be very valuable to the artificial intelligence community.

This survey is now organized into the following parts. In Sect. 2, we describe some basic facts and notations and give a formal definition of the strong coreset guarantee. In Sect. 3, we provide an overview on the four most important techniques used in their construction:

  • Geometric decompositions.

  • Gradient descent.

  • Random sampling.

  • Sketching and projections.

For each technique we will consider an example problem and a coreset construction. We further give an overview on results in this line of research. In Sect. 4, we show how coresets may be used in the context of streaming algorithms. In Sect. 5, we discuss lower bounds on coreset sizes, and provide example problems for which coresets do not exist. We conclude with some open problems.

2 Preliminaries

We assume \(0<\varepsilon ,\delta \le \frac{1}{2}\) for all approximation resp. failure parameters in this paper. We first give the definition of the strong coreset guarantee.

Definition 1

(Strong Coresets) Let A be a subset of points of some universe U and let C be a set of candidate solutions. Let \(f:U\times C\rightarrow \mathbb {R}^{\ge 0}\) be a non-negative measurable function. Then a set \(S\subset U\) is an \(\varepsilon\)-coreset, if for all candidate solutions \(c\in C\) we have

$$\begin{aligned} |f(A,c)-f(S,c)|\le \varepsilon \cdot f(A,c). \end{aligned}$$

This first paper to formalize the coreset framework is that of Agarwal et al. [4]. In many cases, the construction of a coreset is a randomized process. Here, we account for the probability of failure via the parameter \(\delta\), which in turn influences the size of S. We are interested in coresets S whose size is significantly smaller than |A|. Specifically, we aim to find an S such that \(|S|\in \text {polylog}(|A|)\) Footnote 2 or even such that |S| has no dependency on |A|. The size of |S| will almost always depend on \(\varepsilon\), and often on properties of the universe U (e.g. a dependency on d if the points lie d-dimensional Euclidean space) or on properties of f (e.g. a dependency on k, if when we are dealing with the k-means objective function). If there exists no constant \(c>0\) such that \(|S|\in o(|A|^{1-c})\), we say that no coreset exists.

There exists another notion of coreset whose use is not as clear cut as the strong coreset. Roughly speaking, we only require that a solution computed on S is a \((1+\varepsilon )\) approximation to the optimum solution of A. In this case, we say that S is a weak coreset, though we emphasize that there exist no unifying definition on the guarantees of a weak coreset and there are many variations encountered in literature.

We note that while S often is a subset of A, it does not necessarily have to be and we will see examples for this. In most cases the universe U is a Euclidean space, where the \(\ell _2\) norm for \(\mathbb {R}^d\) is defined as \(\Vert x\Vert :=\sqrt{\sum _{i=1}^d |x_i|^2}\).

We will frequently use a standard result on ball-covers for Euclidean spaces. An \(\varepsilon\)-ball-cover of the unit sphere is a set of points B, such that for any point p in the unit sphere the distance to any point of B is at most \(\varepsilon\).

Lemma 1

([85]) Let be the unit sphere in d-dimensional Euclidean space. Then for every \(0<\varepsilon <1\) there exists an \(\varepsilon\)-ball-cover  B of size \(\left( 1+\frac{2}{\varepsilon }\right) ^d\)i.e. for every point \(p\in U\)

$$\begin{aligned} \underset{b\in B}{~\min ~} \Vert p-b\Vert \le \varepsilon . \end{aligned}$$

We note that while this bound is sharp, there exists no known efficient method of constructing such a ball-cover. In most cases, we require only the existence of a small ball-cover for the analysis. If an algorithm uses an actual ball-cover, there are multiple options to construct one using \(\varepsilon ^{-O(d)}\) points, see Chazelle for an extensive overview [26].

3 Construction Techniques

We have identified four main approaches used to obtain coresets. This is not meant to be an absolute classification of algorithms, but rather an overview of what techniques are currently available and an illustration of how we may apply them.

3.1 Geometric Decompositions

Assume that we are given a point set A in Euclidean space. A simple way to reduce the size of A is to compute a discretization S of the space, snap each point of A to its nearest neighbor in S, and use the (weighted) set S to approximate the target function. Many of these decompositions are based on packing arguments and range spaces. We will illustrate one approach via the k-means problem, based on the papers by Har-Peled and Mazumdar [58] and Fichtenberger et al. [51].

Definition 2

(k-Means Clustering) Let A be set of n points in Euclidean space and let C be a set of points with \(|C|=k\). The k-means objective asks for a set C minimizing

$$\begin{aligned} f(A,C):= \sum _{x\in A}\underset{c\in C}{~\min ~}\Vert x-c\Vert ^2. \end{aligned}$$

We first take a closer look at the objective function for Euclidean k-means. Let \(\text {OPT}\) be the value of the optimum solution. We first compute a 10-approximationFootnote 3 \(C'\) to the k-means problem, which can be done in polynomial time [68].

Consider now a sequence of balls with exponentially increasing radius centered around each point of \(C'\), starting at radius \(\frac{1}{n}\cdot \text {OPT}\) and ending at \(10\cdot \text {OPT}\). Our discretization will consist of a suitably scaled \(\varepsilon\)-ball-cover of each ball. To prove correctness, we require the following generalization of the triangle inequality.

Lemma 2

(Generalized Triangle Inequality [71]) Let abc be points in Euclidean space. Then for any  \(\varepsilon \in (0,1)\) we have

$$\begin{aligned} \left| \Vert a-c\Vert ^2-\Vert b-c\Vert ^2 \right| \le \frac{12}{\varepsilon }\cdot \Vert a-b\Vert ^2 + 2\varepsilon \cdot \Vert a-c\Vert ^2. \end{aligned}$$

We now show that the cost of using the discretization is bounded.

Lemma 3

Let A be a set of points, let \(B^i\) be the ball with radius  \(r_i:=\frac{2^i}{n}\cdot \sum _{x\in A}\Vert x\Vert ^2\) centered at the origin and let \(S^i\) be the \(\frac{\varepsilon }{3}\)-ball-cover of  \(B^i\)Denote by \(S=\bigcup _{i=0}^{\log 10n}S^i\).  Then \(\sum _{x\in A}\underset{s\in S}{~\min ~}\Vert x-s\Vert ^2\le \varepsilon ^2\cdot \sum _{x\in A}\Vert x\Vert ^2.\)

Proof

Denote by \(A_{close}\) the points with squared Euclidean norm at most \(\frac{1}{n}\cdot \sum _{x\in A}\Vert x\Vert ^2\) and by \(A_{far}\) the remaining points. Since \(|A_{close}|\le n\), we have \(\sum _{x\in A_{close}}\underset{s\in S^0}{~\min ~}\Vert x-s\Vert ^2\le |A_{close}|\cdot \frac{1}{n}\cdot \sum _{x\in A}\Vert x\Vert ^2\cdot \frac{\varepsilon ^2}{9}\le \frac{\varepsilon ^2}{9}\cdot \sum _{x\in A}\Vert x\Vert ^2\). For the points in \(A_{far}\), consider any point x in \(B^i\setminus B^{i-1}\) for \(i\in \{1,\ldots ,\log 10n\}\). We have \(\underset{s\in S^i}{~\min ~}\Vert x-s\Vert ^2\le \frac{\varepsilon ^2}{9} r_i^2 \le \frac{4\varepsilon ^2}{9} r_{i-1}^2 \le \frac{4\varepsilon ^2}{9}\cdot \Vert x\Vert ^2.\) Summing up over all points, we have \(\sum _{x\in A}\underset{s\in S}{~\min ~}\Vert x-s\Vert ^2\le \frac{\varepsilon ^2}{9}\cdot \sum _{x\in A}\Vert x\Vert ^2 + \frac{4\varepsilon ^2}{9}\cdot \sum _{x\in A_{far}}\Vert x\Vert ^2 < \varepsilon ^2\cdot \sum _{x\in A}\Vert x\Vert ^2\). \(\square\)

We can reapply this analysis for each center of the 10-approximation \(C'\). Observe that the cost of points \(A_c\) assigned to a center \(c\in C'\) is \(\sum _{x\in A_c}\Vert x-c\Vert ^2\), which is equal to \(\sum _{x\in A}\Vert x\Vert ^2\) if we treat c as the origin. Combining this lemma with Lemma 2 and rescaling \(\varepsilon\) shows that S is a coreset.

Theorem 1

For any set of n points A Euclidean space, there exists a coreset for k-means consisting of \(O(k\varepsilon ^{-d}\log n)\) points, where d is the (constant) dimension.

Proof

For each of the k centers from the original 10-factor approximation we have a total of \(\log 10n\) balls with varying radii. For each such ball of radius r, we compute an \(\frac{\varepsilon }{16}\cdot r\) ball-cover. For any point \(x\in A\), let B(x) be the nearest point in the union of all ball-covers. By Lemma 3, we have \(\sum _{x\in A}\Vert x-B(x)\Vert ^2 \le \left( \frac{\varepsilon }{16}\right) ^2\cdot 10\cdot \text {OPT}\). Now consider an arbitrary set of centers C. We have

$$\begin{aligned} {} & \left| \sum _{x\in A} \underset{c\in C}{~\min ~}\Vert x-c\Vert ^2 - \sum _{x\in A} \underset{c\in C}{~\min ~}\Vert B(x)-c\Vert ^2 \right| \\&\quad \le \frac{12}{\varepsilon } \sum _{x\in A}\Vert x-B(x)\Vert ^2 + 2\varepsilon \cdot \sum _{x\in A} \underset{c\in C}{~\min ~}\Vert x-c\Vert ^2 \\&\quad \le \frac{12}{\varepsilon } (\varepsilon /16)^2\cdot 10\cdot \text {OPT} + 2\varepsilon \cdot \sum _{x\in A} \underset{c\in C}{~\min ~}\Vert x-c\Vert ^2 \\&\quad < 2\varepsilon \cdot \text {OPT} + 2\varepsilon \cdot \sum _{x\in A} \underset{c\in C}{~\min ~}\Vert x-c\Vert ^2 \\&\quad \le 4\varepsilon \cdot \sum _{x\in A} \underset{c\in C}{~\min ~}\Vert x-c\Vert ^2, \end{aligned}$$

where the first inequality follows from Lemma 2, the second inequality follows follows from Lemma 1, and the last inequality follows from \(\text {OPT}\le \sum _{x\in A} \underset{c\in C}{~\min ~}\Vert x-c\Vert ^2\) for any set of centers C.

Rescaling \(\varepsilon\) by a factor of 1 / 4 gives the proof. The space bound now follows from Lemma 1 and the fact that we compute a \((\varepsilon /64)\)-ball-cover \(k\cdot \log (10n)\) times. \(\square\)

Bibliographic Remarks Algorithms based on a discretization and then snapping points to the closest point of the discretization are particular common for extent approximation problems such as maintaining \(\varepsilon\)-kernels, the directional width, the diameter, and the minimum enclosing ball problem. In fact, the notion behind coresets was introduced in the context of these problems in the seminal paper by Agarwal et al. [4]. Since then, a number of papers have progressively reduced the space bound required to maintain coresets [8, 23, 24, 99] with Arya and Chan giving an algorithm that stores \(O(\varepsilon ^{-(d-1)/2})\) points [14] in d-dimensional Euclidean space. We will briefly outline in Sect. 5 that this space bound is indeed optimal.

The geometric approach was also popular when coresets were introduced to k-median and k-means clustering and generalizations, see for instance [44, 51, 52, 57, 58]. Due to the exponential dependency inherent in all known constructions, the focus later shifted to sampling. Nevertheless, geometric arguments for correctness proofs are necessary for almost all algorithms, and are present in all the other examples presented in this survey. For an extensive and more complete overview of purely geometric algorithms for coreset construction, we recommend the survey by Agarwal et al. [5]

3.2 Gradient Descent

A quite popular and often only implicitly used technique to build coresets is derived from convex optimization. We begin with an example where a simple application of the well-known sub-gradient method yields the desired coreset. Consider the problem of finding the smallest enclosing ball of a set of points, which is equivalent to the 1-center problem in Euclidean space.

Definition 3

Given \(P\subset \mathbb {R}^d\), the smallest enclosing ball problem (SEB) consist in finding a center \(c^* \in \mathbb {R}^d\) that minimizes the cost function

$$\begin{aligned} f(P,c)=\max \limits _{p \in P} \Vert c-p\Vert . \end{aligned}$$

A standard approach for minimizing the convex function f is to start at an arbitrary point \(c_0\) and to perform iterative updates of the form \(c_{i+1}=c_i - s\cdot g(c_i)\) where \(s\in \mathbb {R}^{>0}\) is an appropriately chosen step size and \(g(c_i)\in \partial f(c_i)\) is a sub-gradient of f at the current point \(c_i\). We will refer to this as the sub-gradient method. This method has been formally developed by several mathematicians in the sixties. See [83, 90, 92] for details and historical remarks. Now, consider the following theorem regarding the convergence behavior of the sub-gradient method.

Theorem 2

([83]) Let f be convex and Lipschitz continuous with constant L and  \(\Vert c_0-c^*\Vert \le R\)Let \(f^* =f(c^*)\) be the optimal solution. Let  \(f^*_l=\min _{i=1..l}\) be the best solution among l iterations of the sub-gradient algorithm with step size \(s=\frac{R}{\sqrt{l+1}}\)Then

$$\begin{aligned} f^*_l - f^*\le \frac{LR}{\sqrt{l+1}}. \end{aligned}$$

From Theorem 2 we learn, that if our function has a small Lipschitz constant and our initial starting point is not too far from the optimal solution, then the best center among a small number of iterations of the sub-gradient algorithm will have a radius that is close to the optimal radius of the smallest enclosing ball. To assess the parameters more closely we begin with finding a sub-gradient at each step. It is easy to see that for any \(p_{\max }\in \mathop {\mathrm{argmax}}\nolimits _{p\in P} \Vert c-p\Vert\) attaining the maximum distance \(g(c) = \frac{c-p_{\max }}{\Vert c-p_{\max }\Vert }\) is a sub-gradient of f at c, i.e. \(g(c)\in \partial f(c)\). Also note that \(c-p_{\max }\ne 0\) unless P is a singleton, in which case the problem is trivial since the only input point is the optimal center. Thus, in all interesting cases, g(c) is well-defined. Also note that g(c) is by definition a normalized vector, i.e., \(\Vert g(c)\Vert =1\) which implies that f is L-Lipschitz with \(L=1\), since by definition of a sub-gradient and applying the Cauchy-Schwarz inequality (CSI) we have

$$\begin{aligned} f(c)-f(x)\le g(c)^T(c-x)\overset{\tiny \mathrm{CSI} }{\le } \Vert g(c)\Vert \Vert c-x\Vert \le 1\cdot \Vert c-x\Vert . \end{aligned}$$

This leads to the following corollary.

Corollary 1

We can choose a starting point and an appropriate step size, such that \(f^*_l\)the best solution among \(l=O(\frac{1}{\varepsilon ^2})\)  iterations of the sub-gradient method for the smallest enclosing ball problem satisfies

$$\begin{aligned} f^*_l - f^*\le \varepsilon f^*. \end{aligned}$$

Proof

Note that if we pick an arbitrary input point as our starting point \(c_0=p_0\in P\) and another point \(p_1\in P\) that maximizes the distance to \(p_0\) then we can deduce that \(\Vert p_0-c^*\Vert \le f^*\le \Vert p_0-p_1\Vert :=R\le 2f^*\). So, in the light of Theorem 2, R is a good approximation for determining an appropriate step size, and at the same time acts as a good parameter for the upper bound on the error. Now plugging \(l=\frac{4}{\varepsilon ^2}-1=O(\frac{1}{\varepsilon ^2})\) and \(s=\frac{R}{\sqrt{l+1}}=\frac{\varepsilon }{2} R\) into Theorem 2, we get

$$\begin{aligned} f^*_l - f^* \le \frac{LR}{\sqrt{l+1}} \le \frac{2f^*}{\sqrt{l+1}} \le \varepsilon f^*. \end{aligned}$$

\(\square\)

Now we would like to argue that the set of points S, that the sub-gradient algorithm selects in each iteration as furthest point from the current center, forms a (weak) coreset for the smallest enclosing ball of P. We will need the following well-known fact for our further analysis.

Lemma 4

([17]) Let B(crbe the smallest enclosing ball of a point set  \(P\subset \mathbb {R}^d\)then any closed halfspace that contains c  must also contain at least one point from P that is at distance  r from c.

Using this lemma we can argue that any \((1+\varepsilon )\)-approximate center must be within \(\sqrt{\varepsilon } f^*\) distance to the optimal center.

Lemma 5

Let B(crbe the smallest enclosing ball of a point set  \(P\subset \mathbb {R}^d\)Let \(\tilde{c}\) be a center such that for all \(p\in P\) we have \(\Vert c-p\Vert \le (1+\varepsilon ) r\)Then  \(\Vert \tilde{c}-c\Vert \le \sqrt{3\varepsilon }r\).

Proof

Let \(\tilde{c}-c\) be normal to the hyperplane H passing through c and let \(H^-\) be the halfspace whose border is H containing c but not containing \(\tilde{c}\). By Lemma 4 we know that there must be some \(p\in P\cap H^-\) such that \(\Vert c-p\Vert =r\). By the law of cosines we have \(\Vert \tilde{c}-c\Vert ^2 = \Vert \tilde{c}-p\Vert ^2 - \Vert p-c\Vert ^2 + 2\Vert \tilde{c}-c\Vert \Vert c-p\Vert \cos \alpha\), where \(\alpha\) is the angle between \(\tilde{c}-c\) and \(p-c\). We further know that \(\cos \alpha \le 0\) since \(p\in H^-\). Thus, \(\Vert \tilde{c}-c\Vert ^2 \le (1+\varepsilon )^2 r^2 - r^2 = (2\varepsilon + \varepsilon ^2) r^2 \le 3\varepsilon r^2\). Taking the square root yields the claim. \(\square\)

Now we are ready to prove the main result, namely that running \(O(\frac{1}{\varepsilon ^4})\) iterations, S is indeed a weak \(\varepsilon\)-coreset for the smallest enclosing ball of P.

Theorem 3

Let S be the set of points that the sub-gradient method chooses as furthest points from the current center in each of  \(O(\frac{1}{\varepsilon ^4})\) iterations. Then S is an  \(\varepsilon\)-coreset for the smallest enclosing ball of P.

Proof

Note, that starting with the same point \(c_0\), the decisions made by the algorithm on input S is the same as on input P. Therefore Corollary 1 applies to both of the sets S and P. (Ties in furthest point queries are assumed to be resolved lexicographically.) Now consider the optimal center \(c^*\) for P and the optimal center \(c_S\) for S as well as the best solution \({\tilde{c}}\) found by the sub-gradient method. We can run the gradient algorithm for a number of \(l=O(\frac{1}{\varepsilon ^4})\) iterations to get a center \({\tilde{c}}\) which is a \((1+\varepsilon ^2)\)-approximation for both sets. By Lemma 5 we have that \(\Vert c^*-{\tilde{c}}\Vert \le \sqrt{3}\varepsilon f^*\) and \(\Vert c_S-{\tilde{c}}\Vert \le \sqrt{3}\varepsilon f^*\). So, by the triangle inequality \(\Vert c^*-c_S\Vert \le \Vert c_S-{\tilde{c}}\Vert + \Vert c^*-{\tilde{c}}\Vert \le 2\sqrt{3}\varepsilon f^*\). Rescaling \(\varepsilon\) by a factor of \(\frac{1}{2\sqrt{3}}\) and again leveraging the triangle inequality yields \(\forall p\in P: \Vert c_S-p\Vert \le \Vert c_S-c^*\Vert + \Vert c^*-p\Vert \le (1+\varepsilon ) f^*\) \(\square\)

Bibliographic Remarks The presented result is rather weak compared to the optimal coreset size of \(\lceil \frac{1}{\varepsilon }\rceil\), cf. [16]. However, we chose to present the method in this way since it shows that the gradient method in its basic form can already yield constant size coresets that do neither depend on the number of input points nor on their dimension. Therefore we see this as a generic method for dimensionality reduction and the design of coresets for computational problems. Putting more effort into geometrically analyzing every single step of the gradient algorithm, i.e., leveraging more problem specific structure in the proof of Nesterov’s theorem (Thm. 2, cf. [83]), can lead to even better results which we want to survey here. Badoiu and Clarkson present in their short and elegant seminal work [15] that their gradient algorithm converges to within \(\varepsilon f^*\) error in \(O(\frac{1}{\varepsilon ^2})\) iterations similar to Corollary 1. Their analysis is stronger in the sense that it implicitly shows that the points selected by the algorithm form an \(\varepsilon\)-coreset of size \(O(\frac{1}{\varepsilon ^2})\). Such a result was obtained before in [17] by picking a point in each iteration which is far enough away from the center of the smallest enclosing ball of the current coreset. Taking the furthest point instead, yields the near-optimal bound of \(\frac{2}{\varepsilon }\) from [15]. The complexity was settled with matching upper and lower bounds of \(\lceil \frac{1}{\varepsilon }\rceil\) in [16]. The smallest enclosing ball problem is by far not the only one for which looking at gradient methods can help. For instance Har-Peled et al. [60] have used similar construction methods to derive coresets for support vector machine training. Clarkson [29] has generalized the method by unifying and strengthening the aforementioned results (among others) into one framework for coresets, sparse approximation and convex optimization. A probabilistic sampling argument was already used in [17] as a dimensionality reduction technique for the 1-median problem which led to the first linear time approximation algorithms for the k-median problem in Euclidean and more general metric spaces [3, 69]. The method itself is closely related to stochastic sub-gradient methods where uniform sub-sampling is used for obtaining an unbiased estimator for the actual sub-gradient in each step [92]. Similar approaches are currently under study to reduce the exponential dependency on the dimension for the probabilistic smallest enclosing ball problem [48].

3.3 Random Sampling

The arguably most straightforward way to reduce the size of dataset is to simply pick as many points as permissible uniformly at random. Though it is rare for an algorithm to produce a coreset in this straightforward fashion, we will see that for outlier-resistant problems this approach can already provide reasonable results. In the following we will examine the geometric median, which is one of the few problems for which uniform sampling gives any form of guarantee. Thereafter, we will show how to obtain strong coresets using more involved sampling distributions.

Definition 4

(Geometric Median) Let A be a set of points in \(\mathbb {R}^d\). The geometric median m(A) minimizes

$$\begin{aligned} f(A,m):= \sum _{x\in A}\Vert x-m(A)\Vert . \end{aligned}$$

We will show that uniform sampling is sufficient to obtain a weak-coreset guarantee, The exposition for uniform sampling follows Thorup [94] and Ackerman et al. [3].

Lemma 6

Let A be a set of points in \(\mathbb {R}^d\) and let S be a uniform sample of AThen for any point b with \(\sum _{x\in A}\Vert x-b\Vert \ge (1+\frac{4\varepsilon }{5}) \sum _{x\in A}\Vert x-m(A)\Vert ,\) we have

$$\begin{aligned} \mathbb {P}\left[ \sum _{x\in S}\Vert x-b\Vert < \sum _{x\in S}\Vert x-m(A)\Vert + \frac{\varepsilon |S|}{5 |A|}\sum _{x\in A}\Vert x-m(A)\Vert \right] \le \exp \left( -\frac{\varepsilon ^2 |S|}{144}\right) . \end{aligned}$$

Proof

We have \(\sum _{x\in A}\left( \Vert x-b\Vert - \Vert x-m(A)\Vert \right)> \frac{4\varepsilon }{5}\sum _{x\in A}\Vert x-m(A)\Vert\) and \(\sum _{x\in A}\left( \Vert x-b\Vert - \Vert x-m(A)\Vert \right)> \frac{4\varepsilon }{5}/(1+\frac{4\varepsilon }{5})\sum _{x\in A}\Vert x-b\Vert\). Then

$$\begin{aligned}&\sum _{x\in A}\left( \Vert x-b\Vert - \Vert x-m(A)\Vert \right) \nonumber \\&\quad>\left( 3-\frac{4\varepsilon }{5}\right) \frac{\varepsilon }{5}\sum _{x\in A}\Vert x-m(A)\Vert + \frac{\varepsilon }{5} \sum _{x\in A}\Vert x-b\Vert \nonumber \\&\quad \ge \left( 2-\frac{4\varepsilon }{5}\right) \frac{\varepsilon }{5}\sum _{x\in A}\Vert x-m(A)\Vert + \frac{\varepsilon }{5} \sum _{x\in A}\Vert p-m(A)\Vert \nonumber \\&\quad \ge \frac{\varepsilon }{5}\sum _{x\in A}\Vert x-m(A)\Vert + \frac{\varepsilon }{5}\left( \sum _{x\in A}\left( \Vert p-m(A)\Vert +\frac{\varepsilon }{5}\Vert x-m(A)\Vert \right) \right) \end{aligned}$$
(1)

Now consider the random variable \(X=\sum _{x\in S} \frac{\Vert x-b\Vert - \Vert x-m(A)\Vert + \Vert m(A)-b\Vert }{2\cdot \left( \Vert m(A)-b\Vert + \frac{\varepsilon }{5|A|}\sum _{x\in A}\Vert x-m(A)\Vert \right) }\). Note that \(\sum _{x\in S} \Vert x-b\Vert \le \sum _{x\in S}\Vert x-m(A)\Vert + \frac{\varepsilon |S|}{5 |A|}\sum _{x\in A}\Vert x-b\Vert \Leftrightarrow |X|\le |S|/2\). Due to the triangle inequality, each summand is in between 0 and 1. Furthermore, \(\mathbb {E}[X] = \frac{|S|}{|A|}\sum _{x\in A}\frac{\Vert x-b\Vert - \Vert x-m(A)\Vert + \Vert m(A)-b\Vert }{2\cdot \left( \Vert m(A)-b\Vert + \frac{\varepsilon }{5|A|}\sum _{x\in A}\Vert x-m(A)\Vert \right) }\). Using Eq. 1, we have

$$\begin{aligned} \mathbb {E}[X]= & {} \frac{|S|}{2}\cdot \frac{1}{|A|}\sum _{x\in A}\frac{\Vert x-b\Vert - \Vert x-m(A)\Vert + \Vert m(A)-b\Vert }{\Vert m(A)-b\Vert + \frac{\varepsilon }{5|A|}\sum _{x\in A}\Vert x-m(A)\Vert }\\\ge & {} \frac{|S|}{2}\cdot \frac{1}{|A|}\sum _{x\in A}\frac{\left( 1+\frac{\varepsilon }{5}\right) \cdot \left( \Vert m(A)-b\Vert + \frac{\varepsilon }{5|A|}\sum _{x\in A}\Vert x-m(A)\Vert \right) }{\Vert m(A)-b\Vert + \frac{\varepsilon }{5|A|}\sum _{x\in A}\Vert x-m(A)\Vert }\\= & {} \frac{|S|}{2}\cdot \left( 1+\frac{\varepsilon }{5}\right) \end{aligned}$$

Applying the Chernoff bound, we have

$$\begin{aligned} \mathbb {P}[X\le |S|/2]\le & {} \mathbb {P}\left[ X\le \left( 1-\frac{\varepsilon }{6}\right) \cdot E[X]\right] \\\le & {} \exp \left( -\frac{\varepsilon ^2\cdot \mathbb {E}[X]}{72}\right) \le \exp \left( -\frac{\varepsilon ^2\cdot |S|}{144}\right) . \end{aligned}$$

\(\square\)

We wish to use Lemma 6 to apply a union bound on all candidate points. In a finite metric consisting of n points, we can set \(|S|\ge 144\varepsilon ^{-2}\log n\) and are done. In the continuous setting, this is not as straightforward. We will use a suitably scaled ball-cover and argue that this already suffices to prove the claim for all points.

Theorem 4

Let A be a set of n points in \(\mathbb {R}^d\) and let S be a uniform sample of A with  \(|S|\in \varOmega (d\varepsilon ^{-2}\log \frac{d}{\varepsilon })\). Let m(Aand m(Sbe the geometric medians of A and S, respectively. Then with constant probability, we have

$$\begin{aligned} \sum _{x\in A}\Vert x-m(S)\Vert \le (1+\varepsilon )\cdot \sum _{x\in A}\Vert x-m(A)\Vert . \end{aligned}$$

Proof

Denote by \(\text {OPT}:=\sum _{x\in A}\Vert x-m(A)\Vert\). By Markov’s inequality and a union bound over S, all points in S will be contained in the ball B of radius \(4|S|\frac{\text {OPT}}{n}\) centered around m(A) with probability at least 1 / 4. Obviously, this means that the optimal median m(S) of |S| will be contained in B. Let C be a \(\frac{\varepsilon }{5} \frac{\text {OPT}}{n}\)-ball-cover of B. Setting \(|S|\ge k\cdot d\varepsilon ^{-2} \log \frac{d}{\varepsilon }\) for a sufficiently large absolute constant k ensures that Lemma 6 holds for all points in C with probability at least 1 / 4.

Now let \(c\in C\) be the closest point in the ball-cover to m(S). We have

$$\begin{aligned} \sum _{x\in S}\Vert x-c\Vert\le & {} \sum _{x\in S}\Vert x-m(S)\Vert + \frac{\varepsilon |S|}{5|A|}\sum _{x\in A}\Vert x-m(A)\Vert \\\le & {} \sum _{x\in S}\Vert x-m(A)\Vert + \frac{\varepsilon |S|}{5|A|}\sum _{x\in A}\Vert x-m(A)\Vert \end{aligned}$$

By Lemma 6 we have \(\sum _{x\in A}\Vert x-c\Vert \le (1+\frac{4\varepsilon }{5})\cdot \sum _{x\in A}\Vert x-m(A)\Vert\). Applying the triangle inequality, then yields \(\sum _{x\in A}\Vert x-m(S)\Vert \le \sum _{x\in A}\Vert x-c\Vert + n \cdot \Vert c-m(S)\Vert \le {(1+\frac{4\varepsilon }{5} + \frac{\varepsilon }{5})}\sum _{x\in A}\Vert x-m(A)\Vert\). \(\square\)

Obviously, the uniform sampling approach can not yield a strong coreset even for the geometric median problem. The reason is that the cost can rely on a small constant number of input points even in one dimension. Given any input P of size \(n-2\), we can place two additional points \(p_{\min } < \min P\) and \(p_{\max }> \max P\) arbitrarily far from the points of P without affecting the median. Clearly, any strong coreset must contain these points to preserve the cost for all possible centers, but the probability of sampling one of them is only \(\frac{2}{n}\). This implies that a strong coreset based on uniform sampling must have linear size. To overcome this limitation, Langberg and Schulman [71] introduced the notion of sensitivity for families of functions. For simplicity of presentation, we consider only the special case of the geometric median problem. For a much more general view on sensitivity sampling for numerous other functions the reader is referred to the original literature [45, 71]

Definition 5

(Sensitivity) We define the sensitivity of a point \(x\in P\) as

$$\begin{aligned} s(x)=\underset{c\in \mathbb {R}^d}{\sup } \frac{\Vert x-c\Vert }{\sum \limits _{p\in P} \Vert p-c\Vert }. \end{aligned}$$

We define the total sensitivity as \(S(P) = \sum _{x\in P} s(x).\)

Informally, the sensitivity of a point measures how important it is for preserving the cost over all possible solutions. Note, that from the definition

$$\begin{aligned} \Vert x-c\Vert \le s(x) \sum \limits _{p\in P} \Vert p-c\Vert \end{aligned}$$
(2)

follows for all \(x\in P\) and all centers \(c\in \mathbb {R}^d\). The total sensitivity will turn out to be a crucial measure for the complexity of sampling from the distribution given by the sensitivities of the input points. We begin with describing the sampling scheme. The following importance or weighted sampling approach yields an unbiased estimator for the cost of any fixed center \(f(P,c) = \sum _{x\in P} \Vert x-c\Vert\). We sample a point \(x\in P\) from the distribution \(q(x)=\frac{s(x)}{S(P)}\) and set \(T=\frac{\Vert x-c\Vert }{q(x)}\). A straightforward calculation of its expected value \(\mathbb {E}[T]=\sum \frac{\Vert x-c\Vert }{q(x)}\cdot q(x)= \sum \limits _{x\in P} \Vert x-c\Vert\) shows that T is unbiased.

Next we bound its variance which is a crucial parameter in deriving concentration results for our sampling procedure.

Lemma 7

\(\mathbb {V}[T]\le (S(P)-1) \mathbb {E}[T]^2\)

Proof

Let \(S=S(P)\). Now \(\frac{1}{\mathbb {E}[T]^2} \mathbb {V}[T]\) equals

$$\begin{aligned}&\frac{1}{\mathbb {E}[T]^2} \sum _{x\in P} \left( \frac{\Vert x-c\Vert }{q(x)} - \mathbb {E}[T]\right) ^2\cdot q(x)\\&\quad =\frac{1}{\mathbb {E}[T]^2} \sum _{x\in P} \left( \frac{\Vert x-c\Vert \cdot S}{s(x)} - \mathbb {E}[T] \right) ^2\cdot \frac{s(x)}{S}\\&\quad = \frac{1}{\mathbb {E}[T]^2} \sum _{x\in P} \left( \frac{\Vert x-c\Vert ^2 \cdot S}{s(x)} - 2\mathbb {E}[T]\Vert x-c\Vert + \frac{s(x)\mathbb {E}[T]^2}{S} \right) \\&\quad = \left( \frac{1}{\mathbb {E}[T]^2} \sum _{x\in P} \frac{\Vert x-c\Vert ^2 \cdot S}{s(x)} \right) - \left( \frac{2}{\mathbb {E}[T]} \sum _{x\in P} \Vert x-c\Vert \right) + \left( \sum _{x\in P} \frac{s(x)}{S} \right) \\&\quad = \left( \frac{1}{\mathbb {E}[T]^2} \sum _{x\in P} \frac{\Vert x-c\Vert ^2 \cdot S}{s(x)} \right) - 2 + 1 \le \frac{S}{\mathbb {E}[T]} \sum _{x\in P} \Vert x-c\Vert - 1 = S-1\,. \end{aligned}$$

The inequality follows from Eq. (2). \(\square\)

In our next lemma we will see why the total sensitivity plays such an important role for the random sampling proportional to the sensitivities.

Lemma 8

Let \(\varepsilon>0\)Let R be a random sample of size \(m \ge \frac{3S}{\varepsilon ^2}\ln (\frac{2\kappa }{\delta })\) drawn i.i.d. from P proportional to the distribution qThen for any fixed set of centers \(C\subset \mathbb {R}^d\) of size \(|C|\le \kappa\) we have

$$\begin{aligned} \mathbb {P}\left[ \exists c\in C : \left| \frac{1}{m}\sum _{x\in R} \frac{\Vert x-c\Vert }{q(x)} - \sum _{x\in P} \Vert x-c\Vert \right| \ge \varepsilon \sum _{x\in P} \Vert x-c\Vert \right] \le \delta . \end{aligned}$$

Proof

Fix any center \(c\in \mathbb {R}^d\). Let \(T_i=\frac{\Vert x-c\Vert }{q(x)}=\frac{\Vert x-c\Vert S}{s(x)}\) with probability q(x) and let \(T=\sum _{i=1}^m T_i\) be their sum. By linearity we have \(\mathbb {E}[T]=m\mathbb {E}[T_i]=m\sum _{x\in P} \Vert x-c\Vert\). By independence of the individual samples and Lemma 7 we have \(\mathbb {V}[T]= m\mathbb {V}[T_i]\le m (S-1) \mathbb {E}[T_i]^2\). Note that by Eq. (2) it holds that \(0\le T_i = \frac{\Vert x-c\Vert S}{s(x)} \le S \sum _{x\in P} \Vert x-c\Vert = S \mathbb {E}[T_i]\) and therefore \(|T_i-\mathbb {E}[T_i]|\le M:= S \mathbb {E}[T_i]\) is a bound that holds almost surely for each i. Now, by an application of Bernstein’s inequality [20] we have \(\mathbb {P}[|T-\mathbb {E}[T]|>\varepsilon \mathbb {E}[T]] \le 2\exp \left( - \frac{\varepsilon ^2 \mathbb {E}[T]^2}{2\mathbb {V}[T]+\frac{2M}{3} \varepsilon \mathbb {E}[T]} \right) .\) We bound the exponent by

$$\begin{aligned} \frac{\varepsilon ^2 \mathbb {E}[T]^2}{2\mathbb {V}[T]+\frac{2M}{3} \varepsilon \mathbb {E}[T]}\ge & {} \frac{\varepsilon ^2 m^2 \mathbb {E}[T_i]^2}{2m (S-1) \mathbb {E}[T_i]^2 + \frac{2}{3} \varepsilon m S \mathbb {E}[T_i]^2}\\\ge & {} \frac{\varepsilon ^2 m }{2 (S-1) + \frac{2}{3} \varepsilon S} \ge \frac{\varepsilon ^2 m }{3 S}\ge \ln \left( \frac{2\kappa }{\delta }\right) . \end{aligned}$$

It follows that \(\mathbb {P}[|T-\mathbb {E}[T]|>\varepsilon \mathbb {E}[T]] \le \frac{\delta }{\kappa }\) which implies our claim by taking a union bound over the set C of size \(|C|\le \kappa\). \(\square\)

Lemma 8 shows that the number of samples that we need, depends linearly on the total sensitivity as well as logarithmically on the number of centers for which we need a guarantee. First we want to bound the total sensitivity for the geometric median problem. If for every input point, there is some center such that the points’ contribution to the cost is large, say bounded below by a constant, then we are lost since in that case the total sensitivity sums up to \(\varOmega (n)\). The next lemma shows that this cannot happen. Actually it turns out that the total sensitivity can be bounded by a constant.

Lemma 9

\(S(P)=\sum _{x\in P} s(x) \le 6.\)

Proof

Let \(c^*\) be the optimal center and let \(\varDelta = \sum _{x\in P} \Vert x-c^*\Vert>0\) denote the optimal cost. Let \(c\in \mathbb {R}^d\) be any center, let \(\varGamma = \sum _{x\in P} \Vert x-c\Vert \ge \varDelta\) denote its cost and let \(\rho = \Vert c-c^*\Vert\) be its distance to the optimal center. Now consider a ball B centered at \(c^*\) with radius \(\frac{2\varDelta }{n}\). By the Markov inequality we have that the number of points inside the ball is at least \(|P\cap B|\ge \frac{n}{2}\). Therefore we have \(\varGamma \ge \frac{n}{2}\max (0,\rho -\frac{2\varDelta }{n})\). Combining both lower bounds we have \(\varGamma \ge \frac{1}{2}[\frac{n}{2}\max (0,\rho -\frac{2\varDelta }{n})+\varDelta ]\). Now,

$$\begin{aligned} s(x)= & {} \underset{c\in \mathbb {R}^d}{\sup }\;\frac{\Vert x-c\Vert }{\sum _{x\in P} \Vert x-c\Vert } \le \underset{c\in \mathbb {R}^d}{\sup }\;\frac{\rho + \Vert x-c^*\Vert }{\frac{1}{2}[\frac{n}{2}\max (0,\rho -\frac{2\varDelta }{n})+\varDelta ]}\\= & {} \underset{\rho \ge 0}{\sup }\;\frac{\rho + \Vert x-c^*\Vert }{\frac{1}{2}[\frac{n}{2}\max (0,\rho -\frac{2\varDelta }{n})+\varDelta ]}= \underset{\rho \ge \frac{2\varDelta }{n}}{\sup }\;\frac{\rho + \Vert x-c^*\Vert }{\frac{1}{2}[\frac{n}{2}(\rho -\frac{2\varDelta }{n})+\varDelta ]}\\= & {} \underset{\rho \ge \frac{2\varDelta }{n}}{\sup }\;\frac{4\rho + 4\Vert x-c^*\Vert }{n\rho } = \underset{\rho \ge \frac{2\varDelta }{n}}{\sup }\;\frac{4}{n} + \frac{4\Vert x-c^*\Vert }{n\rho } := \underset{\rho \ge \frac{2\varDelta }{n}}{\sup }\;G(\rho ). \end{aligned}$$

G, restricted to \(\rho \ge \frac{2\varDelta }{n}\) is clearly monotonously decreasing in \(\rho\) and is thus maximized at \(G(\frac{2\varDelta }{n})=\frac{4}{n} + \frac{2\Vert x-c^*\Vert }{\varDelta }\). We can conclude that S(P) is bounded by \(\sum _{x\in P} s(x) \le \sum _{x\in P} \left( \frac{4}{n} + \frac{2\Vert x-c^*\Vert }{\varDelta } \right) = 4+2\;\frac{\sum _{x\in P}\Vert x-c^*\Vert }{\varDelta } = 6.\) \(\square\)

Now we know that the total sensitivity contributes only a constant factor to our sampling complexity. Before we move to the main result of this section, we will need another technical lemma. It is not immediately clear that we can use the triangle inequality for our samples, due to reweighting. We will therefore establish a relaxation of the triangle inequality for the entire weighted subsample.

Lemma 10

Let \(\varepsilon>0\)Let R be a random sample of size \(m\ge \frac{18}{\varepsilon ^2}\ln (\frac{2}{\delta })\) drawn i.i.d. from  P proportional to the distribution q and reweighted by  \(w(x)=\frac{1}{mq(x)}\)Then with probability at least \(1-\delta\)  for any choice of two centers \(c,c'\in \mathbb {R}^d\) we have

  1. 1.

    \({\text {cost}}(R,c) \le {\text {cost}}(R,c') + (1+\varepsilon )n \Vert c-c'\Vert\)

  2. 2.

    \({\text {cost}}(R,c) \ge (1-\varepsilon )n \Vert c-c'\Vert -{\text {cost}}(R,c').\)

Proof

The tricky part is to bound the total weight of the sample \(\frac{1}{m}\sum _{x\in R} \frac{1}{q(x)}\). In the light of Lemmas 7 and 8, we define the random variables \(T_i=\frac{1}{q(x)}\). Their expectation is \(\mathbb {E}[T_i]=\sum _{x\in P} \frac{1}{q(x)}\cdot q(x) = n\). To bound their variance, we first need an upper bound on \(\frac{1}{s(x)}\). To this end, consider a ball \(B\supset P\) containing all input points, centered at the optimal center \(c^*\). Let \(\varDelta\) be the diameter of B and let \(b\in B\) be the closest point to c located on the surface of B. Note that \(\forall p\in P: \Vert p-b\Vert \le \varDelta\). Then

$$\begin{aligned} \frac{1}{s(x)}= & {} \underset{c\in \mathbb {R}^d}{\inf } \frac{\sum _{p\in P} \Vert p-c\Vert }{\Vert x-c\Vert } \le \underset{c\notin B}{\inf } \frac{\sum _{p\in P} \Vert p-b\Vert +\Vert b-c\Vert }{\Vert b-c\Vert }\nonumber \\\le & {} \underset{c\notin B}{\inf } \left( 1+\frac{\varDelta }{\Vert b-c\Vert }\right) n = n. \end{aligned}$$
(3)

Now, \(\mathbb {V}[T_i]\le (S-1)\mathbb {E}[T_i]\) follows via very similar calculations as in Lemma 7. Inequality (2) is simply replaced by Inequality (3) in the derivation.

Following the arguments from Lemma 8 closely, we can conclude that with high probability \(|\frac{1}{m}\sum _{x\in R} \frac{1}{q(x)}-n|\le \varepsilon n\) holds.

Back to the main claims, the standard triangle inequality yields

$$\begin{aligned} \frac{1}{m}\sum _{x\in R} \frac{\Vert x-c\Vert }{q(x)}\le & {} \frac{1}{m}\sum _{x\in R} \frac{\Vert x-c'\Vert +\Vert c-c'\Vert }{q(x)}\\\le & {} \frac{1}{m}\sum _{x\in R} \frac{\Vert x-c'\Vert }{q(x)}+(1+\varepsilon )n \Vert c-c'\Vert . \end{aligned}$$

The second claim follows similarly using \(\Vert x-c\Vert \ge {\Vert c-c'\Vert} -{\Vert x-c'\Vert}\) in the numerator. \(\square\)

It remains to bound the number of centers for which we will need a guarantee. This seems impossible at first glance, since the strong coreset property asks to hold for every center \(c\in \mathbb {R}^d\) which has infinite cardinality. In the following we will see, that again, it is sufficient to consider a ball of appropriate radius centered at the optimum and decompose it by an \(\varepsilon\)-ball-cover. Lemma 8 is applied to these points only. Now, for every center inside the ball, there is a point of the cover close to it for which the coreset guarantee holds. For the other centers we can argue that they are so far from the optimum center that their distance dominates the cost for both, the original point set as well as for the coreset. This will establish the coreset property for every possible center.

Theorem 5

Let \(\varepsilon>0\), \(c\in \mathbb {R}^d\)Let R be a random sample of P of size \(m \ge \varTheta (\frac{d}{\varepsilon ^2}\ln \frac{1}{\varepsilon \delta })\) drawn i.i.d. proportional to the distribution q and reweighted by  \(w(x)=\frac{1}{mq(x)}\)Then R is a strong coreset for the geometric median of P with probability at least \(1-\delta\).

Proof

Let \(c^*\) be the optimal center for P and let \({\text {OPT}}={\text {cost}}(P,c^*)\) denote the optimal cost. Consider the closed ball B of radius \(r=\frac{{\text {OPT}}}{\varepsilon n}\) centered at \(c^*\). Let C be the set of center points of a \(\varepsilon \frac{{\text {OPT}}}{n}\)-ball-cover of B. For technical reasons we add \(c^*\) to C. Recall from Lemma 1 that \(|C|\le \kappa \le \left( 1+\frac{2 }{\varepsilon ^2}\right) ^d+1 = \left( \frac{1}{\varepsilon }\right) ^{O(d)}.\) We apply Lemma 8 to C with \(m=\frac{18}{\varepsilon ^2} \ln (\frac{2\kappa }{\delta })=\varTheta (\frac{d}{\varepsilon ^2}\ln \frac{1}{\varepsilon \delta })\). Thus we can assume that with probability \(1-\delta\) for all \({\tilde{c}}\in C\) simultaneously and in particular for \(c^*\) we have \(|{\text {cost}}(R,{\tilde{c}})- {\text {cost}}(P,{\tilde{c}})| \le \varepsilon \, {\text {cost}}(P,{\tilde{c}})\). Also we have that the two triangle inequalities from Lemma 10 hold with probability \(1-\delta\).

Now let \(c\in \mathbb {R}^d\) be any center. We distinguish between the cases \(c\in B\) and \(c\notin B\). In the former case let \({\tilde{c}}\in C\) be the closest center to c in our cover, i.e., \(\Vert c-{\tilde{c}}\Vert \le \varepsilon \frac{{\text {OPT}}}{n}\). Using Lemma 10 we have

$$\begin{aligned} {\text {cost}}(R,c)\le & {} {\text {cost}}(R,{\tilde{c}}) + (1+\varepsilon )n\cdot \Vert c-{\tilde{c}}\Vert \\\le & {} (1+\varepsilon )\, {\text {cost}}(P,{\tilde{c}}) + (1+\varepsilon )n\cdot \Vert c-{\tilde{c}}\Vert \\\le & {} (1+\varepsilon ) \,({\text {cost}}(P,c) + n\cdot \Vert c-{\tilde{c}}\Vert ) + (1+\varepsilon )n\cdot \Vert c-{\tilde{c}}\Vert \\\le & {} (1+\varepsilon ) \,{\text {cost}}(P,c) + 2(1+\varepsilon )n\cdot \Vert c-{\tilde{c}}\Vert \\\le & {} (1+\varepsilon ) \,{\text {cost}}(P,c) + 3\varepsilon \,{\text {OPT}}\\\le & {} (1+4\varepsilon ) \,{\text {cost}}(P,c). \end{aligned}$$

The lower bound of \((1-4\varepsilon )\, {\text {cost}}(P, c)\) can be derived similarly and is omitted for brevity of presentation. We are left to deal with the case \(c\notin B\). The inequality to keep in mind is \(\Vert c-c^*\Vert>\frac{{\text {OPT}}}{\varepsilon n}\) or, equivalently \({\text {cost}}(P,c^*)={\text {OPT}}<\varepsilon n\Vert c-c^*\Vert\). From

$$\begin{aligned} {\text {cost}}(R,c)\le & {} {\text {cost}}(R, c^*) + (1+\varepsilon )n\Vert c-c^*\Vert \\\le & {} (1+\varepsilon ) \,({\text {cost}}(P, c^*) + n\Vert c-c^*\Vert )\\\le & {} (1+\varepsilon ) (\varepsilon n\Vert c-c^*\Vert + n\Vert c-c^*\Vert )\\\le & {} (1+3\varepsilon )n\Vert c-c^*\Vert \end{aligned}$$

and

$$\begin{aligned} {\text {cost}}(P,c)\ge & {} n\Vert c-c^*\Vert - {\text {cost}}(P, c^*)\\\ge & {} n\Vert c-c^*\Vert - \varepsilon n\Vert c-c^*\Vert \\\ge & {} (1-\varepsilon )n\Vert c-c^*\Vert \end{aligned}$$

it follows that \(\frac{{\text {cost}}(R,c)}{{\text {cost}}(P,c)}\le \frac{1+3\varepsilon }{1-\varepsilon } \le 1+8\varepsilon .\)

The lower bound of \(1-3\varepsilon\) can be verified very similarly and is again omitted from the presentation. Rescaling \(\varepsilon\) and \(\delta\) by absolute constants concludes the proof. \(\square\)

Bibliographic Remarks The arguably first paper to use random sampling for coreset construction is due to Chen [27]. Like the result on k-means described in Sect. 3.1, he considered balls of exponential increasing radii. Instead of using a ball-cover, he showed that a uniform random sample suffices to approximate the cost and thereby gave the first coresets for k-means and k-median with polynomial dependency on k, d, \(\varepsilon\), and \(\log n\). The sensitivity sampling approach was introduced by Langberg and Schulman [71] and further improved by Feldman and Langberg [45] and Braverman, Feldman and Lang [22].

The technique is now one of the crown jewels of coreset construction, giving the best currently available parameters for a wide range of problems including \(\ell _p\) regression, k-means and k-median clustering, and low rank subspace approximation. It has defined a unified framework for several importance sampling approaches from the early literature of sampling based sublinear approximation algorithms and coresets. One of the first such works is due to Clarkson [28] on \(\ell _1\) regression. After a series of studies on sampling based randomized linear algebra [36,37,38], the approach could also be adapted for \(\ell _2\) regression [40, 41] and was later generalized in [34] to \(\ell _p\) regression for \(p\in [1,\infty )\). In accordance with the statistical meaning of the sampling weights, they were later called (statistical) leverage scores. It was an open problem proposed in [40] to even approximate the \(\ell _2\) leverage scores in less time than needed to solve the regression problem exactly. This problem was resolved in [39] via the oblivious random projection approach we are going to detail in the next section. Low rank subspace approximation was often treated implicitly using the same approaches leading to weak coresets, cf. [91] for a technical review. The first strong coresets of polynomial size for \(\ell _p\) subspace approximation are due to Feldman et al. [47] again achieved via weighted sampling techniques related to the notion of sensitivity. More recently, the sampling based methods from randomized linear algebra [41] were leveraged to develop coresets in [79] for dependency networks [61] and Poisson dependency networks [55].

3.4 Sketches and Projections

It is often useful to view a dataset as a matrix, where the ith row corresponds to the ith point. In the following, let us assume that we have n points in d-dimensional Euclidean space, i.e. we are given a matrix \(A\in \mathbb {R}^{n\times d}\).

A sketch of A is a linear projection obtained by multiplying A with a sketching matrix \(S\in \mathbb {R}^{m\times n}\) for some \(m\ll n\). Our goal is to design a sketching matrix such that SA retains the key properties of A. Many of the previous coreset constructions can also be viewed in terms of sketching matrices. For instance, a sampling algorithm can be viewed as choosing a diagonal matrix \(S\in \mathbb {R}^{n\times n}\) where the diagonal entries are 1 (or some weight) if the point was picked and 0 otherwise in which case the row can as well be deleted.

We are more interested in oblivious sketching matrices, that is the sketching matrix S can be constructed ahead of time without viewing the data. Though it might seem surprising that this yields any results, sketching is now regarded as one of the central tools for streaming. We will illustrate the power of oblivious sketching in the context of subspace embedding and then show how it may be applied to linear regression.

Definition 6

Let \(U\in \mathbb {R}^{n\times d}\) be a matrix with orthogonal unit columns and let \(\varepsilon>0\). An \(\varepsilon\)-subspace embedding of U is a matrix \(SU\in \mathbb {R}^{m\times d}\) for \(m\ll n\) such that for any vector \(x\in \mathbb {R}^d\), we have

$$\begin{aligned} \vert \Vert SUx\Vert ^2 - \Vert Ux\Vert ^2 \vert \le \varepsilon \cdot \Vert x\Vert ^2. \end{aligned}$$

The central ingredient is based around the seminal Johnson-Lindenstrauss lemma [64].

Lemma 11

(Distributional Johnson-Lindenstrauss Lemma) There exists a distribution D over \(m\times n\) matrices with \(m\in O(\varepsilon ^{-2}\log (1/\delta ))\) such that for a matrix \(\varPi\)  drawn from D and any fixed vector \(x\in \mathbb {R}^n\) we have

$$\begin{aligned} \mathbb {P}[(1-\varepsilon )\Vert x\Vert \le \frac{1}{\sqrt{m}}\Vert \varPi x\Vert \le (1+\varepsilon )\Vert x\Vert ]\ge 1 - \delta . \end{aligned}$$

The classic variant of the distributional Johnson-Lindenstrauss Lemma further states that an adequate distribution independently draws each entry of the matrix as a Gaussian with mean 0 and variance 1. Any distribution with mean 0 and variance 1 may be used, and nowadays, it is more common to draw entries from the Rademacher distribution, where each entry is with probability 1 / 2 either 1 or \(-1\). We will briefly discuss advantages of various sketching matrices at the end of this section.

Lemma 11 now gives us a powerful dimension reduction tool for Euclidean spaces. Consider, for instance, the case where we are given \(\eta\) points in an arbitrary number of dimensions. Each distance between two points can be represented by a vector and there are at most \(\eta ^2\) such vectors. Using Lemma 11, we can achieve an \((1\pm \varepsilon )\)-approximate embedding into \(O(\varepsilon ^{-2}\log \eta )\) dimensions with probability at least 2 / 3 by setting \(m\ge c\cdot \varepsilon ^{-2}\log (\eta ^2)\), where c is some absolute constant.

For subspaces, the application of the union bound is not as straightforward as there are infinitely many vectors, even in a 1-dimensional subspace. We first observe that any vector x may be rewritten as \(\Vert x\Vert \cdot \frac{x}{\Vert x\Vert }\) and hence it is sufficient to consider vectors with unit norm. This solves the subspace approximation problem in a single dimension, but even in 2 dimensions we have infinitely many unitary vectors. Instead, we show that applying the union bound on ball-covers is sufficient. Recall that there exists an \(\varepsilon\)-ball-cover of size \((1+2/\varepsilon )^d\) (c.f. Lemma 1). Applying the union bound now gives us \(m\ge \frac{c\log (1+2/\varepsilon )^d}{\varepsilon ^2} = O(d\varepsilon ^{-2}\log \varepsilon ^{-1})\). A slightly more detailed analysis will allow us to remove the \(\log \varepsilon ^{-1}\) factor. The main argument is that we can write every vector as linear combination of the vectors of a 1 / 2-cover. To see this, consider a 1 / 2-cover \(\{x_0,\ldots x_{5^d}\}\) and an arbitrary unit vector y. There exists some \(x_i\) such that \(\Vert y-x_i\Vert \le 1/2\). We then consider the vector \(\Vert y-x_i\Vert x_j\) closest to \((y-x_i)\). Since \(\Vert \frac{1}{\Vert y-x_i\Vert }(y-x_i)-x_j\Vert \le 1/2\), we then have \(\Vert (y-x_i)-\Vert y-x_i\Vert x_j\Vert \le 1/4.\) In each subsequent step, we halve the length of the next vector added and by iterating this sequence into infinity, the length of the remaining vector \(\lim y-\sum \alpha _i x_i\) is 0.

Hence, there exists a subspace approximation of size \(O(d/\varepsilon ^2)\). Now let us consider linear regression, where we aim to minimize \(\Vert Ax-b\Vert\) with \(A\in \mathbb {R}^{n\times d}\) and \(b\in \mathbb {R}^n\). Since all possible vectors \(Ax-b\) lie in a \((d+1)\)-dimensional subspace spanned by the columns of A and b, an oblivious subspace embedding is a coreset for linear regression.

Theorem 6

Let \(A\in \mathbb {R}^{n\times d}\) and \(b\in \mathbb {R}^n\)Choose  \(S\in \mathbb {R}^{m\times n}\) to be a Rademacher matrix, where \(m\in O(d\varepsilon ^{-2})\)Then with constant probability, for all \(x\in \mathbb {R}^d\)we have

$$\begin{aligned} (1-\varepsilon )\Vert Ax-b\Vert \le \Vert S(Ax-b)\Vert \le (1+\varepsilon )\Vert Ax-b\Vert . \end{aligned}$$

Bibliographic Remarks Oblivious subspace embeddings were first introduced by Sarlos [91] both as a means for solving regression and as means for computing low rank approximations. The upper bound of \(O(d/\varepsilon ^2)\) on the target dimension for oblivious sketches was (implicitly) given in the work by Clarkson and Woodruff [31], and this bound was later proved to be optimal by Nelson and Nguyen [82], though even smaller bounds are possible if the sketch is not oblivious, see for instance [33, 49]. It is worth noting that if we are only interested in approximating the optimum using the sketch for linear regression, somewhat akin to a weak coreset guarantee, a target dimension of \(O(d/\varepsilon )\) is sufficient and necessary [31].

Projections and sketches also play an important role for coresets in Euclidean k-means [33, 49]. Cohen et al. [33] showed that an oblivious sketch of target dimension \(O(k/\varepsilon ^2)\) is cost-preserving. As a corollary, this implies that for any coreset construction, the dependency on d may be replaced by a dependency on \(k/\varepsilon ^2\).

The computation time for all sketching approaches can be generally regarded as acceptable. The smallest target dimension is achievable by computing the SVD or a sufficiently good low rank approximation [33, 49]. While this can be done in polynomial time, it is more expensive than the oblivious sketching methods we describe in the following. Conceptually, there are two basic approaches to multiply sketching matrices faster. The first one is to improve the sketching dimension. This, however, is known to be impossible for various regimes, see [12, 63, 65, 72] and very recently for any embedding method by Larsen and Nelson [73]. The other direction is to make the sketching matrix sparse.

Sparse matrices tend to distort sparse vectors. Clearly, the Johnson-Lindenstrauss guarantee cannot hold in such cases. To remedy this problem, Ailon and Chazelle [9] proposed to first increase the density of the input vectors by preforming a randomized Fourier Transform before multiplying with a very sparse embedding matrix. This approach, called the Fast Johnson Lindenstrauss Transform, was later improved and refined, see for instance [10, 11, 21, 95] and references therein.

The second approach for sparse Johnson Lindenstrauss transforms is based around sophisticated ways of combining Count Sketch estimators. The Count Sketch estimator [25] originally proposed for finding heavy hitters is essentially a single row of a Rademacher matrix. Instead of naively repeating the Count Sketch and averaging the results as in a standard Rademacher matrix, we aim to partition the entries of the vectors that are to be embedded, apply a Count Sketch for each partition and thereafter aggregate the results. The degree of sparsity that may be achieved using various partitioning schemes have been studied in [2, 32, 35, 66, 81]. We want to highlight the work by Clarkson and Woodruff [32] who can achieve an embedding in essentially input sparsity time, however at the cost of slightly larger target dimension \(\frac{d^2}{\varepsilon ^2}\). The squared dependency on d was also shown to be necessary by Nelson and Nguyen [80].

Another related direction is generalizing to \(\ell _p\) subspace embeddings. A first step was done by Woodruff and Sohler [93] who designed the first subspace embedding for \(\ell _1\) via Cauchy random variables. The method is in principle generalizable to using p-stable distributions and was improved in [30, 77]. The idea is that the sum of such random variables forms again a random variable from the same type of distribution leading to concentration results for the \(\ell _p\) norm under study. At the same time it is inherently limited to \(1\le p\le 2\) as in [77], since no such distributions exist for \(p>2\), cf. [96]. The first attempts to generalize to \(p>2\) [30] had nearly linear size, namely \(n/{\text {poly}}(d)\), which clearly was not satisfying. A remedy came with a manuscript of Andoni [13], who discovered the max stability of inverse exponential random variables as a means to embed \(\ell _p, p>2\) with little distortion into \(\ell _\infty\). Combining this with the work on \(\ell _2\) embeddings of Clarkson and Woodruff [32] culminated in oblivious subspace embeddings for all p (into \(\ell _2\) resp. \(\ell _\infty\)) and only \({\text {poly(d)}}\) distortion [96]. Note, that the embedding dimension for \(p>2\) is \(n^{1-2/p}{\text {poly}}(d)\) which improved upon the previous \(n/{\text {poly}}(d)\) and is close to optimal given the lower bound of \(\varOmega (n^{1-2/p})\) [86]. The desirable \((1\pm \varepsilon )\) distortion can be achieved using the embeddings for preconditioning and sampling proportional to the \(\ell _p\) leverage scores [30, 34, 96].

Current research has moved beyond strict algorithmic and optimization related characteristics to the study of statistical properties [76, 88]. In particular, not only maximum likelihood estimators are approximated under random projections. Geppert et al. [54] showed that in important classes of Bayesian regression models, the whole structure of the posterior distribution is preserved. This yields much faster algorithms for the widely applicable and flexible, but at the same time computationally demanding Bayesian machinery.

Recently a series of optimization software based on random projections and sketches have appeared. A parallel least squares regression solver LSRN was developed in [78, 97]. An implementation of some of the presented sketching techniques named RaProR was made available for the statistics programming language R [53, 54, 87].

4 Streaming

Streaming algorithms process the data point by point (or entry by entry) and aim to (approximately) answer queries to the data using as little space as possible. Though we describe the coreset constructions in the last chapters with no streaming implementation in mind, it turns out that if we can find a strong coreset for a problem, there also exists a streaming algorithm with little overhead in terms of space requirement. Once a coreset construction is known for a problem, it is often not necessary to develop a specific streaming algorithm, since one can rely on the following black box reduction:

For most functionsFootnote 4, coresets have the very useful property of being closed under union, that is, for two point sets, the union of coresets for both point sets is a coreset for the entire point set. To get an idea of how the reduction works, assume that we partition the input sequence into \(n/\log n\) batches of size \(\log n\). These batches form the leaves of a binary tree of height \(h\le \log n\). Whenever we process an entire batch, we compute a coreset. Whenever we have computed a coreset for two children of a node, we aggregate them by recomputing a coreset of the union of the children and storing it in the parent node. The children can be deleted at this point. Thus, we only have to store at most two coresets at each level bounding the number of coresets in memory to \(O(\log n)\).

So, this framework comes not for free, but its blow up remains bounded. If we apply the merging step as a black box, the coreset contained at the root node, i.e. the final output of the algorithm will have an approximation guarantee of \(\left(1+\varepsilon \right)^{\log n}\). Rescaling \(\varepsilon\) by \(2\log n\), we have the desired \(\left(1+\frac{\varepsilon }{2\log n}\right)^{\log n} \le \exp \left(\frac{\varepsilon }{2}\right) \le \left(1+\varepsilon \right)\) approximation ratio. Finally, many known constructions require randomization and have some adjustable failure probability \(\delta\). To limit the overall failure probability when processing a stream, \(\delta\) is rescaled by the number of coreset constructions. Since the space dependency on \(\delta\) is typically \(\log \frac{1}{\delta }\), we incur another factor of \(O(\log n)\). In total, the space dependency on \(\log n\) is increased by a factor of \(\log ^{c+p} n\), where c is the exponent of \(\varepsilon\) in the offline coreset construction and \(p=1\) if the construction is deterministic and \(p=2\) if it is randomized.

Heinrich et al. have extended this streaming construction of coresets to an asymptotic error guarantee of \(\varepsilon \rightarrow 0\) as \(n\rightarrow \infty\) while the memory remains bounded by \({\text {polylog}}(n)\). This has led to the notion of asymptotically exact streaming algorithms [62].

The framework, called merge and reduce, was originally introduced by Bentley and Saxe [19] and first applied to streaming by Agarwal, Har-Peled and Varadarajan in their seminal paper [4]. Nowadays, many papers have given more efficient streaming implementations for their coresets. For extent approximation algorithms and k-center clustering, we now have constructions with no dependency on n, see  [8, 23, 24, 99], with the currently best algorithms storing \(O(\varepsilon ^{-(d-1)/2})\) points [14] for \(\varepsilon\)-kernels and \(O(k\varepsilon ^{-d})\) points for k-center [98]. The dependency on the dimension d is exponential for all these algorithms and Agarwal and Sharathkumar [7] showed that no algorithm with polynomial dependency on d can exist, see also Sect. 5, unless one is willing to drop the \((1+\varepsilon )\) guarantee for a weaker fixed constant.

k-median and k-means clustering thus far are more reliant on the merge and reduce technique. Certain geometric decompositions avoid this, see [51, 52], but have a much larger offline space dependency compared to other constructions. See Table 1 for an overview.

The oblivious sketching algorithms avoid the merge and reduce framework entirely and immediately translate to the streaming algorithm. In fact, they can operate even when processing deletions and entry-wise modifications to an input matrix. Li, Nelson and Woodruff [74] showed that that essentially any such streaming algorithm may be reformulated in terms of linear sketches.

Table 1 Comparison of memory demands, where lower order factors are suppressed and the memory to store a d-dimensional point is not specified. The constructions for high dimensions do not treat d as a constant and succeed with constant probability

5 Lower Bounds

Lower bounds for coresets come in two flavors: (1) space complexity in terms of points or bits and (2) impossibility results. We will sketch examples for both. First, let us consider the extent approximation problem.

Definition 7

Let A be a set of points in \(\mathbb {R}^d\) and let \(\varepsilon>0\). An extent approximation is a subset S of A such that for any unit vector w, we have

$$\begin{aligned} \left| \underset{x\in A}{~\min ~}w^Tx - \underset{x\in A}{~\max ~}w^Tx\right| \le (1+\varepsilon )\cdot \left| \underset{x\in S}{~\min ~}w^Tx - \underset{x\in S}{~\max ~}w^Tx\right| \end{aligned}$$

Arya and Chan gave an algorithm storing \(O(\varepsilon ^{-(d-1)/2})\) points. We will briefly outline why this is indeed optimal (see Agarwal et al. [5] for details). Consider the unit sphere and the spherical cap with angular radius \(\sqrt{\varepsilon }\). The height of this cap is \(1-\cos (\sqrt{\varepsilon })\le \frac{\varepsilon }{2}\). Thus, the extent approximation must contain at least one point from every cap. The radius of each such cap is \(\sin (\sqrt{\varepsilon })=\varTheta (\sqrt{\varepsilon })\). We know that the bound of Lemma 1 is asymptotically tight, i.e., we require \(\varOmega (\sqrt{\varepsilon }^{-d})\) space for a \(\sqrt{\varepsilon }\)-ball-cover in d-dimensional space. Combining this with the fact that the unit sphere is a \(d-1\)-dimensional space, we know that \(\varOmega (\sqrt{\varepsilon }^{-(d-1)}) = \varOmega (\varepsilon ^{-(d-1)/2})\) points are necessary.

We now present an impossibility result for logistic regression.

Definition 8

(Logistic Regression) Let A be a set of points in \(\mathbb {R}^d\). Then logistic regression objective aims to find a vector w minimizing

$$\begin{aligned} \sum _{x_i\in A}\ln \left( 1+\exp (-w^Tx_i)\right) . \end{aligned}$$

Note, that we assume that the label \(y_i\in \{-1,1\}\) of point \(x_i\) is already folded into \(x_i\).

Let us first recall the reduction of streaming algorithms to the construction of strong coresets from the last section. Taking it the opposite direction, a lower bound for streaming algorithms gives us a lower bound for strong coresets. Thus, a wide variety of tools from communication complexity becomes available to us. A complete review of all techniques is well out of scope, for further reading we refer to the book by Kushilevitz and Nisan [70]. We will focus on the following communication problem.

The indexing problem is a two party communication game, where the first player Alice has a binary bit string \(x\in \{0,1\}^n\) and the second player Bob has an index \(k\in \{1,\ldots ,n\}\). Alice is allowed to send one message to Bob, whereupon Bob has to output the kth bit. The number of bits of the transmitted message required by any randomized protocol succeeding with probability at least 2 / 3 over the random choices of the players is in \(\varOmega (n)\), see [1].

We will use the indexing problem to show that no strong coresets for logistic regression exist.

We first show a reduction to the convex hull membership problem. Here, the stream consist of sequence of points A and at any given time we want to be able to answer whether a given point x lies in the convex hull C(A) or not.

Lemma 12

Let P be a set of n points in \(\mathbb {R}^2\)Let A be a subset of P arriving one after the other in a stream. Then any single pass randomized algorithm deciding with probability 2 / 3 whether some point \(b\in P\) lies in C(Arequires at least \(\varOmega (n)\) space.

Proof

Let \(x\in \{0,1\}^n\) be Alice’s bit string and let k be Bob’s index. For each \(i\in \{1,\ldots ,n\}\), define the point \(p_i=(\sin (i/n),\cos (i/n))\). By construction, all points lie on the unit sphere and therefore \(p_k\) is in the convex hull of any point set \(\bigcup _{i\in I} p_i\) with \(I\subseteq \{1,\ldots ,n\}\) if and only if \(k\in I\). For each entry \(x_i=1\), Alice constructs \(p_i\). She then runs a streaming algorithm on all generated points and sends the memory of the streaming algorithm to Bob. Bob then checks whether \(p_k\) is in the convex hull generated by Alice. Since this solves the indexing problem, the communication complexity of indexing is a lower bound to the space complexity of convex hull membership. \(\square\)

Corollary 2

Let A and B be two sets of a total n points in 2-dimensional space arriving in a data stream. Then any single pass randomized algorithm deciding with probability 2 / 3 whether A and B are linearly separable requires at least \(\varOmega (n)\) space.

We now return to logistic regression. We have the following theorem.

Theorem 7

For any \(\delta>0\) and any integer n there exists a set of n  points \(C=A\cup B\)such that any strong \(\varepsilon\)-coreset of  C for logistic regression must consist of \(\varOmega (n^{1-\delta })\) points.

Proof

Let A and B be linearly separable, we have at least one misclassified point. The cost of this point is lower bounded by \(\ln (1+\exp (0)) = \ln (2)\). Otherwise, let w be a separating hyperplane. Then \(\lim _{\Vert w\Vert \rightarrow \infty }\sum _{x_i\in A}\ln (1+\exp (-w^Tx_i)) = 0\). Given a single pass randomized algorithm for logistic regression we can distinguish between these two cases. Corollary 2 implies that the space complexity must therefore be \(\varOmega (n)\). Since the merge and reduce framework incurs a \({\text {polylog}}(n)\) blowup, this implies a lower bound of \(\varOmega (n/{\text {polylog}}(n))\subset \varOmega (n^{1-\delta })\) for the space complexity of strong coresets for logistic regression. \(\square\)

A very similar impossibility result was derived recently for Poisson regression by reduction from communication complexity problems [79]. The paper discusses and demonstrates how coresets can be useful in practice anyway, going beyond the worst-case perspective.

6 Conclusion and Open Problems

We have outlined techniques and limitations of coreset construction methods. To summarize their benefits in a short statement: Coresets are arguably the state of the art technique to turnFootnote 5 “Big Data into tiny data!” Their design and analysis should be considered whenever a new statistical model or learning task is designed. They make the algorithmic assessment more efficient saving time, space, communication, and energy, to tackle the most common resource restrictions associated with Big Data.

There are further lines of research employing coresets, concerning topics like privacy issues [43] or distributed computation [18, 67, 97], that we have not covered here, but encourage the interested reader to investigate. In the following, we would like to conclude with three important open problems.

Problem 1

Let A be a set of n points in d-dimensional Euclidean space. Is it possible to deterministically compute a coreset S for the k-median or k-means objective in polynomial time, such that \(|S|\in \text {poly}(k,\varepsilon ^{-1},d,\log n)\)?

All known coreset constructions for k-median and k-means clustering with polynomial dependency on \(k,\varepsilon ^{-1},d,\) and \(\log n\) are randomized. The best known deterministic constructions are either exponential in d, or exponential in k and \(\varepsilon ^{-1}\). Since we know that small coresets exist, it is possible to compute them by brute force enumeration, which is deterministic but also clearly infeasible.

Problem 2

Let A be a set of n points in d-dimensional Euclidean space. Is it possible to compute a coreset S for the geometric median objective such that |S| is independent of d and n?

It is a simple exercise to show that coresets with no dependency on n, d, or for that matter \(\varepsilon\) exist for the centroid or mean of a point set. Indeed, there exist coresets for the more general Euclidean k-means problem of size \(\text {poly}(k/\varepsilon )\) [33, 49]. The algorithms and proofs are heavily reliant on connections to linear algebra and in particular the fact that Euclidean k-means is a constrained form of low-rank approximation with respect to the Frobenius norm. The unconstrained low-rank approximation of a matrix can be determined via singular value decomposition (SVD). Computational and mathematical aspects of the SVD are well understood. In contrast, far less is known by the k-median analogue of a low-rank approximation and its computation is invariably harder than the SVD. Currently, we neither know of a lower bound stating that a dependency of d is required for a geometric median coreset, nor of a coreset construction consisting of even \(\exp (\varepsilon ^{-1})\) many points. It is known that for the minimum enclosing ball (i.e. 1-center clustering), an exponential dependency on d is necessary.

Problem 3

Let A be a subset of points of some universe U and let C be a set of candidate solutions. Let \(f:U\times C\rightarrow \mathbb {R}^{\ge 0}\) be a non-negative measurable function. Let S be the smallest possible \(\varepsilon\)-coreset with respect to A and f. Is it possible to always compute a \(\varepsilon\)-coreset \(S'\), such that \(|S'|\le \alpha \cdot |S|\) for some approximation factor \(\alpha\)?

Though we have algorithms that are optimal in the worst case for some problems, the worst case complexity may nevertheless be unappealing, such as is the case for extent problems. It would be nice to have algorithms with instance optimal running times and output sizes. To be more specific, consider the minimum enclosing ball problem. Is it possible to compute a coreset \(S'\) in time \(O(|A|+\alpha \cdot |S|)\) such that \(|S'|\le \alpha \cdot |S|\), where S is the optimal coreset and \(\alpha\) a small (ideally constant) factor?