Scenario reduction revisited: fundamental limits and guarantees
- 170 Downloads
The goal of scenario reduction is to approximate a given discrete distribution with another discrete distribution that has fewer atoms. We distinguish continuous scenario reduction, where the new atoms may be chosen freely, and discrete scenario reduction, where the new atoms must be chosen from among the existing ones. Using the Wasserstein distance as measure of proximity between distributions, we identify those n-point distributions on the unit ball that are least susceptible to scenario reduction, i.e., that have maximum Wasserstein distance to their closest m-point distributions for some prescribed \(m<n\). We also provide sharp bounds on the added benefit of continuous over discrete scenario reduction. Finally, to our best knowledge, we propose the first polynomial-time constant-factor approximations for both discrete and continuous scenario reduction as well as the first exact exponential-time algorithms for continuous scenario reduction.
KeywordsScenario reduction Wasserstein distance Constant-factor approximation algorithm k-median clustering k-means clustering
The authors are indebted to the referees and the guest editors for their comments that considerably improved the manuscript. This research was funded by the SNSF Grant BSCGI0_157733 and the EPSRC Grants EP/M028240/1 and EP/M027856/1.
- 4.Charikar, M., Li, S.: A dependent LP-rounding approach for the \(k\)-median problem. In: Proceedings of the 39th International Colloquium Conference on Automata, Languages, and Programming, pp. 194–205 (2012)Google Scholar
- 6.Dasgupta, S.: CSE 291: Topics in unsupervised learning. http://cseweb.ucsd.edu/~dasgupta/291-unsup/ (2008)
- 10.Gao, R., Kleywegt, A.: Distributionally robust stochastic optimization with Wasserstein distance. (2016) arXiv:1604.02199
- 13.Hanasusanto, G., Kuhn, D.: Conic programming reformulations of two-stage distributionally robust linear programs over Wasserstein balls. (2016) arXiv:1609.07505
- 18.Hopcroft, J., Kannan, R.: Computer science theory for the information age. (2012) https://www.cs.cmu.edu/~venkatg/teaching/CStheory-infoage/
- 26.Mahajan, M., Nimbhorkar, P., Varadarajan, K.: The planar \(k\)-means problem is NP-hard. In: Proceedings of the 3rd International Workshop on Algorithms and Computation, pp. 274–285 (2009)Google Scholar
- 27.Mohajerin Esfahani, P., Kuhn, D.: Data-driven distributionally robust optimization using the Wasserstein metric: Performance guarantees and tractable reformulations. Math. Program. (2017). https://doi.org/10.1007/s10107-017-1172-1
- 30.Pflug, G., Pichler, A.: Approximations for probabilitydistributions and stochastic optimization problems. In: Bertocchi, M., Consigli, G., Dempster, M.A.H. (eds.) Stochastic Optimization Methods in Finance and Energy: New Financial Products and Energy Market Strategies, pp. 343–387. Springer, Berlin (2011)CrossRefGoogle Scholar
- 38.Zhao, C., Guan, Y.: Data-driven risk-averse stochastic optimization with Wasserstein metric. (2015) Available on Optimization OnlineGoogle Scholar