Abstract
We propose an accelerated gradient-free method with a non-Euclidean proximal operator associated with the p-norm (1 ⩽ p ⩽ 2). We obtain estimates for the rate of convergence of the method under low noise arising in the calculation of the function value. We present the results of computational experiments.
Similar content being viewed by others
References
Rosenbrock, H.H., An Automatic Method for Finding the Greatest or Least Value of a Function, Comput. J., 1960, vol. 3, no. 3, pp. 175–184. doi: https://doi.org/10.1093/comjnl/3.3.175
Brent, R.P., Algorithms for Minimization Without Derivatives, Dover Books on Mathematics, Mineola: Dover, 1973, ISBN 9780486419985. https://books.google.de/books?id=6Ay2biHG-GEC
Spall, J.C., Introduction to Stochastic Search and Optimization, New York: Wiley, 2003.
Rumelhart, D.E., Hinton, G.E., and Williams, R.J., Learning Representations by Back-Propagating Errors, Nature, 1986, no. 323, pp. 533–536.
Schmidhuber, J., Deep Learning in Neural Networks: An Overview, Neural Networks, 2015, vol. 61, pp. 85–117. arXiv:1404.7828.
Goodfellow, I., Bengio, Y., and Courville, A., Deep Learning, Cambridge: MIT Press, 2016.
Nikolenko, S., Kadurin, A., and Arkhangel’skaya, E., Glubokoe obuchenie. Pogruzhenie v mir neironnykh setei (Deep Learning: An Immersion in the World of Neural Networks), St. Petersburg: Piter, 2018.
Nesterov, Yu., Random Gradient-Free Minimization of Convex Functions, Universite catholique de Louvain, Center for Operations Research and Econometrics (CORE), no. 2011001, 2011.
Nesterov, Yu. and Spokoiny, V., Random Gradient-Free Minimization of Convex Functions, Found. Comput. Math., 2017, vol. 17, no. 2, pp. 527–566.
Nesterov, Yu.E., A Method of Solving a Convex Programming Problem with Convergence Rate O(1/k 2), Soviet Math. Dokl., 1983, vol. 27, no. 2, pp. 372–376.
Gasnikov, A.V., Dvurechenskii, P.E., and Nesterov, Yu.E., Stochastic Gradient Methods with an Imprecise Oracle, Tr. MFTI, 2016, vol. 8, no. 1, pp. 41–91, arXiv preprint arXiv:1411.4218.
Chopra, P., Reinforcement Learning without Gradients: Evolving Agents using Genetic Algorithms, https://towardsdatascience.com/reinforcement-learning-without-gradients-evolving-agents-usinggenetic-algorithms-8685817d84f
Vorontsova, E.A., Gasnikov, A.V., and Gorbunov, E.A., Accelerated Directional Search with Non- Euclidean Prox-Structure, Autom. Remote Control, 2019, vol. 80, no. 4, pp. 693–707.
Allen-Zhu, Z. and Orecchia, L., Linear Soupling: An Ultimate Unification of Gradient and Mirror Descent, arXiv preprint arXiv:1407.1537.
Dvurechensky, P., Gasnikov, A., and Tiurin, A., Randomized Similar Triangles Method: A Unifying Framework for Accelerated Randomized Optimization Methods (Coordinate Descent, Directional Search, Derivative-Free Method), arXiv preprint arXiv:1707.08486.
Nemirovskii, A.S. and Yudin, D.B., Slozhnost’ zadach i effektivnost’ metodov optimizatsii (Complexity of Problems and Efficiency of Optimization Methods), Moscow: Nauka, 1979.
Gorbunov, E.A., Vorontsova, E., and Gasnikov, A.V., On the Upper Bound for the Expectation of the Norm of a Vector Uniformly Distributed on the Sphere and the Phenomenon of Concentration of Uniform Measure on the Sphere, Mat. Zam., 2019, vol. 106, no. 1, pp. 13–23.
Bogolubsky, L., Dvurechensky, P., Gasnikov, A., Gusev, G., Nesterov, Y., Raigorodskii, A., Tikhonov, A., and Zhukovskii, M., Learning Supervised PageRank with Gradient-Based and Gradient-Free Optimization Methods, in Lee, D.D., Sugiyama, M., Luxburg, U.V., Guyon, I., Garnett, R., Eds., Proc. 30th Int. Conf. on Neural Information Processing Systems (NIPS’16), pp. 4914–4922, Curran Associates, Inc., 2016, arXiv:1603.00717.
ACDF method Python code. https://github.com/evorontsova/ACDF
Gasnikov, A.V., Effektivnye chislennye metody poiska ravnovesii v bol’shikh transportnykh setyakh (Efficient Numerical Methods for Finding Equilibria in Large-Scale Transportation Networks), Doctoral Dissertation, Moscow: MFTI, 2016, arXiv preprint arXiv:1607.03142.
Acknowledgments
The work shown in Section 3 was supported by the Russian Science Foundation, project no. 17-11-01027. In the remaining sections, the work of A.V. Gasnikov was funded within the framework of the State Support of the Leading Universities of the Russian Federation “5-100” and was supported by the Russian Foundation for Basic Research, project no. 18-31-20005 mol-a-ved, the work of E.A. Gorbunov was supported by the grant of the President of the Russian Federation MD-1320.2018.1, the work of P.E. Dvurechenskii and E.A. Vorontsova was supported by the Russian Foundation for Basic Research, project no. 18-29-03071 mk.
Author information
Authors and Affiliations
Corresponding authors
Additional information
Russian Text © The Author(s), 2019, published in Avtomatika i Telemekhanika, 2019, No. 8, pp. 149–168.
Rights and permissions
About this article
Cite this article
Vorontsova, E.A., Gasnikov, A.V., Gorbunov, E.A. et al. Accelerated Gradient-Free Optimization Methods with a Non-Euclidean Proximal Operator. Autom Remote Control 80, 1487–1501 (2019). https://doi.org/10.1134/S0005117919080095
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1134/S0005117919080095