Skip to main content
Log in

PPA-Like Contraction Methods for Convex Optimization: A Framework Using Variational Inequality Approach

  • Published:
Journal of the Operations Research Society of China Aims and scope Submit manuscript

Abstract

Linearly constrained convex optimization has many applications. The first-order optimal condition of the linearly constrained convex optimization is a monotone variational inequality (VI). For solving VI, the proximal point algorithm (PPA) in Euclidean-norm is classical but abstract. Hence, the classical PPA only plays an important theoretical role and it is rarely used in the practical scientific computation. In this paper, we give a review on the recently developed customized PPA in H-norm (H is a positive definite matrix). In the frame of customized PPA, it is easy to construct the contraction-type methods for convex optimization with different linear constraints. In each iteration of the proposed methods, we need only to solve the proximal subproblems which have the closed form solutions or can be efficiently solved up to a high precision. Some novel applications and numerical experiments are reported. Additionally, the original primal-dual hybrid gradient method is modified to a convergent algorithm by using a prediction-correction uniform framework. Using the variational inequality approach, the contractive convergence and convergence rate proofs of the framework are more general and quite simple.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6

Similar content being viewed by others

Notes

  1. About the \( \varepsilon \)-approximate solution, readers can consult the paper by Nesterov [16] (see (2.5)).

References

  1. Higham, N.J.: Computing the nearest correlation matrix—a problem from finance. IMA J. Numer. Anal. 22, 329–343 (2002)

  2. Cai, J.F., Candès, E.J., Shen, Z.W.: A singular value thresholding algorithm for matrix completion. SIAM J. Optim. 20, 1956–1982 (2010)

  3. Chen, C.H., He, B.S., Yuan, X.M.: Matrix completion via alternating direction method. IMA J. Numer. Anal. 32, 227–245 (2012)

  4. Martinet, B.: Regularisation, d’inéquations variationelles par approximations succesives. Rev. Francaise d’Inform. Recherche Oper. 4, 154–159 (1970) (In French)

  5. Rockafellar, R.T.: Monotone operators and the proximal point algorithm. SIAM J. Control. Optim. 14, 877–898 (1976)

  6. Zhu, M., Chan, T.F.: An efficient primal-dual hybrid gradient algorithm for total variation image restoration, CAM Report 08–34, UCLA, (2008)

  7. He, B.S., You, Y.F., Yuan, X.M.: On the convergence of primal-dual hybrid gradient algorithm. SIAM J. Imaging Sci. 7, 2526–2537 (2014)

  8. Chambolle, A., Pock, T.: A first-order primal-dual algorithms for convex problem with applications to imaging. J. Math. Imaging Vis. 40, 120–145 (2011)

  9. Pock, T., Chambolle, A.: Diagonal preconditioning for first order primal-dual algorithms in convex optimization. In: 2011 IEEE International Conference on Computer Vision, pp. 1762–1769 (2011)

  10. He, B.S., Yuan, X.M.: Convergence analysis of primal-dual algorithms for a saddle-point problem: from contraction perspective. SIAM J. Imaging Sci. 5, 119–149 (2012)

  11. Gu, G.Y., He, B.S., Yuan, X.M.: Customized proximal point algorithms for linearly constrained convex minimization and saddle-point problems: a unified approach. Comput. Optim. Appl. 59, 135–161 (2014)

  12. He, B.S., Yuan, X.M., Zhang, W.X.: A customized proximal point algorithm for convex minimization with linear constraints. Comput. Optim. Appl. 56, 559–572 (2013)

  13. You, Y.F., Fu, X.L., He, B.S.: Lagrangian-PPA based contraction methods for linearly constrained convex optimization. Pacific. J. Opt. 10, 199–213 (2014)

  14. Facchinei, F., Pang, J.S.: Finite-Dimensional Variational Inequalities and Complementarity Problems, Volume 1. Springer Series in Operations Research. Springer, New York (2003)

  15. He, B.S., Yuan, X.M.: On the \(O(1/t)\) convergence rate of the alternating direction method. SIAM J. Numer. Anal. 50, 700–709 (2012)

  16. Nesterov, Y.E.: Gradient methods for minimizing composite objective function. Math. Program. Ser. B. 140, 125–161 (2013)

  17. He, B.S., Shen, Y.: On the convergence rate of customized proximal point algorithm for convex optimization and saddle-point problem. Sci. Sin. Math. 42, 515–525 (2012) (In Chinese)

  18. Nemirovski, A.: Prox-method with rate of convergence \(O(1/t)\) for variational inequalities with Lipschitz continuous monotone operators and smooth convex-concave saddle point problems. SIAM J. Optim. 15, 229–251 (2004)

Download references

Acknowledgments

The author is grateful to the two anonymous referees for their careful reading and valuable comments which have helped me substantially improve the presentation of this paper.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Bing-Sheng He.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

He, BS. PPA-Like Contraction Methods for Convex Optimization: A Framework Using Variational Inequality Approach. J. Oper. Res. Soc. China 3, 391–420 (2015). https://doi.org/10.1007/s40305-015-0108-9

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s40305-015-0108-9

Keywords

Mathematics Subject Classification

Navigation