Lagrangian Relaxations on Networks by ε-Subgradient Methods

Article

Abstract

The efficiency of the network flow techniques can be exploited in the solution of nonlinearly constrained network flow problems by means of approximate subgradient methods. The idea is to solve the dual problem by using ε-subgradient methods, where the dual function is estimated by minimizing approximately a Lagrangian function, which relaxes the side constraints and is subject only to network constraints. In this paper, convergence results for some kind of ε-subgradient methods are put forward. Moreover, in order to evaluate the quality of the solution and the efficiency of these methods some of them have been implemented and their performances computationally compared with codes that are able to solve the proposed test problems.

Keywords

Nonlinear programming Lagrangian relaxation Approximate subgradient methods Network flows 

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. 1.
    Mijangos, E.: An implementation of Newton-like methods on nonlinearly constrained networks. Comput. Oper. Res. 32(2), 181–199 (2004) CrossRefMathSciNetGoogle Scholar
  2. 2.
    Mijangos, E.: An efficient method for nonlinearly constrained networks. Eur. J. Oper. Res. 161(3), 618–635 (2005) CrossRefMATHMathSciNetGoogle Scholar
  3. 3.
    Kiwiel, K.: Convergence of approximate and incremental subgradient methods for convex optimization. SIAM J. Optim. 14(3), 807–840 (2004) CrossRefMATHMathSciNetGoogle Scholar
  4. 4.
    Mijangos, E.: Approximate subgradient methods for nonlinearly constrained network flow problems. J. Optim. Theory Appl. 128(1), 167–190 (2006) CrossRefMATHMathSciNetGoogle Scholar
  5. 5.
    Mijangos, E.: On the efficiency of the ε-subgradient methods over nonlinearly constrained networks. In: System Modeling and Optimization. IFIP Advances in Information and Communication Technology, vol. 199, pp. 101–111 (2006) CrossRefGoogle Scholar
  6. 6.
    Nedić, A., Ozdaglar, A.: Approximate primal solutions and rate analysis for dual subgradient methods. SIAM J. Optim. 19(4), 1757–1780 (2009) CrossRefMATHGoogle Scholar
  7. 7.
    Shor, N.Z.: Minimization Methods for Nondifferentiable Functions. Springer, Berlin (1985) CrossRefGoogle Scholar
  8. 8.
    Ermoliev, Yu.M.: Methods for solving nonlinear extremal problems. Cybernetics 2, 1–17 (1966) CrossRefGoogle Scholar
  9. 9.
    Poljak, B.T.: Introduction to Optimization. Optimization Software Inc., New York (1987) Google Scholar
  10. 10.
    Correa, R., Lemaréchal, C.: Convergence of some algorithms for convex minimization. Math. Program. 62, 261–275 (1993) CrossRefMATHGoogle Scholar
  11. 11.
    Nedić, A., Bertsekas, D.P.: Incremental subgradient methods for nondifferentiable optimization. SIAM J. Optim. 12(1), 109–138 (2001) CrossRefMATHMathSciNetGoogle Scholar
  12. 12.
    Brännlund, U.: On relaxation methods for nonsmooth convex optimization. Ph.D. Thesis, Royal Institute of Technology, Stockholm, Sweden (1993) Google Scholar
  13. 13.
    Goffin, J.L., Kiwiel, K.: Convergence of a simple subgradient level method. Math. Program. 85, 207–211 (1999) CrossRefMATHMathSciNetGoogle Scholar
  14. 14.
    Wächter, A., Biegler, L.T.: On the implementation of a primal-dual interior point filter line search algorithm for large-scale nonlinear programming. Math. Program. 106(1), 25–57 (2006) CrossRefMATHMathSciNetGoogle Scholar
  15. 15.
    Byrd, R.H., Nocedal, J., Waltz, R.A.: KNITRO: an integrated package for nonlinear optimization. In: Di Pillo, G., Roma, M. (eds.) Large-Scale Nonlinear Optimization. Springer, Berlin (2006), pp. 35–59 CrossRefGoogle Scholar
  16. 16.
    Bertsekas, D.P., Nedić, A., Ozdaglar, A.E.: Convex Analysis and Optimization. Athena Scientific, Belmont (2003) MATHGoogle Scholar
  17. 17.
    Bertsekas, D.P.: Nonlinear Programming, 2nd edn. Athena Scientific, Belmont (1999) MATHGoogle Scholar
  18. 18.
    Toint, Ph.L., Tuyttens, D.: On large scale nonlinear network optimization. Math. Program. 48, 125–159 (1990) CrossRefMATHMathSciNetGoogle Scholar
  19. 19.
    Bertsekas, D.P.: Constrained Optimization and Lagrange Multiplier Methods. Academic Press, New York (1982) MATHGoogle Scholar
  20. 20.
    Murtagh, B.A., Saunders, M.A.: Large-scale linearly constrained optimization. Math. Program. 14, 41–72 (1978) CrossRefMATHMathSciNetGoogle Scholar
  21. 21.
    Mijangos, E., Nabona, N.: The application of the multipliers method in nonlinear network flows with side constraints. Technical Report 96/10, Dept. of Statistics and Operations Research Universitat Politècnica de Catalunya, Spain (1996) (downloadable from website http://www.ehu.es/~mepmifee/)
  22. 22.
    Czyzyk, J., Mesnier, M.P., Moré, J.: The NEOS server. IEEE Comput. Sci. Eng. 5(3), 68–75 (1998) CrossRefGoogle Scholar
  23. 23.
    Fourer, R., Gay, D.M., Kernighan, B.W.: AMPL, a Modelling Language for Mathematical Programming. Boyd & Fraser, Danvers (1993) Google Scholar
  24. 24.
    DIMACS: The first DIMACS international algorithm implementation challenge: The bench-mark experiments. Technical Report, DIMACS, New Brunswick, NJ, USA (1991) Google Scholar

Copyright information

© Springer Science+Business Media, LLC 2011

Authors and Affiliations

  1. 1.Department of Applied Mathematics and Statistics and Operations ResearchUniversity of the Basque CountryBilbaoSpain

Personalised recommendations