Skip to main content
Log in

Note on a method of conjugate subgradients for minimizing nondifferentiable functions

  • Short Communication
  • Published:
Mathematical Programming Submit manuscript

Abstract

An algorithm is described for finding the minimum of any convex, not necessarily differentiable, functionf of several variables. The algorithm yields a sequence of points tending to the solution of the problem, if any; it requires the calculation off and one subgradient off at designated points. Its rate of convergence is estimated for convex and also for twice differentiable convex functions. It is an extension of the method of conjugate gradients, and terminates whenf is quadratic.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

References

  1. D.P. Bertsekas and S.K. Mitter, “A descent numerical method for optimization problems with nondifferentiable cost functionals”,SIAM Journal on Control 11 (1973) 637–652.

    Google Scholar 

  2. M. Hled, R.M. Karp and P. Wolfe, “Large-scale optimization and the relaxation method”, Proceedings of the 25th National ACM Meeting, Boston, Mass., August, 1972.

  3. M. Held, P. Wolfe and H.P. Crowder, “Validation of subgradient optimization”,Mathematical Programming 6 (1974) 62–88.

    Google Scholar 

  4. M.R. Hestenes and E. Stiefel, “Methods of conjugate gradients for solving linear systems”,Journal of Research of the National Bureau of Standards 49 (1952) 409–436.

    Google Scholar 

  5. C. Lemarechal, “An algorithm for minimizing convex functions”, in: J.L. Rosenfeld, ed.,Information processing '74 (North-Holland, Amsterdam) pp. 552–556.

  6. C. Lemarechal, “An extension of “Davidson” methods for minimizing nondifferentiable functions”, in: M.L. Balinski and P. Wolfe, eds.,Nondifferentiable optimization, Mathematical Programming Study 3 (North-Holland, Amsterdam), to appear.

  7. R.T. Rockafellar,Convex analysis (Princeton University Press, Princeton, N.J., 1970).

    Google Scholar 

  8. P. Wolfe, “Convergence theory in nonlinear programming”, in: J. Abadie, ed.,Integer and nonlinear programming (North-Holland, Amsterdam, 1970) pp. 1–36.

    Google Scholar 

  9. P. Wolfe, “A method of conjugate subgradients for minimizing nondifferentiable functions”, in: M.L. Balinski and P. Wolfe, eds.,Nondifferentiable optimization, Mathematical Programming Study 3 (North-Holland, Amsterdam), to appear.

  10. P. Wolfe, “Finding the nearest point in a polytope”, RC4887, IBM Research Center, Yorktown Heights, N.Y. (June, 1974).

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Additional information

Editor's note. The communications of Wolfe and Lemarechal which follow — received almost simultaneously — display different points of view, but deal with the same problem and use similar techniques. They are preliminary versions of promising attacks on the problem of minimizing a convex, but not necessarily differentiable, function of many variables. MATHEMATICAL PROGRAMMING STUDY 3 entitledNondifferentiable optimization is to be devoted to this subject.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Wolfe, P. Note on a method of conjugate subgradients for minimizing nondifferentiable functions. Mathematical Programming 7, 380–383 (1974). https://doi.org/10.1007/BF01585533

Download citation

  • Received:

  • Revised:

  • Issue Date:

  • DOI: https://doi.org/10.1007/BF01585533

Keywords

Navigation