Subgradient method

This is an old revision of this page, as edited by Kiefer.Wolfowitz (talk | contribs) at 21:25, 13 October 2010 (Convergence results: link Dimitri P. Bertsekas and Shor). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

Subgradient methods are iterative methods for solving convex minimization problems. Originally developed by Naum Z. Shor and others in the 1960s and 1970s, subgradient methods are convergent when applied even to a non-differentiable objective function. When the objective function is differentiable, subgradient methods for unconstrained problems use the same search direction as the method of steepest descent.

Although subgradient methods can be much slower than interior-point methods and Newton's method in practice, they can be immediately applied to a far wider variety of problems and require much less memory. Moreover, by combining the subgradient method with primal or dual decomposition techniques, it is sometimes possible to develop a simple distributed algorithm for a problem.

Basic subgradient update

Let   be a convex function with ___domain  . The subgradient method uses the iteration

 

where   denotes a subgradient of   at  . If   is differentiable, its only subgradient is the gradient vector   itself. It may happen that   is not a descent direction for   at  . We therefore maintain a list   that keeps track of the lowest objective function value found so far, i.e.

 

Step size rules

Many different types of step size rules are used in the subgradient method. Five basic step size rules for which convergence is guaranteed are:

  • Constant step size,  
  • Constant step length,  , which gives  
  • Square summable but not summable step size, i.e. any step sizes satisfying
 
  • Nonsummable diminishing, i.e. any step sizes satisfying
 
  • Nonsummable diminishing step lengths, i.e.  , where
 

Notice that the step sizes listed above are determined before the algorithm is run and do not depend on any data computed during the algorithm. This is very different from the step size rules found in standard descent methods, which depend on the current point and search direction.

Convergence results

For constant step-length and scaled subgradients having Euclidean norm equal to one, the subgradient method converges to an arbitrarily close approximation to the minimum value, that is

  by a result of Shor.[1]

Constrained optimization

Projected subgradient

One extension of the subgradient method is the projected subgradient method, which solves the constrained optimization problem

minimize   subject to
 

where   is a convex set. The projected subgradient method uses the iteration

 

where   is projection on   and   is any subgradient of   at  

General constraints

The subgradient method can be extended to solve the inequality constrained problem

minimize   subject to
 

where   are convex. The algorithm takes the same form as the unconstrained case

 

where   is a step size, and   is a subgradient of the objective or one of the constraint functions at   Take

 

where   denotes the subdifferential of  . If the current point is feasible, the algorithm uses an objective subgradient; if the current point is infeasible, the algorithm chooses a subgradient of any violated constraint.

References

  1. ^ The approximate convergence of the constant step-size (scaled) subgradient method is stated as Exercise 6.3.14(a) in Bertsekas (page 636): Bertsekas, Dimitri P. (1999). Nonlinear Programming (Second ed.). Cambridge, MA.: Athena Scientific. ISBN 1-886529-00-0. On page 636, Bertsekas attributes this result to Shor: Shor, Naum Z. (1985). Minimization Methods for Non-differentiable Functions. Springer-Verlag. ISBN 0-387-12763-1.