A globally convergent linearly constrained Lagrangian methods for nonlinear optimization

M. P. Friedlander and M. A. Saunders
SIAM J. on Optimization 15(3), 863–897, 2005 PDF


For optimization problems with nonlinear constraints, linearly constrained Lagrangian (LCL) methods solve a sequence of subproblems of the form “minimize an augmented Lagrangian function subject to linearized constraints.” Such methods converge rapidly near a solution but may not be reliable from arbitrary starting points. Nevertheless, the well-known software package MINOS has proved effective on many large problems. Its success motivates us to derive a related LCL algorithm that possesses three important properties: it is globally convergent, the subproblem constraints are always feasible, and the subproblems may be solved inexactly. The new algorithm has been implemented in Matlab, with an option to use either MINOS or SNOPT (Fortran codes) to solve the linearly constrained subproblems. Only first derivatives are required. We present numerical results on a subset of the COPS, HS, and CUTE test problems, which include many large examples. The results demonstrate the robustness and efficiency of the stabilized LCL procedure.


  Author = {M. P. Friedlander and M. A. Saunders},
  Journal = {SIAM J. on Optimization},
  Number = 3,
  Pages = {863--897},
  Title = {A globally convergent linearly constrained {L}agrangian
           method for nonlinear optimization},
  Volume = 15,
  Year = 2005,
  Doi = {10.1137/S1052623402419789}