TY - JOUR
T1 - Forward–backward quasi-Newton methods for nonsmooth optimization problems
AU - Stella, Lorenzo
AU - Themelis, Andreas
AU - Patrinos, Panagiotis
N1 - Funding Information:
This work was supported by the KU Leuven Research Council under BOF/STG-15-043.
Publisher Copyright:
© 2017, Springer Science+Business Media New York.
PY - 2017/7/1
Y1 - 2017/7/1
N2 - The forward–backward splitting method (FBS) for minimizing a nonsmooth composite function can be interpreted as a (variable-metric) gradient method over a continuously differentiable function which we call forward–backward envelope (FBE). This allows to extend algorithms for smooth unconstrained optimization and apply them to nonsmooth (possibly constrained) problems. Since the FBE can be computed by simply evaluating forward–backward steps, the resulting methods rely on a similar black-box oracle as FBS. We propose an algorithmic scheme that enjoys the same global convergence properties of FBS when the problem is convex, or when the objective function possesses the Kurdyka–Łojasiewicz property at its critical points. Moreover, when using quasi-Newton directions the proposed method achieves superlinear convergence provided that usual second-order sufficiency conditions on the FBE hold at the limit point of the generated sequence. Such conditions translate into milder requirements on the original function involving generalized second-order differentiability. We show that BFGS fits our framework and that the limited-memory variant L-BFGS is well suited for large-scale problems, greatly outperforming FBS or its accelerated version in practice, as well as ADMM and other problem-specific solvers. The analysis of superlinear convergence is based on an extension of the Dennis and Moré theorem for the proposed algorithmic scheme.
AB - The forward–backward splitting method (FBS) for minimizing a nonsmooth composite function can be interpreted as a (variable-metric) gradient method over a continuously differentiable function which we call forward–backward envelope (FBE). This allows to extend algorithms for smooth unconstrained optimization and apply them to nonsmooth (possibly constrained) problems. Since the FBE can be computed by simply evaluating forward–backward steps, the resulting methods rely on a similar black-box oracle as FBS. We propose an algorithmic scheme that enjoys the same global convergence properties of FBS when the problem is convex, or when the objective function possesses the Kurdyka–Łojasiewicz property at its critical points. Moreover, when using quasi-Newton directions the proposed method achieves superlinear convergence provided that usual second-order sufficiency conditions on the FBE hold at the limit point of the generated sequence. Such conditions translate into milder requirements on the original function involving generalized second-order differentiability. We show that BFGS fits our framework and that the limited-memory variant L-BFGS is well suited for large-scale problems, greatly outperforming FBS or its accelerated version in practice, as well as ADMM and other problem-specific solvers. The analysis of superlinear convergence is based on an extension of the Dennis and Moré theorem for the proposed algorithmic scheme.
UR - http://www.scopus.com/inward/record.url?scp=85017207558&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85017207558&partnerID=8YFLogxK
U2 - 10.1007/s10589-017-9912-y
DO - 10.1007/s10589-017-9912-y
M3 - Article
AN - SCOPUS:85017207558
VL - 67
SP - 443
EP - 487
JO - Computational Optimization and Applications
JF - Computational Optimization and Applications
SN - 0926-6003
IS - 3
ER -