convex composite optimization
Recently Published Documents


TOTAL DOCUMENTS

23
(FIVE YEARS 2)

H-INDEX

9
(FIVE YEARS 0)

2020 ◽  
Vol 45 (3) ◽  
pp. 1164-1192
Author(s):  
James V. Burke ◽  
Abraham Engle

This work concerns the local convergence theory of Newton and quasi-Newton methods for convex-composite optimization: where one minimizes an objective that can be written as the composition of a convex function with one that is continuiously differentiable. We focus on the case in which the convex function is a potentially infinite-valued piecewise linear-quadratic function. Such problems include nonlinear programming, mini-max optimization, and estimation of nonlinear dynamics with non-Gaussian noise as well as many modern approaches to large-scale data analysis and machine learning. Our approach embeds the optimality conditions for convex-composite optimization problems into a generalized equation. We establish conditions for strong metric subregularity and strong metric regularity of the corresponding set-valued mappings. This allows us to extend classical convergence of Newton and quasi-Newton methods to the broader class of nonfinite valued piecewise linear-quadratic convex-composite optimization problems. In particular, we establish local quadratic convergence of the Newton method under conditions that parallel those in nonlinear programming.


Sign in / Sign up

Export Citation Format

Share Document