Abstract
This paper studies the minimization of a broad class of nonsmooth nonconvex objective functions subject to nonlinear functional equality constraints, where the gradients of the differentiable parts in the objective and the constraints are only locally Lipschitz continuous. We propose a specific proximal linearized alternating direction method of multipliers in which the proximal parameter is generated dynamically, and we design an explicit and tractable backtracking procedure to generate it. We prove subsequent convergence of the method to a critical point of the problem, and global convergence when the problem’s data are semialgebraic. These results are obtained with no dependency on the explicit manner in which the proximal parameter is generated. As a byproduct of our analysis, we also obtain global convergence guarantees for the proximal gradient method with a dynamic proximal parameter under local Lipschitz continuity of the gradient of the smooth part of the nonlinear sum composite minimization model.
Original language | English |
---|---|
Pages (from-to) | 324-353 |
Number of pages | 30 |
Journal | Journal of Optimization Theory and Applications |
Volume | 193 |
Issue number | 1-3 |
DOIs | |
State | Published - Jun 2022 |
Keywords
- Augmented Lagrangian-based methods
- Global convergence
- Kurdyka-Lojasiewicz property
- Nonconvex and nonsmooth minimization
- Proximal gradient method
ASJC Scopus subject areas
- Management Science and Operations Research
- Control and Optimization
- Applied Mathematics