TY - JOUR
T1 - Frank–Wolfe-type methods for a class of nonconvex inequality-constrained problems
AU - Zeng, Liaoyuan
AU - Zhang, Yongle
AU - Li, Guoyin
AU - Pong, Ting Kei
AU - Wang, Xiaozhou
N1 - Publisher Copyright:
© Springer-Verlag GmbH Germany, part of Springer Nature and Mathematical Optimization Society 2024.
PY - 2024/11
Y1 - 2024/11
N2 - The Frank–Wolfe (FW) method, which implements efficient linear oracles that minimize linear approximations of the objective function over a fixed compact convex set, has recently received much attention in the optimization and machine learning literature. In this paper, we propose a new FW-type method for minimizing a smooth function over a compact set defined as the level set of a single difference-of-convex function, based on new generalized linear-optimization oracles (LO). We show that these LOs can be computed efficiently with closed-form solutions in some important optimization models that arise in compressed sensing and machine learning. In addition, under a mild strict feasibility condition, we establish the subsequential convergence of our nonconvex FW-type method. Since the feasible region of our generalized LO typically changes from iteration to iteration, our convergence analysis is completely different from those existing works in the literature on FW-type methods that deal with fixed feasible regions among subproblems. Finally, motivated by the away steps for accelerating FW-type methods for convex problems, we further design an away-step oracle to supplement our nonconvex FW-type method, and establish subsequential convergence of this variant. Numerical results on the matrix completion problem with standard datasets are presented to demonstrate the efficiency of the proposed FW-type method and its away-step variant.
AB - The Frank–Wolfe (FW) method, which implements efficient linear oracles that minimize linear approximations of the objective function over a fixed compact convex set, has recently received much attention in the optimization and machine learning literature. In this paper, we propose a new FW-type method for minimizing a smooth function over a compact set defined as the level set of a single difference-of-convex function, based on new generalized linear-optimization oracles (LO). We show that these LOs can be computed efficiently with closed-form solutions in some important optimization models that arise in compressed sensing and machine learning. In addition, under a mild strict feasibility condition, we establish the subsequential convergence of our nonconvex FW-type method. Since the feasible region of our generalized LO typically changes from iteration to iteration, our convergence analysis is completely different from those existing works in the literature on FW-type methods that deal with fixed feasible regions among subproblems. Finally, motivated by the away steps for accelerating FW-type methods for convex problems, we further design an away-step oracle to supplement our nonconvex FW-type method, and establish subsequential convergence of this variant. Numerical results on the matrix completion problem with standard datasets are presented to demonstrate the efficiency of the proposed FW-type method and its away-step variant.
KW - 68Q25
KW - 90C06
KW - 90C26
KW - Away-step oracles
KW - Frank–Wolfe variants
KW - Generalized linear-optimization oracles
KW - Nonconvex constraint sets
UR - https://www.scopus.com/pages/publications/85183736234
U2 - 10.1007/s10107-023-02055-y
DO - 10.1007/s10107-023-02055-y
M3 - Journal article
AN - SCOPUS:85183736234
SN - 0025-5610
VL - 208
SP - 717
EP - 761
JO - Mathematical Programming
JF - Mathematical Programming
IS - 1-2
ER -