Abstract
This study considers a proximal Newton-type method to solve the minimization of a composite function that is the sum of a smooth nonconvex function and a nonsmooth convex function. In general, the method uses the Hessian matrix of the smooth portion of the objective function or its approximation. The uniformly positive definiteness of the matrix plays an important role in establishing the global convergence of the method. In this study, an inexact proximal memoryless quasi-Newton method is proposed based on the memoryless Broyden family with the modified spectral scaling secant condition. The proposed method inexactly solves the subproblems to calculate scaled proximal mappings. The approximation matrix is shown to retain the uniformly positive definiteness and the search direction is a descent direction. Using these properties, the proposed method is shown to have global convergence for nonconvex objective functions. Furthermore, the R-linear convergence for strongly convex objective functions is proved. Finally, some numerical results are provided.
Similar content being viewed by others
References
Al-Baali, M.: On measure functions for the self-scaling updating formulae for quasi-Newton methods. Optimization 32, 59–69 (1995)
Beck, A.: First-Order Method in Optimization, MOS-SIAM Series on Optimization. SIAM, New Delhi (2017)
Beck, A., Teboulle, M.: A fast iterative shrinkage-thresholding algorithm for linear inverse problems. SIAM J. Imag. Sci. 2, 183–202 (2009)
Becker, S., Candés, E.J., Grant, M.C.: Templates for convex cone problems with applications to sparse signal recovery. Math. Program. Comput. 3, 165–218 (2011)
Becker, S., Candés, E.J., Grant M.C.: TFOCS website. http://cvxr.com/tfocs/download/ (latest access: Jun 5, 2020)
Becker, S., Fadili, J., Ochs, P.: On quasi-newton forward-backward splitting: proximal calculus and convergence. SIAM J. Optim. 29, 2445–2481 (2019)
Bradley, P.S., Mangasarian, O.L.: Massive data discrimination via linear support vector machines. Optim. Methods Softw. 13, 1–10 (2000)
Chang, C.C., Lin, C.J.: LIBSVM: a library for support vector machines. ACM Trans. Intell. Syst. Technol. 2, 1–27 (2011)
Chang, C.C., Lin, C.J.: LIBSVM data: classification, regression, and multi-label. https://www.csie.ntu.edu.tw/~cjlin/libsvmtools/datasets/ (latest access: Jun 5, 2020)
Cheng, W.Y., Li, D.H.: Spectral scaling BFGS method. J. Optim. Theory Appl. 146, 305–319 (2010)
Dolan, E.D., Moré, J.J.: Benchmarking optimization software with performance profiles. Math. Program. 91, 201–213 (2002)
Fukushima, M., Mine, H.: A generalized proximal point algorithm for certain non-convex minimization problems. Int. J. Syst. Sci. 12, 989–1000 (1981)
Gibbons, L.E., Hearn, D.W., Pardalos, P.M., Ramana, M.V.: Continuous characterizations of the maximal clique problem. Math. Oper. Res. 22, 754–768 (1997)
Hoshino, S.: A formulation of variable metric methods. IMA J. Appl. Math. 10, 394–403 (1972)
Koh, K., Kim, S.J., Boyd, S.: An interior-point method for large-scale \(\ell _1\)-regularized logistic regression. J. Mach. Learn. Res. 8, 1519–1555 (2007)
Kou, C.X., Dai, Y.H.: A modified self-scaling memoryless Broyden–Fletcher–Goldfarb–Shanno method for unconstrained optimization. J. Optim. Theory Appl. 165, 209–224 (2015)
Lee, J.D., Sun, Y., Saunders, M.: Proximal Newton-type methods for minimizing composite functions. SIAM J. Optim. 24, 1420–1443 (2014)
Lee, J.D., Sun, Y., Saunders, M.A.: PNOPT website. https://web.stanford.edu/group/SOL/software/pnopt/ (latest access: Jun 5, 2020)
Li, J., Andersen, M.S., Vandenberghe, L.: Inexact proximal Newton methods for self-concordant functions. Math. Methods Oper. Res. 85, 19–41 (2017)
Li, D.H., Fukushima, M.: A modified BFGS method and its global convergence in nonconvex minimization. J. Comput. Appl. Math. 129, 15–35 (2001)
Liu, X., Hsieh, C.J., Lee, J.D., Sun, Y.: An inexact subsampled proximal Newton-type method for large-scale machine learning. arXiv:1708.08552 (2017)
Markowitz, H.: Portfolio selection. J. Financ. 7, 77–91 (1952)
Modarres, F., Hassan, M.A., Leong, W.J.: Memoryless modified symmetric rank-one method for large-scale unconstrained optimization. Am. J. Appl. Sci. 6, 2054–2059 (2009)
Moyi, A.U., Leong, W.J.: A sufficient descent three-term conjugate gradient method via symmetric rank-one update for large-scale optimization. Optimization 65, 121–143 (2016)
Nakayama, S.: A hybrid method of three-term conjugate gradient method and memoryless quasi-Newton method for unconstrained optimization. SUT J. Math. 54, 79–98 (2018)
Nakayama, S., Narushima, Y., Yabe, H.: A memoryless symmetric rank-one method with sufficient descent property for unconstrained optimization. J. Oper. Res. Soc. Jpn. 61, 53–70 (2018)
Nakayama, S., Narushima, Y., Yabe, H.: Memoryless quasi-Newton methods based on spectral-scaling Broyden family for unconstrained optimization. J. Ind. Manag. Optim. 15, 1773–1793 (2019)
Nesterov, Y.: Introductory Lectures on Convex Optimization: A Basic Course, 87. Springer, Berlin (2003)
Scheinberg, K., Tang, X.: Practical inexact proximal quasi-Newton method with global complexity analysis. Math. Program. 160, 495–529 (2016)
Shanno, D.F.: Conjugate gradient methods with inexact searches. Math. Oper. Res. 3, 244–256 (1978)
Shevade, S.K., Keerthi, S.S.: A simple and efficient algorithm for gene selection using sparse logistic regression. Bioinformatics 19, 2246–2253 (2003)
Stella, L., Themelis, A., Patrinos, P.: Forward-backward quasi-Newton methods for nonsmooth optimization problems. Comput. Optim. Appl. 67, 443–487 (2017)
Sun, W., Yuan, Y.: Optimization Theory and Methods: Nonlinear Programming. Springer, Berlin (2006)
Themelis, A., Stella, L., Patrinos, P.: Forward-backward envelope for the sum of two nonconvex functions: further properties and nonmonotone linesearch algorithms. SIAM J. Optim. 28, 2274–2303 (2018)
Themelis, A., Ahookhosh, M., Patrinos, P.: On the acceleration of forward-backward splitting via an inexact Newton method. In: Bauschke, H.H., Burachik, R.S., Luke, D.R. (eds.) Splitting Algorithms, Modern Operator Theory, and Applications, pp. 363–412. Springer, Cham (2019)
Tibshirani, R.: Regression shrinkage and selection via the lasso. J. R. Stat. Soc. Ser. B 58, 267–288 (1996)
Tibshirani, R.: Regression shrinkage and selection via the lasso: a retrospective. J. R. Stat. Soc. Ser. B 73, 273–282 (2011)
Wen, B., Chen, X., Pong, T.K.: Linear convergence of proximal gradient algorithm with extrapolation for a class of nonconvex nonsmooth minimization problems. SIAM J. Optim. 27, 124–145 (2017)
Wright, S.J., Nowak, R.D., Figueiredo, M.A.T.: Sparse reconstruction by separable approximation. IEEE Trans. Signal Process. 57, 2479–2493 (2009)
Yuan, M.M., Lin, Y.: Model selection and estimation in regression with grouped variables. J. R. Stat. Soc. Ser. B 68, 49–67 (2006)
Zhang, Y., Tewarson, P.: Quasi-Newton algorithms with updates from the preconvex part of Broyden’s family. IMA J. Numer. Anal. 8, 487–509 (1988)
Acknowledgements
This research was supported in part by JSPS KAKENHI (Grant Numbers 18K11179, 20K11698 and 20K14986) and the Research Institute for Mathematical Sciences in Kyoto University.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Nakayama, S., Narushima, Y. & Yabe, H. Inexact proximal memoryless quasi-Newton methods based on the Broyden family for minimizing composite functions. Comput Optim Appl 79, 127–154 (2021). https://doi.org/10.1007/s10589-021-00264-9
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10589-021-00264-9