Proximal iterative Gaussian smoothing algorithm for a class of nonsmooth convex minimization problems
Author(s) -
Sanming Liu,
Zhijie Wang,
Chongyang Liu
Publication year - 2015
Publication title -
numerical algebra control and optimization
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.303
H-Index - 20
eISSN - 2155-3289
pISSN - 2155-3297
DOI - 10.3934/naco.2015.5.79
Subject(s) - smoothing , lipschitz continuity , mathematics , differentiable function , convex function , gaussian , constant (computer programming) , mathematical optimization , proximal gradient methods for learning , function (biology) , convex optimization , regular polygon , optimization problem , simple (philosophy) , algorithm , mathematical analysis , subderivative , computer science , geometry , physics , philosophy , statistics , epistemology , quantum mechanics , evolutionary biology , biology , programming language
In this paper, we consider the problem of minimizing a convex objective which is the sum of three parts: a smooth part, a simple non-smooth Lipschitz part, and a simple non-smooth non-Lipschitz part. A novel optimization algorithm is proposed for solving this problem. By making use of the Gaussian smoothing function of the functions occurring in the objective, we smooth the second part to a convex and differentiable function with Lipschitz continuous gradient by using both variable and constant smoothing parameters. The resulting problem is solved via an accelerated proximal-gradient method and this allows us to recover approximately the optimal solutions to the initial optimization problem with a rate of convergence of order $O(\frac{\ln k}{k})$ for variable smoothing and of order $O(\frac{1}{k})$ for constant smoothing.
Accelerating Research
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom
Address
John Eccles HouseRobert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom