On Graduated Optimization for Stochastic Non-Convex Problems
Author(s): Hazan, Elad; Levy, Kfir Y; Shalev-Shwartz, Shai
DownloadTo refer to this page use:
http://arks.princeton.edu/ark:/88435/pr1nn96
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Hazan, Elad | - |
dc.contributor.author | Levy, Kfir Y | - |
dc.contributor.author | Shalev-Shwartz, Shai | - |
dc.date.accessioned | 2021-10-08T19:49:35Z | - |
dc.date.available | 2021-10-08T19:49:35Z | - |
dc.date.issued | 2016 | en_US |
dc.identifier.citation | Hazan, Elad, Kfir Yehuda Levy, and Shai Shalev-Shwartz. "On Graduated Optimization for Stochastic Non-Convex Problems." In Proceedings of The 33rd International Conference on Machine Learning (2016): pp. 1833-1841. | en_US |
dc.identifier.issn | 2640-3498 | - |
dc.identifier.uri | http://proceedings.mlr.press/v48/hazanb16.pdf | - |
dc.identifier.uri | http://arks.princeton.edu/ark:/88435/pr1nn96 | - |
dc.description.abstract | The graduated optimization approach, also known as the continuation method, is a popular heuristic to solving non-convex problems that has received renewed interest over the last decade.Despite being popular, very little is known in terms of its theoretical convergence analysis. In this paper we describe a new first-order algorithm based on graduated optimization and analyze its performance. We characterize a family of non-convex functions for which this algorithm provably converges to a global optimum. In particular, we prove that the algorithm converges to an ε-approximate solution within O(1 / ε^2) gradient-based steps. We extend our algorithm and analysis to the setting of stochastic non-convex optimization with noisy gradient feedback, attaining the same convergence rate. Additionally, we discuss the setting of “zero-order optimization", and devise a variant of our algorithm which converges at rate of O(d^2/ ε^4). | en_US |
dc.format.extent | 1833 - 1841 | en_US |
dc.language.iso | en_US | en_US |
dc.relation.ispartof | Proceedings of The 33rd International Conference on Machine Learning | en_US |
dc.rights | Final published version. Article is made available in OAR by the publisher's permission or policy. | en_US |
dc.title | On Graduated Optimization for Stochastic Non-Convex Problems | en_US |
dc.type | Conference Article | en_US |
pu.type.symplectic | http://www.symplectic.co.uk/publications/atom-terms/1.0/conference-proceeding | en_US |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
GradOptStochasticProblems.pdf | 545.72 kB | Adobe PDF | View/Download |
Items in OAR@Princeton are protected by copyright, with all rights reserved, unless otherwise indicated.