Skip to main content

A Linearly Convergent Variant of the Conditional Gradient Algorithm under Strong Convexity, with Applications to Online and Stochastic Optimization

Author(s): Garber, Dan; Hazan, Elad

Download
To refer to this page use: http://arks.princeton.edu/ark:/88435/pr11k0h
Full metadata record
DC FieldValueLanguage
dc.contributor.authorGarber, Dan-
dc.contributor.authorHazan, Elad-
dc.date.accessioned2021-10-08T19:48:36Z-
dc.date.available2021-10-08T19:48:36Z-
dc.date.issued2016en_US
dc.identifier.citationGarber, Dan, and Elad Hazan. "A linearly convergent variant of the conditional gradient algorithm under strong convexity, with applications to online and stochastic optimization." SIAM Journal on Optimization 26, no. 3 (2016): 1493-1528. doi:10.1137/140985366en_US
dc.identifier.issn1052-6234-
dc.identifier.urihttps://arxiv.org/pdf/1301.4666.pdf-
dc.identifier.urihttp://arks.princeton.edu/ark:/88435/pr11k0h-
dc.description.abstractLinear optimization is many times algorithmically simpler than nonlinear convex optimization. Linear optimization over matroid polytopes, matching polytopes, and path polytopes are examples of problems for which we have simple and efficient combinatorial algorithms but whose nonlinear convex counterpart is harder and admits significantly less efficient algorithms. This motivates the computational model of convex optimization, including the offline, online, and stochastic settings, using a linear optimization oracle. In this computational model we give several new results that improve on the previous state of the art. Our main result is a novel conditional gradient algorithm for smooth and strongly convex optimization over polyhedral sets that performs only a single linear optimization step over the domain on each iteration and enjoys a linear convergence rate. This gives an exponential improvement in convergence rate over previous results. Based on this new conditional gradient algorithm we give the first algorithms for online convex optimization over polyhedral sets that perform only a single linear optimization step over the domain while having optimal regret guarantees, answering an open question of Kalai and Vempala and of Hazan and Kale. Our online algorithms also imply conditional gradient algorithms for nonsmooth and stochastic convex optimization with the same convergence rates as projected (sub)gradient methods.en_US
dc.format.extent1493 - 1528en_US
dc.language.isoen_USen_US
dc.relation.ispartofSIAM Journal on Optimizationen_US
dc.rightsAuthor's manuscripten_US
dc.titleA Linearly Convergent Variant of the Conditional Gradient Algorithm under Strong Convexity, with Applications to Online and Stochastic Optimizationen_US
dc.typeJournal Articleen_US
dc.identifier.doi10.1137/140985366-
dc.identifier.doi1095-7189-
pu.type.symplectichttp://www.symplectic.co.uk/publications/atom-terms/1.0/journal-articleen_US

Files in This Item:
File Description SizeFormat 
LinearlyConvConditionalGradient.pdf394.07 kBAdobe PDFView/Download


Items in OAR@Princeton are protected by copyright, with all rights reserved, unless otherwise indicated.