Skip to main content

Label optimal regret bounds for online local learning

Author(s): Awasthi, Pranjal; Charikar, Moses S; Lai, Kevin A; Risteski, Andrej

Download
To refer to this page use: http://arks.princeton.edu/ark:/88435/pr17k0k
Full metadata record
DC FieldValueLanguage
dc.contributor.authorAwasthi, Pranjal-
dc.contributor.authorCharikar, Moses S-
dc.contributor.authorLai, Kevin A-
dc.contributor.authorRisteski, Andrej-
dc.date.accessioned2021-10-08T19:44:42Z-
dc.date.available2021-10-08T19:44:42Z-
dc.date.issued2015en_US
dc.identifier.citationAwasthi, Pranjal, Moses Charikar, Kevin A. Lai, and Andrej Risteski. "Label optimal regret bounds for online local learning." In Conference on Learning Theory (2015): pp. 150-166.en_US
dc.identifier.issn2640-3498-
dc.identifier.urihttp://proceedings.mlr.press/v40/Awasthi15a.html-
dc.identifier.urihttp://arks.princeton.edu/ark:/88435/pr17k0k-
dc.description.abstractWe resolve an open question from Christiano (2014b) posed in COLT’14 regarding the optimal dependency of the regret achievable for online local learning on the size of the label set. In this framework, the algorithm is shown a pair of items at each step, chosen from a set of n items. The learner then predicts a label for each item, from a label set of size L and receives a real valued payoff. This is a natural framework which captures many interesting scenarios such as online gambling and online max cut. Christiano (2014a) designed an efficient online learning algorithm for this problem achieving a regret of O(\sqrtnL^3 T), where T is the number of rounds. Information theoretically, one can achieve a regret of O(\sqrtn \log L T). One of the main open questions left in this framework concerns closing the above gap. In this work, we provide a complete answer to the question above via two main results. We show, via a tighter analysis, that the semi-definite programming based algorithm of Christiano (2014a) in fact achieves a regret of O(\sqrtnLT). Second, we show a matching computational lower bound. Namely, we show that a polynomial time algorithm for online local learning with lower regret would imply a polynomial time algorithm for the planted clique problem which is widely believed to be hard. We prove a similar hardness result under a related conjecture concerning planted dense subgraphs that we put forth. Unlike planted clique, the planted dense subgraph problem does not have any known quasi-polynomial time algorithms. Computational lower bounds for online learning are relatively rare, and we hope that the ideas developed in this work will lead to lower bounds for other online learning scenarios as well.en_US
dc.format.extent150 - 166en_US
dc.language.isoen_USen_US
dc.relation.ispartofProceedings of The 28th Conference on Learning Theoryen_US
dc.relation.ispartofseriesProceedings of Machine Learning Research;-
dc.rightsFinal published version. Article is made available in OAR by the publisher's permission or policy.en_US
dc.titleLabel optimal regret bounds for online local learningen_US
dc.typeConference Articleen_US
pu.type.symplectichttp://www.symplectic.co.uk/publications/atom-terms/1.0/conference-proceedingen_US

Files in This Item:
File Description SizeFormat 
LabelOptimalRegretBoundsOnlineLocalLearning.pdf322.76 kBAdobe PDFView/Download


Items in OAR@Princeton are protected by copyright, with all rights reserved, unless otherwise indicated.