Contextual bandits with linear Payoff functions
Author(s): Chu, W; Li, L; Reyzin, L; Schapire, Robert E
DownloadTo refer to this page use:
http://arks.princeton.edu/ark:/88435/pr18v6f
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Chu, W | - |
dc.contributor.author | Li, L | - |
dc.contributor.author | Reyzin, L | - |
dc.contributor.author | Schapire, Robert E | - |
dc.date.accessioned | 2021-10-08T19:47:21Z | - |
dc.date.available | 2021-10-08T19:47:21Z | - |
dc.date.issued | 2011-12-01 | en_US |
dc.identifier.citation | Chu, W, Li, L, Reyzin, L, Schapire, RE. (2011). Contextual bandits with linear Payoff functions. Journal of Machine Learning Research, 15 (208 - 214 | en_US |
dc.identifier.issn | 1532-4435 | - |
dc.identifier.uri | http://arks.princeton.edu/ark:/88435/pr18v6f | - |
dc.description.abstract | In this paper we study the contextual bandit problem (also known as the multi-armed bandit problem with expert advice) for linear Payoff functions. For T rounds, K actions, and d dimensional feature vectors, we prove an O(√Td ln 3(KT ln(T)/δ)) regret bound that holds with probability 1-δ for the simplest known (both conceptually and computationally) efficient upper confidence bound algorithm for this problem. We also prove a lower bound of Ω( √Td) for this setting, matching the upper bound up to logarithmic factors. Copyright 2011 by the authors. | en_US |
dc.format.extent | 208 - 214 | en_US |
dc.language.iso | en_US | en_US |
dc.relation.ispartof | Journal of Machine Learning Research | en_US |
dc.rights | Final published version. This is an open access article. | en_US |
dc.title | Contextual bandits with linear Payoff functions | en_US |
dc.type | Conference Article | en_US |
dc.identifier.eissn | 1533-7928 | - |
pu.type.symplectic | http://www.symplectic.co.uk/publications/atom-terms/1.0/conference-proceeding | en_US |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
ContextualBanditsLinearPayoffFunctions.pdf | 1.3 MB | Adobe PDF | View/Download |
Items in OAR@Princeton are protected by copyright, with all rights reserved, unless otherwise indicated.