Bounded regret in stochastic multi-armed bandits
Author(s): Bubeck, Sébastien; Perchet, Vianney; Rigollet, Philippe
DownloadTo refer to this page use:
http://arks.princeton.edu/ark:/88435/pr16j5x
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Bubeck, Sébastien | - |
dc.contributor.author | Perchet, Vianney | - |
dc.contributor.author | Rigollet, Philippe | - |
dc.date.accessioned | 2020-03-02T22:14:56Z | - |
dc.date.available | 2020-03-02T22:14:56Z | - |
dc.date.issued | 2013-02 | en_US |
dc.identifier.citation | Bubeck, Sébastien, Perchet, Vianney, Rigollet, Philippe. (2013). Bounded regret in stochastic multi-armed bandits. https://arxiv.org/abs/1302.1611v2 | en_US |
dc.identifier.uri | https://arxiv.org/abs/1302.1611v2 | - |
dc.identifier.uri | http://arks.princeton.edu/ark:/88435/pr16j5x | - |
dc.description.abstract | We study the stochastic multi-armed bandit problem when one knows the value $\mu^{(\star)}$ of an optimal arm, as a well as a positive lower bound on the smallest positive gap $\Delta$. We propose a new randomized policy that attains a regret {\em uniformly bounded over time} in this setting. We also prove several lower bounds, which show in particular that bounded regret is not possible if one only knows $\Delta$, and bounded regret of order $1/\Delta$ is not possible if one only knows $\mu^{(\star)}$ | en_US |
dc.format.extent | 1 - 14 | en_US |
dc.language.iso | en_US | en_US |
dc.relation.ispartof | Journal of Machine Learning Research | en_US |
dc.rights | Author's manuscript | en_US |
dc.title | Bounded regret in stochastic multi-armed bandits | en_US |
dc.type | Journal Article | en_US |
pu.type.symplectic | http://www.symplectic.co.uk/publications/atom-terms/1.0/journal-article | en_US |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
BoundedRegretMultiArmStochastic.pdf | 188.74 kB | Adobe PDF | View/Download |
Items in OAR@Princeton are protected by copyright, with all rights reserved, unless otherwise indicated.