Skip to main content

A sensing policy based on confidence bounds and a restless multi-armed bandit model

Author(s): Oksanen, Jan; Koivunen, Visa; Poor, H Vincent

Download
To refer to this page use: http://arks.princeton.edu/ark:/88435/pr1x77k
Full metadata record
DC FieldValueLanguage
dc.contributor.authorOksanen, Jan-
dc.contributor.authorKoivunen, Visa-
dc.contributor.authorPoor, H Vincent-
dc.date.accessioned2020-02-19T21:59:50Z-
dc.date.available2020-02-19T21:59:50Z-
dc.date.issued2012-11en_US
dc.identifier.citationOksanen, Jan, Visa Koivunen, and H. Vincent Poor. "A sensing policy based on confidence bounds and a restless multi-armed bandit model." In 2012 Conference Record of the Forty Sixth Asilomar Conference on Signals, Systems and Computers (ASILOMAR), (2012): 318-323. doi:10.1109/ACSSC.2012.6489015en_US
dc.identifier.issn1058-6393-
dc.identifier.urihttp://arks.princeton.edu/ark:/88435/pr1x77k-
dc.description.abstractA sensing policy for the restless multi-armed bandit problem with stationary but unknown reward distributions is proposed. The work is presented in the context of cognitive radios in which the bandit problem arises when deciding which parts of the spectrum to sense and exploit. It is shown that the proposed policy attains asymptotically logarithmic weak regret rate when the rewards are bounded independent and identically distributed or finite state Markovian. Simulation results verifying uniformly logarithmic weak regret are also presented. The proposed policy is a centrally coordinated index policy, in which the index of a frequency band is comprised of a sample mean term and a confidence term. The sample mean term promotes spectrum exploitation whereas the confidence term encourages exploration. The confidence term is designed such that the time interval between consecutive sensing instances of any suboptimal band grows exponentially. This exponential growth between suboptimal sensing time instances leads to logarithmically growing weak regret. Simulation results demonstrate that the proposed policy performs better than other similar methods in the literature.en_US
dc.format.extent318 - 323en_US
dc.language.isoen_USen_US
dc.relation.ispartof2012 Conference Record of the Forty Sixth Asilomar Conference on Signals, Systems and Computers (ASILOMAR)en_US
dc.rightsAuthor's manuscripten_US
dc.titleA sensing policy based on confidence bounds and a restless multi-armed bandit modelen_US
dc.typeConference Articleen_US
dc.identifier.doi10.1109/ACSSC.2012.6489015-
dc.date.eissued2013-03-28en_US
dc.identifier.eissn1058-6393-
pu.type.symplectichttp://www.symplectic.co.uk/publications/atom-terms/1.0/journal-articleen_US

Files in This Item:
File Description SizeFormat 
OA_SensingPolicyBasedConfidenceBoundsRestlessMultiArmedBanditModel.pdf522.48 kBAdobe PDFView/Download


Items in OAR@Princeton are protected by copyright, with all rights reserved, unless otherwise indicated.