Skip to main content

A Discrepancy Lower Bound for Information Complexity

Author(s): Braverman, Mark; Weinstein, O

Download
To refer to this page use: http://arks.princeton.edu/ark:/88435/pr16x1f
Full metadata record
DC FieldValueLanguage
dc.contributor.authorBraverman, Mark-
dc.contributor.authorWeinstein, O-
dc.date.accessioned2018-07-20T15:09:54Z-
dc.date.available2018-07-20T15:09:54Z-
dc.date.issued2016-11-01en_US
dc.identifier.citationBraverman, M, Weinstein, O. (2016). A Discrepancy Lower Bound for Information Complexity. Algorithmica, 76 (846 - 864. doi:10.1007/s00453-015-0093-8en_US
dc.identifier.urihttp://arks.princeton.edu/ark:/88435/pr16x1f-
dc.description.abstractThis paper provides the first general technique for proving information lower bounds on two-party unbounded-rounds communication problems. We show that the discrepancy lower bound, which applies to randomized communication complexity, also applies to information complexity. More precisely, if the discrepancy of a two-party function f with respect to a distribution μ is Discμf, then any two party randomized protocol computing f must reveal at least Ω (log (1 / Discμf)) bits of information to the participants. As a corollary, we obtain that any two-party protocol for computing a random function on { 0 , 1 } n× { 0 , 1 } n must reveal Ω (n) bits of information to the participants. In addition, we prove that the discrepancy of the Greater-Than function is Ω(1/n), which provides an alternative proof to the recent proof of Viola (Proceedings of the twenty-fourth annual ACM-SIAM symposium on discrete algorithms, SODA 2013, New Orleans, LA, USA, 6–8 Jan 2013, pp 632–651, 2013) of the Ω (log n) lower bound on the communication complexity of this well-studied function and, combined with our main result, proves the tight Ω (log n) lower bound on its information complexity. The proof of our main result develops a new simulation procedure that may be of an independent interest. In a followup breakthrough work of Kerenidis et al. (53rd annual IEEE symposium on foundations of computer science, FOCS 2012, New Brunswick, NJ, USA, 20–23 Oct 2012, pp 500–509, 2012), our simulation procedure served as a building block towards a proof that almost all known lower bound techniques for communication complexity (and not just discrepancy) apply to information complexity as well. © 2015, Springer Science+Business Media New York.en_US
dc.format.extent846 - 864en_US
dc.language.isoen_USen_US
dc.relation.ispartofAlgorithmicaen_US
dc.rightsAuthor's manuscripten_US
dc.titleA Discrepancy Lower Bound for Information Complexityen_US
dc.typeJournal Articleen_US
dc.identifier.doidoi:10.1007/s00453-015-0093-8-
dc.date.eissued2015-11-30en_US
pu.type.symplectichttp://www.symplectic.co.uk/publications/atom-terms/1.0/journal-articleen_US

Files in This Item:
File Description SizeFormat 
A Discrepancy Lower Bound for Information Complexity.pdf217.59 kBAdobe PDFView/Download


Items in OAR@Princeton are protected by copyright, with all rights reserved, unless otherwise indicated.