Correlation in hard distributions in communication complexity

We study the effect that the amount of correlation in a bipartite distribution has on the communication complexity of a problem under that distribution. We introduce a new family of complexity measures that interpolates between the two previously studied extreme cases: the (standard) randomised comm...

Full description

Saved in:
Bibliographic Details
Main Authors: Klauck, Hartmut, Bottesch, Ralph Christian, Gavinsky, Dmitry
Other Authors: School of Physical and Mathematical Sciences
Format: Article
Language:English
Published: 2018
Subjects:
Online Access:https://hdl.handle.net/10356/87921
http://hdl.handle.net/10220/46886
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-87921
record_format dspace
spelling sg-ntu-dr.10356-879212023-02-28T19:35:24Z Correlation in hard distributions in communication complexity Klauck, Hartmut Bottesch, Ralph Christian Gavinsky, Dmitry School of Physical and Mathematical Sciences Communication Complexity Information Theory DRNTU::Science::Mathematics We study the effect that the amount of correlation in a bipartite distribution has on the communication complexity of a problem under that distribution. We introduce a new family of complexity measures that interpolates between the two previously studied extreme cases: the (standard) randomised communication complexity and the case of distributional complexity under product distributions. - We give a tight characterisation of the randomised complexity of Disjointness under distributions with mutual information k, showing that it is Theta(sqrt(n(k+1))) for all 0 <= k <= n. This smoothly interpolates between the lower bounds of Babai, Frankl and Simon for the product distribution case [k=0], and the bound of Razborov for the randomised case. The upper bounds improve and generalise what was known for product distributions, and imply that any tight bound for Disjointness needs Omega(n) bits of mutual information in the corresponding distribution. - We study the same question in the distributional quantum setting, and show a lower bound of Omega((n(k+1))^{1/4}), and an upper bound (via constructing communication protocols), matching up to a logarithmic factor. - We show that there are total Boolean functions f_d that have distributional communication complexity O(log(n)) under all distributions of information up to o(n), while the (interactive) distributional complexity maximised over all distributions is Theta(log(d)) for n <= d <= 2^{n/100}. This shows, in particular, that the correlation needed to show that a problem is hard can be much larger than the communication complexity of the problem. - We show that in the setting of one-way communication under product distributions, the dependence of communication cost on the allowed error epsilon is multiplicative in log(1/epsilon) - the previous upper bounds had the dependence of more than 1/epsilon. This result, for the first time, explains how one-way communication complexity under product distributions is stronger than PAC-learning: both tasks are characterised by the VC-dimension, but have very different error dependence (learning from examples, it costs more to reduce the error). NRF (Natl Research Foundation, S’pore) MOE (Min. of Education, S’pore) Published version 2018-12-07T08:53:13Z 2019-12-06T16:52:09Z 2018-12-07T08:53:13Z 2019-12-06T16:52:09Z 2015 Journal Article Bottesch, R. C., Gavinsky, D. & Klauck, H. (2015). Correlation in hard distributions in communication complexity. LIPIcs–Leibniz International Proceedings in Informatics, 40, 544-572. doi:10.4230/LIPIcs.APPROX-RANDOM.2015.544 https://hdl.handle.net/10356/87921 http://hdl.handle.net/10220/46886 10.4230/LIPIcs.APPROX-RANDOM.2015.544 en LIPIcs–Leibniz International Proceedings in Informatics © 2015 Ralph Christian Bottesch, Dmitry Gavinsky, and Hartmut Klauck; licensed under Creative Commons License CC-BY. 29 p. application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Communication Complexity
Information Theory
DRNTU::Science::Mathematics
spellingShingle Communication Complexity
Information Theory
DRNTU::Science::Mathematics
Klauck, Hartmut
Bottesch, Ralph Christian
Gavinsky, Dmitry
Correlation in hard distributions in communication complexity
description We study the effect that the amount of correlation in a bipartite distribution has on the communication complexity of a problem under that distribution. We introduce a new family of complexity measures that interpolates between the two previously studied extreme cases: the (standard) randomised communication complexity and the case of distributional complexity under product distributions. - We give a tight characterisation of the randomised complexity of Disjointness under distributions with mutual information k, showing that it is Theta(sqrt(n(k+1))) for all 0 <= k <= n. This smoothly interpolates between the lower bounds of Babai, Frankl and Simon for the product distribution case [k=0], and the bound of Razborov for the randomised case. The upper bounds improve and generalise what was known for product distributions, and imply that any tight bound for Disjointness needs Omega(n) bits of mutual information in the corresponding distribution. - We study the same question in the distributional quantum setting, and show a lower bound of Omega((n(k+1))^{1/4}), and an upper bound (via constructing communication protocols), matching up to a logarithmic factor. - We show that there are total Boolean functions f_d that have distributional communication complexity O(log(n)) under all distributions of information up to o(n), while the (interactive) distributional complexity maximised over all distributions is Theta(log(d)) for n <= d <= 2^{n/100}. This shows, in particular, that the correlation needed to show that a problem is hard can be much larger than the communication complexity of the problem. - We show that in the setting of one-way communication under product distributions, the dependence of communication cost on the allowed error epsilon is multiplicative in log(1/epsilon) - the previous upper bounds had the dependence of more than 1/epsilon. This result, for the first time, explains how one-way communication complexity under product distributions is stronger than PAC-learning: both tasks are characterised by the VC-dimension, but have very different error dependence (learning from examples, it costs more to reduce the error).
author2 School of Physical and Mathematical Sciences
author_facet School of Physical and Mathematical Sciences
Klauck, Hartmut
Bottesch, Ralph Christian
Gavinsky, Dmitry
format Article
author Klauck, Hartmut
Bottesch, Ralph Christian
Gavinsky, Dmitry
author_sort Klauck, Hartmut
title Correlation in hard distributions in communication complexity
title_short Correlation in hard distributions in communication complexity
title_full Correlation in hard distributions in communication complexity
title_fullStr Correlation in hard distributions in communication complexity
title_full_unstemmed Correlation in hard distributions in communication complexity
title_sort correlation in hard distributions in communication complexity
publishDate 2018
url https://hdl.handle.net/10356/87921
http://hdl.handle.net/10220/46886
_version_ 1759855972221714432