This post is based on [1].
For a distribution μ, we use Dϵμ(F) to denote the ϵ-error distributional communication complexity that is the minimum cost of deterministic protocol Π such that
Pμ[Π(X,Y)=F(X,Y)]≥1−ϵ.
On the other hand, the cost if for the best deterministic protocol that gives the correct answer for F on at least (1−ϵ) fraction of all inputs of X×Y given μ.
We say that DISJ(X,Y)=1 if and only if X∩Y=∅. We will prove the following theorem.
Theorem 1
Theorem 1. There exists a distribution μ, such that Dϵμ(DISJn)≥Ω(n) where ϵ<0.01.
Notice that μ is not a product distribution (i.e., μX(x)μY(y)=μ(x,y)) because the common choice of i. Therefore, it does not violate the O~(n) algorithm for DISJ.
Proof. We first give the distribution μ. Let n=4t−1 and let π=(SX,SY,{i}) be an arbitrary partition of [n] such that ∣SX∣=∣SY∣=2t−1 (i∈[n]). Now, let us construct the input for the first player, Alice. We select t elements from SX∪{i} uniformly at random as X. For the player Bob, we select t elements from SY∪{i} uniformly at random as Y. We know that with probability 1/2, we can select the i in the input. For simplicity, we use X0 to denote the input X such that i∈X. We use X1 to denote the input X such that i∈X. Similarly, we define Y0 and Y1.
In the above distribution μ, we have four types of input, i.e., (X1,Y1) (we use I to denote the set of such an input) and other three cases (X0,Y0),(X0,Y1),(X1,Y0) (we use J to denote such inputs) . It is easy to find that μ(J)=3/4.
The following lemma is the key to proving this theorem.
Lemma 1. Given any rectangle R, we have
μ(R∩J)≥αμ(R∩I)−2−δn
where α,δ are some constants.
Suppose Dϵμ=k. Then, protocols will partition {0,1}n×{0,1}n into at most 2k rectangles. Let R1,R2,…,Rℓ denote the monochromatic rectangles with value1 where ℓ=2k. Because we allow ϵ error for protocols, μ(∪i=1ℓRi∩I)≥43−ϵ and μ(∪i=1ℓRi∩J)≤ϵ (Recall that μ(I)=3/4 and μ(J)=1/4. Notice that here J is the set of announcing 0, so the probability of announcing 1 is 0).
To prove Lemma 1, we need to connect pX0(π)pX1(π) and pY0(π)pY1(π) together. Intuitively speaking, they should be close. It is not hard to find the connection.
Step 2> Define Bad Events
To build the connection, we define bad events as follows.
Bad Event. We say a partition π=(SX,SY,{i}) is X-bad if
pX1(π)<31pX0(π)−2−ϵn (1)
Claim 1. Let Bad(π) be the indicator of the event π=(SX,SY,{i}) being Bad. We have
Eπ[pX0(π)pY0(π)Bad(π)]≤54Eπ[pX0(π)pY0(π)] (2)
Combining with the definition of X-bad, we get that
P[i∈s]<31P[i∈s]−2p(π)2−ϵn
So,
P(i∈s)<1/4. (3)
Now let {i1,…,i2t} denote [n]∖SY and let s={s1,…,s2t}. We say that sj=1 if ij∈s.
i). It is easy to see that P(sj)≤1/2 (s has size t).
ii). P(sj)<1/4 from (3).
We prove this cliam by contradiction. Suppose that the bad event happens with probability at least 1/5. So, 1/5 of {s1,…,s2t} is X-bad, which corresponds to ii). For the remaining, it corresponds to i).
[1]. Razborov, Alexander A. “On the distributional complexity of disjointness.” International Colloquium on Automata, Languages, and Programming. Berlin, Heidelberg: Springer Berlin Heidelberg, 1990.
[2]. Kushilevitz, E. and Nisan, N. (1996). Communication Complexity. Cambridge.