PFR

6 Entropy version of PFR

Definition 6.1
#

η:=1/9.

Throughout this chapter, G=F2n, and X10,X20 are G-valued random variables.

Definition 6.2 τ functional
#

If X1,X2 are two G-valued random variables, then

τ[X1;X2]:=d[X1;X2]+ηd[X10;X1]+ηd[X20;X2].
Lemma 6.3 τ depends only on distribution

If X1,X2 are copies of X1,X2, then τ[X1;X2]=τ[X1;X2].

Proof
Definition 6.4 τ-minimizer
#

A pair of G-valued random variables X1,X2 are said to be a τ-minimizer if one has

τ[X1;X2]τ[X1;X2]

for all G-valued random variables X1,X2.

Proposition 6.5 τ has minimum
#

A pair X1,X2 of τ-minimizers exist.

Proof

6.1 Basic facts about minimizers

In this section we assume that X1,X2 are τ-minimizers. We also write k:=d[X1;X2].

Lemma 6.6 Distance lower bound

For any G-valued random variables X1,X2, one has

d[X1;X2]kη(d[X10;X1]d[X10;X1])η(d[X20;X2]d[X20;X2]).
Proof
Lemma 6.7 Conditional distance lower bound

For any G-valued random variables X1,X2 and random variables Z,W, one has

d[X1|Z;X2|W]kη(d[X10;X1|Z]d[X10;X1])η(d[X20;X2|W]d[X20;X2]).
Proof

6.2 First estimate

We continue the assumptions from the preceding section.

Let X1,X2,X~1,X~2 be independent random variables, with X1,X~1 copies of X1 and X2,X~2 copies of X2. (This is possible thanks to Lemma 3.7.)

We also define the quantity

I1:=I[X1+X2:X~1+X2|X1+X2+X~1+X~2].
Lemma 6.8 Fibring identity for first estimate
#

We have

d[X1+X~2;X2+X~1]+d[X1|X1+X~2;X2|X2+X~1]+I[X1+X2:X~1+X2|X1+X2+X~1+X~2]=2k.
Proof
Lemma 6.9 Lower bound on distances
#

We have

d[X1+X~2;X2+X~1]kη(d[X10;X1+X~2]d[X10;X1])η(d[X20;X2+X~1]d[X20;X2])
Proof
Lemma 6.10 Lower bound on conditional distances
#

We have

d[X1|X1+X~2;X2|X2+X~1]kη(d[X10;X1|X1+X~2]d[X10;X1])η(d[X20;X2|X2+X~1]d[X20;X2]).
Proof
Lemma 6.11 Upper bound on distance differences

We have

d[X10;X1+X~2]d[X10;X1]12k+14H[X2]14H[X1]d[X20;X2+X~1]d[X20;X2]12k+14H[X1]14H[X2],d[X10;X1|X1+X~2]d[X10;X1]12k+14H[X1]14H[X2]d[X20;X2|X2+X~1]d[X20;X2]12k+14H[X2]14H[X1].
Proof
Lemma 6.12 First estimate
#

We have I12ηk.

Proof

One can also extract the following useful inequality from the proof of the above lemma.

Lemma 6.13 Entropy bound on quadruple sum
#

With the same notation, we have

H[X1+X2+X~1+X~2]12H[X1]+12H[X2]+(2+η)kI1.
1

Proof

6.3 Second estimate

We continue the assumptions from the preceding section. We introduce the quantity

I2:=I[X1+X2:X1+X~1|X1+X2+X~1+X~2].
Lemma 6.14 Distance between sums
#

We have

d[X1+X~1;X2+X~2]kη2(d[X1;X1]+d[X2;X2]).
Proof
Lemma 6.15
#

We have

d[X1;X1]+d[X2;X2]2k+2(2ηkI1)1η.
Proof
Lemma 6.16 Second estimate
#

We have

I22ηk+2η(2ηkI1)1η.
Proof

6.4 Endgame

Let X1,X2,X~1,X~2 be as before, and introduce the random variables

U:=X1+X2,V:=X~1+X2,W:=X1+X~1

and

S:=X1+X2+X~1+X~2.
Lemma 6.17 Symmetry identity
#

We have

I(U:W|S)=I(V:W|S).
Proof
Lemma 6.18 Bound on conditional mutual informations

We have

I(U:V|S)+I(V:W|S)+I(W:U|S)6ηk15η1η(2ηkI1).
Proof
Lemma 6.19 Bound on distance increments
#

We have

i=12A{U,V,W}(d[Xi0;A|S]d[Xi0;Xi])(63η)k+3(2ηkI1).
Proof
Lemma 6.20 Key identity
#

We have U+V+W=0.

Proof

For the next two lemmas, let (T1,T2,T3) be a G3-valued random variable such that T1+T2+T3=0 holds identically. Set

δ:=1i<j3I[Ti;Tj].
3

Lemma 6.21 Constructing good variables, I
#

One has

kδ+η(d[X10;T1]d[X10;X1])+η(d[X20;T2]d[X20;X2])+12ηI[T1:T3]+12ηI[T2:T3].

(Note: in the paper, this lemma was phrased in a more intuitive formulation that is basically the contrapositive of the one here. Similarly for the next two lemmas.)

Proof
Lemma 6.22 Constructing good variables, II
#

One has

kδ+η3(δ+i=12j=13(d[Xi0;Tj]d[Xi0;Xi])).
Proof
Theorem 6.23 τ-decrement
#

Let X1,X2 be tau-minimizers. Then d[X1;X2]=0.

Proof

6.5 Conclusion

Theorem 6.24 Entropy version of PFR
#

Let G=F2n, and suppose that X10,X20 are G-valued random variables. Then there is some subgroup HG such that

d[X10;UH]+d[X20;UH]11d[X10;X20],

where UH is uniformly distributed on H. Furthermore, both d[X10;UH] and d[X20;UH] are at most 6d[X10;X20].

Proof

Note: a “stretch goal” for this project would be to obtain a ‘decidable‘ analogue of this result (see the remark at the end of Section 2 for some related discussion).