6 Analysis on the p-biased cube

Let p[0,1]. We define a measure μp on {1,1}n by picking each xi independently to be 1 with probability q and 1 with probability p where q=1p. This convention looks counterintuitive at first (placing p for the probability of the smaller value), but recall that under the correspondence {1,1}n{0,1}n, we have 10 and 11. So this convention means that we “take the identity most of the time, and take the non-identity with probability p”.

Consider the random variable xi (where xμp). Then

μ=𝔼xi=qp=2q1=12p

and

σ2=𝔼(xiμ)2=q(1μ)2+p(1μ)2=q(2p)2+p(2q)2=4pq,

so σ=2pq.

Let ϕi=xiμσ. We write ϕ(t) for tμσ, so ϕi=ϕ(xi).

Let ϕA=iAϕi. The ϕA will play the role of the xA in the unbiased case.

Fix p. Then given f,g:{1,1}n, we define

f,g=𝔼xμpf(x)g(x)

and

fr=(𝔼xμp|f(x)|r)1r.

Lemma 6.1. ϕA,ϕB=δAB for every A,B[n].

Proof.

ϕA,ϕB=𝔼xμpiAϕi(x)iBϕi(x)=𝔼xμpiABϕi(x)iABϕi(x)2=δAB.

Definition 6.2 (p-biased Fourier coefficient). Let f:{1,1}n. The p-biased Fourier coefficient f^(A) is defined by f^(A)=f,ϕA=𝔼xμpf(x)ϕA(x).

Because the ϕA form an orthonormal basis, it follows that

f,g=Af^(A)g^(A)(Plancherel)f=Af^(A)ϕA(inversion formula)

Definition 6.3. Let f:{1,1}n. Then define Dif by

Dif(x)=σ2(f(xi1)f(xi1)).

Then define Infif to be 𝔼xμpDif(x)2=Dif22.

Now,

DiϕA(x)=σ2(ϕA(xi1)ϕA(xi1)).

If iA then this is 0. Otherwise, it is

σ2ϕA{1}(x)(ϕ(1)ϕ(1))=σ2ϕA{1}(1μσ1μσ)=ϕA{1}(x).

It follows that

Dif=Di(Af^(A)ϕA)=Aif^(A)ϕA{1},

and therefore that

Dif22=Aif^(A)2,

and therefore that

I(f)=iInfif=A|A|f^(A)2.
Noise and stability

Fix p, so that we don’t have to write subscripts everywhere.

Let x{1,1} and let ρ[0,1]. We say that yNρ(x) if with probability ρ yi=xi and with probability 1ρ yi is μp-random, and all yis independent.

If xμp, then

[yi=1]=pq+(1p)q=q.

So yμp. Then we say that x and y are ρ-correlated, and write xρy.

Note: ρ depends on p, but we don’t write ρ,p (i.e. we omit the p for convenience).

Definition 6.4. The p-biased noise operator Tρ is given by the formula

Tρf(x)=𝔼yNρ(x)f(y).

Lemma 6.5. For every A[n], TρϕA=ρ|A|ϕA.

Reminders:

ϕ(t)=tμσμ=qp=2q1=12pσ=2pqϕ(1)=pqϕ(1)=qpϕA(x)=iAϕ(xi)

Proof.

TρϕA(x)=𝔼yTρ(x)iAϕ(yi)=iA𝔼yTρ(x)ϕ(yi)=iA(ρϕ(xi)+(1ρ)𝔼ϕ)=ρ|A|ϕA(x)

Corollary 6.6. Tρf^(A)=ρ|A|f^(A).

Proof. Tρf=Af^(A)TρϕA=Aρ|A|f^(A)ϕA.

Definition 6.7. Let ρ[0,1], f:{1,1}n. Then

Stabρf=f,Tρf=𝔼xρyf(x)f(y).

Remark. By Corollary 6.6, Stabρf=Aρ|A|f^(A)2, as in unbiased case.

The Margulis–Russo formula

We shall be considering more than one value of p.

Convention: Let f:{1,1}n. If we write f(p), then all definitions should be understood to be p-biased.

Lemma 6.8. Let f:n be a multilinear function, and write f also for its restriction to {1,1}n. Then 𝔼xμpf(p)(x)=f(μ¯,μ,,μ).

Note: whenever we write f(p), it is implicit that we are restricting to {1,1}n, since μp is only defined there.

We will give 3 proofs!

Proof 1. Write f=A𝜃AxA. Then 𝔼xμpxA=iA𝔼xμpxi=iA(qp)=μ|A|=xA(μ,μ,,μ). Then by linearity, we’re done.

You might think the above proof is a bit odd, since it uses the xA even though we’re in the p-biased case. I would agree with you!

Proof 2. Write f(p)=Af^(A)ϕA. Then

𝔼xμpϕA=iA𝔼xμpϕi={0A1A==ϕA(μ,μ,,μ)

Proof 3. Induction on n.

𝔼f(p)(x)=𝔼(qf(p)(xn1)+pf(p)(xn1))=𝔼f(p)(xnμ)=f(p)(μ,,μ).

Where the second equality used linearity in the last coordinate, and the last equality is by induction hypothesis.

Theorem 6.9 (The Margulis–Russo formula). Let f be as above. Then

ddμ𝔼f(p)=1σi=1nf(p)^(i).

Remark. Later we will care instead about ddp. But for now it is easier to work with ddμ.

Proof. By Lemma 6.8,

ddμ𝔼f(p)=ddμf(μ,μ,,μ)=i=1nxif(μ,μ,,μ)=i=1n12(f(μi1)f(μi1))(by multilinearity)=1σi=1nDif(μ)=1σi=1n𝔼xμpDif(p)(x)

But Dif=Aif^(A)ϕA{i}, so 𝔼Dif=f^(i). The result follows.

Corollary 6.10. Let f:{1,1}n{1,1} be a monotone Boolean function. Then

ddp[f(p)(x)=1]=1σ2I(f(p)).

Proof. 𝔼f(p)=12[f(p)(x)=1], so

[f(p)(x)=1]=1𝔼f(p)2.

Therefore,

ddp[f(p)(x)=1]=12ddμ𝔼f(p)dμdp=1σi=1nf(p)^(i)(by Theorem 6.9)

From the proof of Theorem 6.9, this is 1σi=1n𝔼Dif(p). Since f is monotone, Dif(p)(x){0,σ}, so this equals

1σ2i=1n𝔼(Dif(p))2=1σ2iDif(p)22=I(f(p)).

Remark. Suppose that p1<p2 are such that [f(p1)(x)=1]=𝜀, [f(p2)=1]=1𝜀. Then by MVT there exists p(p1,p2) such that

ddp[f(p)(x)=1]=12𝜀p2p1.

So by The Margulis–Russo formula, there exists p(p1,p2) such that I(f(p))=σ2(12𝜀p2p1).

So if p2p1 isn’t small, then there exists p such that 𝜀[f(p)=1]1𝜀 and I(f(p)) isn’t too large.

Let f(p):{1,1}. Then for each i[n], Eif is defined by

Eif(x)=qf(xi1)+pf(xi1).

Lemma 6.11. For every f:{1,1}n and every i[n], f=Eif+ϕiDif, and Eif and ϕiDif are orthogonal.

Proof. Reminders:

μ=qp=2q1=12pσ=2pq

So

f(x)Eif(x)={p(f(xi1)f(xi1))xi=1q(f(xi1)f(xi1))xi=1

Also,

ϕi(x)={2pσxi=12qσxi=1

Therefore,

f(x)Eif(x)=σ2ϕi(x)(f(xi1)f(xi1))=ϕiDif(x).

Orthogonality is easy (Eif and Dif don’t depend on xi, and then use the fact that ϕi has average 0).

Lemma 6.12 (p-biased Bonami Lemma). Let f(p):{1,1}n have degree at most k. Then f4Ck2f2, where C=4σ2.

Proof. Induction on n. Let g=Enf, h=Dnf. By orthogonality,

f22=g22+ϕnh22=g22+h22.

Note also that Dnf has degree at most k1.

So

f44=𝔼x(g(x)4+4ϕn(x)g(x)3h(x)+6ϕn(x)2g(x)2h(x)2+4ϕn(x)2g(x)h(x)3+ϕn(x)4h(x)4)g44+6g42h42+4|𝔼ϕ3|g4h43+𝔼ϕ4h44

Using that ϕ(1)=pq, ϕ(1)=qp, we get

𝔼ϕ3=q(pq)32p(qp)32=p2q2pq=2(pq)2pq,

so |𝔼ϕ3|2σ. Also,

𝔼ϕ4=qp2q2+pq2p2=p3+q3pq4σ2.

So

f44C2k(g24+6C1g24h22+8σC32g2h23+4σ2C2h24).

Apply aba2+b22 with a=2g2h2C12, b=4σC1h22. Then

f44C2k(g24+8C1g22h22+12σ2C2h24).

Choose C such that 8C12 and 12σ2C21. C=4σ2 will do.

So f44C2k(g22+h22)2=C2kf24.

Remark. This proof does not recover the p=12 case that we saw before (Bonami’s Lemma): in Lemma 4.1 we proved Lemma 6.12 for p=12 but with C=3, whereas in the above proof we only get C=4 in the case p=12.

Corollary 6.13. Let ρ=σ2. Then for every f:{1,1}n we have Tρf4f2.

Proof.

Tρf4k=0nTρf(=k)4k=nρkf(=k)4k=0nρkCk2f(=k)2=k=0nf(=k)2nf2

By the tensor power trick, the result follows.

Corollary 6.14. For every f:{1,1}n with ρ=σ2, we also have Tρf2f43.

Proof. Identical to uniform case, but with a different ρ.

Remark. As before, this gives us that Stabρ2ff432, i.e. Stabσ24f432.

Theorem 6.15 (p-biased Friedgut junta theorem). Let f(p):{1,1}n{1,1} be a Boolean function and suppose that f(k)221𝜀. Then there exists an m-junta g:{1,1}n with gf222𝜀 and mρ2kI(f)3𝜀2σ2.

Proof. Let τ>0 (to be chosen later) and let

J={i:Infifτ}.

Let ρ=σ24. Then

iJStabρ(Dif)iJDif432=iJDif22σ23=iJ(Infif)32σ1σ1τ12I(f)

Let g=AJ|A|kf^(A)ϕ(A). Then

fg22BJ|B|kf^(B)2+|B|>kf^(B)2.

By hypothesis, the second term is at most 𝜀. But

iJStabρ(Dif)=ρ1B|BJ|ρ|B|f^(B)2ρ1BJ|B|kρkf(B)2.

Therefore,

BJ|B|kf(B)2ρ(k1)σ1τ12I(f).

Set τ=𝜀2σ2ρ2kI(f)2. Then we get the bound.

Remark. As in unbiased case, we always have that f(k)221𝜀 if kI(f)𝜀.