4 Lovasz-Erdős Local Lemma

What is the idea fo a local lemma? It’s a general statement in probability about when a certain probability is non-zero.

A silly example: Say I have 100 independent coins. How do I prove that there exists a configuration of these coins such that everything is heads?

Our earlier techniques don’t work on something like this: before we either started with an event that holds with probability, or something that is close to an event that holds with high probability. 100 heads is neither of these things, but obviously it exists because everything is independent.

The Local Lemma allows us to show that the probability of an event is non-zero provided we have “enough” independence, which we make precise below.

Lemma 4.1 (Erdős-Lovasz, 70s). Assuming that:

  • F={Ai} is a family of events in a probability space

  • (Ai)1e(Δ+1), where Δ=Δ(G) is the max degree of G

Then
(iAic)>0.

Definition 4.2 (Dependency graph). If F={Ai} is a family of events, a dependency graph G is a graph with vertex set F, with the property that: for all i, the event Ai is independent from {Aj:AjAi}.

Note that a family of events F can have many possible dependency graphs, but in applications there is often a particularly natural choice to take.

Picture:

PIC

Reminder: Ai is independent from {Aj:AjAi} if E is formed by taking intersections of Aj,AjcAi, we have

(AiE)=(Ai)(E).

Theorem 4.3 (Spencer). R(k)(1+o(1))2ke2k2.

Proof. Let δ>0 and let n=(1δ)2ke2k2. Choose a colouring uniformly at random. For every S[n](k), define the event

AS=S(2) are monochromatic”.

We clearly have (AS)=2k2+1.

Let G be a dependency graph on {AS}S, where we say ASAT if S(2)T(2).

Δ=Δ(G)k2n2k2k4(1n2)nk

We now check:

Δ2k2+12k4(1n2)(enk)k2k2=[(1+o(1))1n2k(en2k2k2)]k.

This 0. So the condition in Local Lemma holds for sufficiently large k.

Remark. This is the best known lower bound for diagonal Ramsey.

Definition 4.4 (k-uniform hypergraph). Say H is a k-uniform hypergraph on vertex set X if HX(k).

Definition 4.5 (Colourable hypergraph). Say a hypergraph H is 2-colourable if there exists χ:X{red,blue} such that there is no monochromatic eH.

PIC

Theorem 4.6. Assuming that:

Then H is 2-colourable.

Note. This theorem has no condition on |H| in the assumptions. Thus, this theorem is not just picking out a high probability event out of all uniformly random colourings: if |H| is large and has a reasonable number of edges, then with high probability there will be a monochromatic edge.

Proof. We choose our colouring of the ground set X uniformly at random. For each eH, define the event

Ae=e monochromatic”.

We have (Ae)=2k+1, so we want 2k+11e(Δ+1). Note if we define our dependency graph G by AeAf if ef, then

Δ(G)dk.

Now just use the bound on d in the hypothesis.

Remark. Actually, our second application implies the first on R(k): just let H be the k2-uniform hypergraph on X=[n](2), defined by

H={S(2):S[n](k)}.

We now move onto proving Local Lemma.

Theorem 4.7 (Lopsided Local Lemma). Assuming that:

  • F={Ai} a family of events

  • x1,,xn(0,1) be such that i,

    (Ai)xiAiAj(1xj)

    (where AiAj denotes adjacency in the dependency graph).

Then
(i=1nAic)i=1n(1xi).

This is slightly more mysterious than the first form, but it can be much more useful. Specifically, if the probabilities (Ai) aren’t all the same, then this lopsided form allows us to tune the values of xi more precisely than if we just use Local Lemma. We’ll see an example of this at the end of this section.

Proof. We use

(AB)=(A|B)(B).

Applying this many times, we have

(i=1nAic)=i=1n(Aic|A1cAi1c)=i=1n(1(Ai|A1cAi1c)xi?)

We will prove that this term is xi by instead proving the more general statement:

()=(Ai | jSAjc)xi,

where S[n]{i}. We prove by induction on |S|.

If S=, then done by hypothesis.

If S, then define

D=jSjiAjc,I=jSjiAjc.

Note that

()=(Ai|DI)=(AiDI)(DI)(AiI)(DI)=(Ai)(I)(DI)=(Ai)(D|I)

Let {Ai1,,Ais} be the events in the intersection D. Then

(D|I)=j=1s(Aijc|IAi1cAij1c)=j=1s(1(Aij|IAi1cAij1c))

If no events in the intersection, then we are already done. Otherwise, we may apply the induction hypothesis to each term to get

(Ai)(D|I)xiji(1xj)j=1s(1xij)xi.

Proof of Lopsided Local Lemma. Apply Lopsided Local Lemma with weights xi=1Δ+1 for all i. We just need to check that the probability upper bounds in Lopsided Local Lemma hold.

By calculus, we have

(ΔΔ+1)Δ1e.

Hence

(Ai)1e(Δ+1)1Δ+1(ΔΔ+1)Δ=1Δ+1(11Δ+1)Δ1Δ+1AiAj(11Δ+1)

Now we will see another proof of R(3,k)ck2(logk)2 (Theorem 3.1), in order to give some intuition as to what kind of situations are good for Local Lemma. Roughly speaking, it works well if your events are “weakly dependent”, or there are “not too many dependencies” among them.

Proof of R(3,k)ck2(logk)2 using Lopsided Local Lemma. Recall that it is enough to show that for sufficiently large n, there exists a triangle-free graph on n vertices with α(G)Cnlogn (this was the statement of Theorem 3.2).

Set k=Cnlogn, and sample GG(n,p) where p=γn, with γ>0 is chosen later.

Define the events:

  • For each T[n](3), define

    AT=T(2)G.
  • For each I[n](k), define

    BI=I independent in G.

Define the dependency graph G:

  • ATAS for S,T[n](3) if S(2)T(2).

  • ATBI for T[n](3),I[n](k) if S(2)I(2).

  • BIBJ for I,J[n](k) if I(2)J(2).

Now note:

  • The event AT is to 3n AS, S[n](3).

  • The event AT is to 3nk2 BI, I[n](k).

  • The event BI is to k2n AS, S[n](3).

  • The event BI is to k2nk2 BJ, J[n](k).

PIC

We choose xT=2p3 for all T[n](3) (“we want to make it a bit bigger than the probability of it occuring so that we can afford the product stuff”) and xI=n10k for all I[n](k).

Then

xTS[n](3)TS(1xS)I[n](k)IT(1xI)2p3(12p3)3n(1n10k)3nk22p3exp((1+o(1))[2p3(3n)+3n10knk2])=2p3(1+o(1))p3

for large enough n. We have (BI)=(1p)k2epk2, so

xI(12p3)k2n(1n10k)k2nk2n10kexp((1+o(1))[2p3k2n+n10kk2nk20])=n10kexp((1+o(1))(2γ2)pk2)

Note

epk2=exp(γnkCn( log n))=nγCk.

First choose γ small so e2γ2pkepk212, and then choose C1γ, then done.