5 Upper bounds on R(3,k)

Theorem 5.1 (Ajtai-Komlós-Szemerédi, 1980s + Shearer 80s). R(3,k)(1+o(1))k2logk.

Theorem 5.2 (Shearer, 1980s). Assuming that:

  • G a triangle-free graph on n vertices

  • max degree d

Then
α(G)(1+o(1))nd(logd),

where o(1)0 as d.

Remark. Given G on n vertices with maximum degree d, we can use greedy algorithm to show α(G)nd+1: pick a vertex not adjacent to anything picked so far, and chuck away its neighbours. At each iteration, we use up at most d+1 vertices, hence we can run the algorithm for at least nd+1 steps.

Theorem 5.2 beats this simple argument by a factor of logd, under the additional assumption that G is triangle-free.

Remark. If we take GG(n,dn) and modify, for dn, one can show that there exist graphs with no triangles and

α(G)(2+o(1))ndlogd.

So Theorem 5.2 is in some sense sharp.

In fact, understanding what happens between these 1+o(1) and 2+o(1) cases is a big open problem.

Sketch pad:

Fact (Jensen’s inequality). Let φ be a convex function on . Let X be a random variable. Then

𝔼φ(X)φ(𝔼X).

Fact. Let G be a graph with average degree d=2e(G)n. Then

xV(G)yxd(y)=xV(G)d(x)2nd2

Proof.

xV(G)yxd(y)=xyd(y)𝟙xy=yd(y)x𝟙xy=yd(y)2

The inequality follows by Jensen.

Now we will prove Shearer, 1980s (in fact a slight strengthening). We discussed earlier that the proof will involve repeatedly removing vertices and tracking a density increase. Shearer found a very elegant way of tracking this, that makes the proof very clean and magical looking, but the key idea is really the sketch argument given above.

Define

f(d)=dlogdd+1(d1)2.

Note f(d)(0,1) for 1<d, f(d)<0 and f(d)0. We have

f(x)f(x0)+(xx0)f(x0)

by remainder form of Taylor’s theorem.

Theorem 5.3 (Shearer on f(d)). Assuming that:

  • G an n vertex graph

  • triangle-free

  • average degree d

Then
α(G)f(d)n.

Proof. Let vV(G). Define G=GN(v)v. We have

α(G)1+α(G)1+(nd(v)1)f(d),

where

d=average degree of G=2e(G)(nd(v)1)

and

e(G)=e(G)yvd(y)

by triangle-free property.

PIC

Using the earlier working, we have

α(G)1+f(d)(nd(v)1)+(dd)f(d)(nd(v)1)f(d)nf(d)(d(v)+1)+f(d)(2e(G)2e(G)+dd(v)+d)+1f(d)nf(d)(d(v)+1)+f(d)(2e(G)2yvd(y)2e(G)+dd(v)+d)+1

We want

f(d)(d(v)+1)f(d)(2yvd(y)+dd(v)+d)+1.

We claim this holds on average. Average of left hand side is

1nvV(G)f(d)(d(v)+1)=f(d)(d+1).

Average of right hand side is (using f(d)<0),

1nRHS=1+f(d)(2nvyvd(y)+d2+d)1+f(d)(2d2+d2+d)=1+f(d)(dd2)

In fact, f satisfies the differential equation

f(x)(x+1)=f(x)(xx2)+1.

So the inequality holds on average, and thus there exists a good choice of v for the induction.

In this proof, note that this function f is not essential to the proof and is just helpful to make the presentation as short as possible.

We now give another proof of Theorem 5.2, but with some constant c instead of 1+o(1), and this time we’ll assume the maximum degree is bounded rather than the average degree:

Second proof of Theorem 5.2. Choose an independent set I uniformly at random among all independent sets. We’ll show

𝔼|I|cndlogd.

This is an example of what is known as the “hard-core model”. It is called “hard” because we have a hard constraint on every edge (we require that each edge is not present).

Let Xv, for vV(G), be defined by

Xv=d𝟙vI+|N(v)I|.

Note, for fixed I,

vV(G)Xv=vV(G)d𝟙vI+|N(v)I|d|I|+d|I|=2d|I|.

So

v𝔼Xv2d𝔼|I|,

and so enough to show

𝔼Xvclogd,

for vV(G) and some c>0.

PIC

Now fix vV(G) and define

L=N(v){v},F=V(G)vN(v).

We now bound

𝔼XvminJF𝔼(Xv|IF=J).

So now fix some JF. We consider IL where L=LN(J) and N(J) means {y:xJ,xy}. Let l=|L|.

Now observe that IL is uniform over all independent sets in G[L]. So since IL is either {v} or any subset of L{v}, we have

𝔼(Xv|IF=J)=d2l1+1+(l12)(2l12l1+1).

We are happy with this inequality, because if l is large, then the second term is large, and if l is small then the first term is large.

Continuing the above calculation, we have

𝔼(Xv|IF=J)max{d2l1+1,l14}.

Now solve

d2l1+1=l14,

which is 4d=(l1)(2l1+1), so l=clogd. So

𝔼(Xv|IF=J)clogd,

and thus 𝔼Xvclogd as desired.

5.1 Recap of R(3,k) bounds proved in this course

We have shown

ck2(logk)2R(3,k)(1+o(1))k2logk.

Kim showed R(3,k)=Θ(k2logk).

Triangle-free process: Define the random graph process (Gi)i. Given Gi define

Oi={e[n](2):Gi+eK3}.

Now sample ei+1Oi uniformly at random, and define Gi+1=Gi+ei+1.

This method is not what Kim used, but it was used by others later to improve the lower bound.

PIC

Up until very recently, every proof of lower bound on R(3,k) has used a variant of the triangle-free process.

Remark. There is some similarity between the triangle-free process and the first proof of Theorem 5.2 that we saw: both are about a process where each step is uniformly random once we condition on a desired property. So we might expect a related theorem to hold for the triangle-free process, and indeed this is the case.

The Ramsey numbers R(3,k) are still of interest, as there is now interest in determining the right constant.

We now finish this section by using Theorem 5.2 to deduce the upper bound on R(3,k):

Theorem 5.4 (Shearer upper bound on R(3,k)). R(3,k)(1+o(1))k2logk.

Proof. Let n=(1+δ)k2logk. If there is a vertex of degree k in the blue graph then we have an independent set of size k because the blue graph must be triangle-free, otherwise done. So the max degree of blue graph is k.

So apply Shearer, 1980s to get (for large k) that

α(G)(1+o(1))nklogk=(1+o(1))(1+δ)k2(logk)klogkk

(where G is the blue graph).