6 Entropy in additive combinatorics
We shall need two “simple” results from additive combinatorics due to Imre Ruzsa.
Definition (Sum set / difference set / etc).
Let
be an abelian group and let .
The sumset
is the set .
The difference set
is the set .
We write
for ,
for ,
etc.
Definition (Ruzsa distance).
The Ruzsa distance
is
Lemma 6.1 (Ruzsa triangle inequality).
.
Proof.
This is equivalent to the statement
For each ,
pick ,
such
that .
Define a map
Adding the coordinates of
gives , so we
can calculate
(and ) from
, and hence
can calculate .
So is an
injection. □
Lemma 6.2 (Ruzsa covering lemma).
Assuming that:
-
an abelian group
-
finite subsets of
Then can be
covered by at most
translates of .
Proof.
Let
be a maximal subset of
such that the sets
are disjoint.
Then if ,
there exists
such that .
Then .
So can be
covered by
translates of .
But
|
Let ,
be discrete random variables taking values in an abelian group. What is
when
and
are
independent?
For each ,
.
Writing
and for
and
respectively,
this givesim
where
and .
So, sums of independent random variables
convolutions.
Definition (Entropic Ruzsa distance).
Let
be an abelian group and let ,
be
-valued random variables.
The entropic Ruzsa distance
is
where ,
are independent copies of
and .
Lemma 6.3.
Assuming that:
-
,
are finite subsets of
-
,
are uniformly distributed on ,
respectively
Proof.
Without loss of generality ,
are
indepent. Then
Lemma 6.4.
Assuming that:
Then
|
Proof.
By symmetry we also have
Corollary.
Assuming that:
Then:
|
Corollary 6.5.
Assuming that:
Proof.
Without loss of generality ,
are independent.
Then ,
so
Lemma 6.6.
Assuming that:
Then if and only if
there
is some (finite) subgroup
of such that
and
are uniform
on cosets of .
Proof.
-
If ,
are uniform
on ,
, then
is uniform
on ,
so
So .
-
Suppose that ,
are independent and .
From the first line of the proof of Lemma 6.4, it follows that
. Therefore,
and
are independent.
So for every
and every ,
|
where ,
, i.e. for
all ,
So
is constant on .
In particular, .
By symmetry, .
So
for any .
So for every ,
,
,
so .
So
is the same for every .
Therefore,
for every .
It follows that
So
is a subgroup. Also, ,
so
is a coset of .
,
so
is also a coset of .
□
Recall Lemma 1.16: If ,
then:
Lemma 6.7 (The entropic Ruzsa triangle inequality).
Assuming that:
Proof.
We must show that (assuming without loss of generality that
,
and
are
independent) that
|
i.e. that
|
Since is a function
of and is also
a function of ,
we get using Lemma 1.16 that
|
This is the same as
|
By independence, cancelling common terms and Subadditivity, we get ().
□
Lemma 6.8 (Submodularity for sums).
Assuming that:
Then
|
Proof.
is a
function of and
also a function of .
Therefore (using Lemma 1.16),
|
Hence
|
By independence and cancellation, we get the desired inequality. □
Lemma 6.9.
Assuming that:
Proof.
Let ,
,
be independent
copies of .
Then
(as are all
copies of ).
□
Corollary 6.10.
Assuming that:
Proof.
Conditional Distances
Definition (Conditional distance).
Let
be -valued random
variables (in fact,
and don’t have
to be -valued
for the definition to make sense). Then the conditional distance is
|
The next definition is not completely standard.
Definition (Simultaneous conditional distance).
Let
be
-valued random variables. The
simultaneous conditional distance of
to given
is
|
We say that ,
are conditionally
independent trials of ,
given
if:
-
is distributed like .
-
is distributed like .
-
For each ,
is distributed like ,
-
For each ,
is distributed like .
-
and
are independent.
Then
|
(as can be seen directly from the formula).
Lemma 6.11 (The entropic BSG theorem).
Assuming that:
Then
|
Remark.
The last few terms look like .
But they aren’t equal to it, because
and
aren’t (necessarily) independent!
Proof.
|
where ,
are conditionally
independent trials of ,
given
. Now
calculate
Similarly, is
the same, so
is also the same.
Let and
be conditionally
independent trials of
given .
Then .
By Submodularity,
Finally,
Adding or subtracting as appropriate all these terms gives the required inequality. □