In a recent breakthrough Hao Huang proved the sensitivity conjecture, that had remained open for past 30 years despite some serious effort from various computer scientists and mathematicians. The proof can be described in a single tweet (if you have the basic background), and it has appeared on the ELI5 page of reddit. The main ideas behind the proof are from spectral graph theory, and one of the starting points (see for example this comment) is the following bound on the independence number of a graph due to Cvetković:
Lemma 1 (Inertia Bound). Let be a graph on
vertices and let
be its eigenvalues. Let
be the number of non-negative eigenvalues of
, and
the number of non-positive eigenvalues (counted with multiplicity). Then
.
The proof of this Lemma is a nice application of a classical result in mathematics known as Cauchy’s interlacing theorem (I have talked about eigenvalue interlacing earlier in this blog), and I leave it to the reader to fill in the details*. This spectral upper bound on the independence number of a graph is relatively less known than Hoffman/Delsarte bound which says . Since many combinatorial problems can be rephrased as finding a nice upper bound on the independence number of some graph, such spectral results can be quite useful, when one has enough structure to say something about the eigenvalues. Until recently, I didn’t see any applications of the inertia bound in extremal combinatorics, whereas Hoffman’s bound is widely used; for example, in proving Erdős-Ko-Rado theorems. Interestingly, the classical Erdős-Ko-Rado theorem can also be proved using the Lemma 1 (as I learned from Hao Huang’s talk at our combinatorics seminar).
One interesting observation made by Huang regarding Lemma 1 (which was already known) is that instead of using the eigenvalues of the adjacency matrix of the graph, we could use any pseudo-adjacency matrix and still arrive at the same conclusion. A pseudo-adjacency matrix of a graph on
vertices is any
real symmetrix matrix
with the property that
whenever the
-th vertex is non-adjacent to the
-th vertex (in an adjacency matrix we also require
for every adjacent pair). If
are the number non-negative/positive eigenvalues of
, then from the proof of Lemma 1 using interlacing we again get
.
The flexibility of using the pseudo-adjacency matrix is crucial to Huang’s breakthrough result. In fact, before the proof of the sensitivity conjecture, Huang (in collaboration with Oleksiy Klurman and Cosmin Pohoata) used the same method to give a spectral proof of another theorem on the hypercube, known as Kleitman’s theorem, which is what I want to discuss in this blog post. In particular, I want to give a slightly simplified proof due to Aditya Potukuchi and Amey Bhangale, that my friend Aditya shared with me a few days after Huang’s sensitivity paper appeared on arXiv as an application of his methods, without realising that a spectral proof of this theorem had already appeared in an earlier paper of Huang. Here is the theorem that we will prove:
Theorem 2 (Kleitman). Let be a set of vectors in
such that the hamming distance between any two vectors in
is at most
. Then
.
(This can be seen as the discrete analog of the isodiametric theorems in real geometry where one wants to find the largest volume of a set of points with a given diameter.)
To use Lemma 1 above, we will naturally consider the graph on
where two vectors are adjacent if the hamming distance between them is at least
, because then any independent set in this graph corresponds to such a set
in the theorem. What are the eigenvalues of the graph
?
The boolean hypercube, , and the graphs defined on it using hamming distance are well studied objects. One way of computing the eigenvalues of
is to look at it as the Cayley graph Cay
, where
is the Hamming ball of radius
, that is, all vectors whose Hamming weight is at most
. From the basic theory of Cayley graphs, it then follows that
has
distinct eigenvalues
with
appearing with multiplicity , where
is the Kravchuk polynomial evaluated at
, defined as
Note that is a polynomial of degree
. For ease of notation, we will skip the superscript from now on and just write
instead. These polynomials have several nice properties, and we will use the following symmetry (or should it be called anti-symmetry?):
which lets us see the evaluation of a degree polynomial at
as the evaluation of a degree
polynomial at
, up-to a factor of
. This fact will come in handy later.
Now if you apply Lemma 1 to using these eigenvalues, you will realise (after some calculation) that it does not give the required bound (what bounds do you get?). That’s where pseudo-adjacency matrices come into picture. We will define a weight function
, such that every vector with
non-zero values is given the weight
, and
for all
. From this weight function we will get the pseudo-adjacency matrix
of
where
if the Hamming distance between the
-th and the
-th vectors is equal to
. Again, from the basic theory of Cayley graphs (in particular, calculating Fourier coefficients), or more directly (Lemma 2.3 here), we get that
has
distinct eigenvalues,
with multiplicity , with
. Now, we just have to choose the function
cleverly so that either
, or
, is at most what we need it to be. Here is when Huang, Klurman and Pohoata make the clever choice of
, for the case of
. For the odd case,
, there is a different choice, but from now on we will only stick to the even case of Theorem 2. From this choice, they get the following three properties,
, for
, for
,
from which the result then follows by an easy analysis. The way they prove these properties is using generating functionology, and it is not immediately clear how the choice of came about. What Aditya and Amey do instead, is start with the properties that one wants from the eigenvalues, and then show the existence of the function
, without ever computing it explicitly!
Here is how this is done. Observe that using the symmetric property of Kravchuk polynomials, we can write
In the second formulation, we can see as
multiplied with a linear combination of polynomials of degrees
, evaluated at
. These polynomials are in fact linearly independent, and hence any polynomial
of degree at most
can be written as a linear combination of them. Therefore, we can choose a nice
, for which the eigenvalues will be
, and then get the weights
‘s automatically.
Pick as the unique monic polynomial whose roots are
, for
to
. WLOG, assume that
(if not, then we can take
). Since
is a polynomial, the sign of its values oscillates between the roots. In particular, this shows that
has the same sign for
, that is,
if
is odd (since
and
have the same sign), and
if
is even, for all of these
‘s. In the former case, we can carefully count the number of negative eigenvalues with multiplicities, and in the latter case we can count the number of positive eigenvalues with multiplicities, giving us Theorem 2.
Note that as far as the signs of ‘s are concerned, this trick of picking the polynomial
explicitly, and then getting the weights, gives us the same three properties as we had earlier where Huang, Klurman and Pohoata made the explicit choice of the weight function. For the case of Theorem 2 when
, we have to do something different, and I leave it to the reader to figure that case out.
For some open questions that can potentially be solved using these methods, check out Section 5 of the Huang-Klurman-Pohoata paper.
(*) Fedor Petrov has given a “low-level” proof that avoids using the interlacing theorem (congrats to Fedya for his new blog!).
For discussions on Huang’s proof of the sensitivity conjecture, see the following blog posts: Terry Tao, Gil Kalai, Scott Aaronson, Ken Reagan.
Pingback: Huangs’ Breakthrough, Cvetković’s Bound, Godsil’s Question, and Sinkovic’s Answer | Ratio Bound – A Combinatorics Blog
Pingback: Amazing: Hao Huang Proved the Sensitivity Conjecture! | Combinatorics and more
“For the case of Theorem 2 when d = 2t + 1, we have to do something different, and I leave it to the reader to figure that case out.”
I am currently at a conference with Anurag and Aditya. When Anurag wrote the above. I thought that Anurag meant the following easy argument. As they were not aware of it, here it is:
Let S be the independent set as above. Let
,
,
,
.
My claim is that
is a subset of
such that all elements have pairwise distance at most
and that the same holds for
.
Elements of
have even weight, so they have even pairwise distance and their minimum distance is at most
. Similarly, elements of
have pairwise distance at most
to each other. Elements in
have even weight and vertices in
have odd weight, so a vertex
in
has an odd distance to a vertex
in
. This distance cannot be
as then
and
have distance
in
. This shows the claim.
The same argument works for
. Hence, we can just apply the bound for
and are done.
Or maybe shorter:
Let
be the set with only even weight vertices. Let
be
. Then the pairwise distances in
are all
. Hence, one can apply the bound in
for even d to
.
Same works for the subset of all even weight vertices. That gives the bound.
Really nice proof! One thing that needs to be verified as well is that
is equal to (or just at least) the size of the set of vectors in
that they correspond to, that is, you do not have an
such that both
and
are elements of
(which does not happen because
can’t have both even and odd weight).