The title of this blog and enumerative combinatorics
by t0rajir0u, Aug 13, 2008, 5:28 AM
The title of this blog is inspired by a quote from Professor Stevens at PROMYS, who likes to say that "mathematicians are annoyingly precise." I really liked this sentiment when I first heard it, especially since PROMYS is where I first learned to really be precise (that is, rigorous) about mathematics.
Professor Stevens also told us to "think deeply of simple things." At PROMYS (as at Ross), that meant thinking very hard about the integers - about addition, about multiplication, about what it meant for an integer to be greater than another integer - but of course this principle applies everywhere. So I'd like to think deeply about this simple thing:
Why does
equal
?
This is such a basic fact (it's certainly an indispensable bit of competition math) that it doesn't really appear to be worth paying attention to. The usual proofs boil down to manipulation or induction. At a later stage, one can also view this identity as the statement of the factorization

which is also basic enough. So why even bother thinking about this identity further?
Maybe an "algebraic" proof just isn't satisfying. Is there a combinatorial way to interpret the above identity? Let me give an example. In proving that

we could just cite the binomial theorem on
and be done with it, or we could recall that the LHS counts the number of subsets of size
of a set of size
for every
, and therefore counts the total number of subsets. The RHS also counts the total number of subsets: for each of the
elements of a set, either that element is or is not in that set. This is perhaps the prototypical example of counting the same thing in two ways, or the method of bijection.
Bijection is the preferred proof technique of enumerative combinatorialists. "Manipulative" techniques - series, in other words - don't really tell you what's going on from this point of view. Let me illustrate this point by pushing the above identity further. We started with the generating function
and substituted
. Differentiating before substituting we obtain the identity
.
(Sums like this are also discussed in Engel.) So what's the combinatorial interpretation here? Well, what we're doing on the LHS is essentially picking a subset and then distinguishing an element of a subset. We can think of this in terms of picking a committee out of
people and then picking a President of that committee. We can also do this by picking the President first, and then picking the committee out of the
people left; that's the RHS. Would you rather elect a President or differentiate?
Differentiating again or picking a committee and electing a President and a Secretary immediately leads to the identity

and
differentiations or picking a committee and electing a board of directors of size
gives

which is a nice lemma in its own right, and summing over all possible values of
we have

because we are splitting up our set into
categories: non-committee, committee but non-director, and director. In terms of generating functions, what we did boils down to computing a Taylor series (or using the binomial theorem twice), but isn't bureaucracy so much simpler than calculus? Or algebra? (See also the similar argument here.)
Now that we've discussed at length the kind of argument we're looking for, what could be counted by
? Well, since we're dividing by
on the RHS, perhaps it's better to multiply out, since multiplying is more natural when counting than division: hence we want to combinatorially interpret
.
The RHS might now remind us of base-
representation, which is good inspiration, but we don't have to refer to it explicitly: let's continue the bureaucracy interpretation. Suppose that we are trying to divide some subset of
people among
committees (so each person has
states; a committee or no committee) such that at least one person is in a committee. On the one hand, there are
ways to do this. On the other hand, given any such division let's order the people and ignore the leftmost people who aren't in a committee (that is, zero digits); say there are
of them. There are
ways to distribute the first person who is in a committee and
ways to distribute the rest. There it is!
Perhaps that was too simple of an answer. Here's another based on the observation that when
is (a power of a) prime,
counts the number of points in the affine space
of dimension
over the finite field
(which means we're implicitly taking
a power of a prime; this turns out not to matter). In other words,
counts the number of points in the affine spaces over
with dimension at most
.
It turns out that we've already encountered a structure that behaves like this: it is the projective space
. Recall that this is defined as the equivalence classes of points in
up to multiplication by a nonzero constant, excluding the origin. In other words, this is the set of lines through the origin. Since we have the advantage of working over finite fields, this set of lines is finite, and so we can count it. But first: over any field, recall that when
we have a projective line, which is an affine line plus a point at infinity. This occured because the equivalence classes of ordered pairs
can be "normalized" to the ordered pairs
when
by appropriate multiplication, but when
we have an extra point
. Generally, a projective space of dimension
contains a copy of an affine space of dimension
(when the last coordinate is not zero) and a projective
-space at infinity (when it is). Symbolically,

in the appropriate sense, for any field
and for all
. Expanding the RHS,
,
so in fact the projective space
contains a copy of each affine space with dimension less than
! (This essentially occurs because we're monitoring how many of the starting components of the vector are zero, which is the same observation we made about people not in committees.) It therefore follows that when
is the finite field
we know that
.
On the other hand, there are
points in
(the affine space we start our construction in) besides the origin, and
nonzero elements of
we can divide by. In other words,
.
Voila!
Of course, we haven't pushed far enough. The combinatorial interpretation in terms of projective space, while pretty, does not provide an obvious generalization to the identity

which we obtain by differentiating with respect to
. Does bureaucracy have more hope? Well, first we have to wonder if we're even asking the right question. Perhaps a different formulation of this question (in which we want to ask how to evaluate sums such as
) would be more susceptible to a counting argument; indeed, the complexity of the RHS suggests that perhaps we want to do some simplification.
Instead of starting with algebra, though, let's just push our combinatorial argument. Suppose we now want to exclude configurations where only one person is in a committee; that is, at least two people must be in a committee. This can clearly be done in
ways (where we are using binomial coefficients to suggest the eventual form of the generalization). On the other hand, in any committee assignment we can consider the positions (from left to right) of the first two people in a committee. For fixed
the second person can be in position
whereas the first person can be in any of the first
positions, leaving
other people. Since the first two people must be in committees, this can be done in
ways. Thus we conclude that

.
This is, in fact, the identity we get if we take
and differentiate with respect to
instead of
, and then plug in
. It generalizes immediately: if we require that at least
people be in a committee (or differentiate
times), we conclude that
.
The polynomial that begins to appear on the RHS is precisely the expansion of
in powers of
. Note that unlike in the other example, where the generating functions approach requires less work, repeated use of the quotient rule gets very tedious on the problem as is. Of course, the answer suggests that the correct approach is to expand the RHS in powers of
, and indeed we have already seen an example of how the binomial theorem and differentiation are closely related.
Back to the binomial theorem: here's another example of a simple problem, but this time I'd like to use it to introduce a more sophisticated technique in the guise of a simple one:
Why does
equal
?
Remember, an answer like
is no longer satisfactory. What this statement means is that there are exactly as many subsets of even size as subsets of odd size, so ideally, what we'd like is a bijection between the two. This is very easy when
is odd: the complement of an even set is odd, and the complement of an odd set is even. But what happens when
is even?
Ideally we'd still like to produce a bijection. Here's an idea: if we add an element to a set, we change its parity! Well, unless that element's already there. But if we remove an element from a set, we change its parity! That's a remarkably simple idea, and it works beautifully. But what may not be obvious is that this strategy and the complement are really the same idea.
The operation we performed above was almost a union - except when our chosen element was in a set, in which case it became the union minus the intersection. The union minus the intersection of two sets is called their symmetric difference, a slightly less intuitive notion than union and intersection, but a very useful one. The complement of a subset is just its symmetric difference with the whole set, so our above two strategies are not really different.
But the story doesn't end here. Let me introduce a convenient notation for subsets of a set with
elements. We will put them in some order and associate with each subset an
-digit binary string, with a
if the
element is contained in that subset and a
otherwise. For example, the subset
is the binary string
. This notation has two huge conveniences, easily verified by comparing truth tables:
- The intersection of two subsets is the componentwise multiplication of their binary strings
.
- The symmetric difference of two subsets is the componentwise addition of their binary strings
.
Several properties I was too lazy to prove now follow immediately: intersection distributes over symmetric difference, symmetric difference is associative, etc. Intersection and symmetric difference together turn the power set of a set into a structure called a Boolean ring. This turns out to characterize all finite Boolean rings.
This is a lovely reformulation of the problem, but we still don't quite have a nice way of expressing "the parity of a subset." The next step is to think of an
-digit binary string as a vector in
(a vector space, to be distinguished from an affine space!), in which case the parity of a subset can be equivalently phrased as the dot product of a vector with
(the vector that represents the entire set) or as the dot product of a vector with itself. It's the first formulation that leads us to the (in hindsight, obvious) conclusion that parity is additive, and therefore symmetric difference is now a completely obvious choice of technique: adding an odd vector (taking the symmetric difference with an odd subset) changes parity.
This has been an extremely "simple" use of linear algebra in combinatorics, but my hope is that it has been simple enough that the ideas are well-motivated. Another simple example can be found here, and an alternate approach to a generalization of the above question can be found here.
The natural extension of the above question is to ask about the number of subsets with size congruent to
. Unfortunately, as we have seen, the power set of a finite set is naturally equipped with a notion of parity but doesn't appear to handle any larger primes. In particular, the vector space
no longer has a simple interpretation in terms of subsets: we knew that
meant "not in this set" and
meant "in this set," but what exactly does
mean?
The standard ("manipulative") technique to answer this question is to apply the roots of unity filter to the coefficients of the polynomial
. For example, the number of subsets with size congruent to
(which we will call
-sets) is the sum of the binomial coefficients
, which is

where
is a primitive third root of unity. (What the above manipulation does is force monomials
to evaluate to zero if
and evaluate to
otherwise, therefore "filtering" out precisely the sum of the coefficients of the terms
.) If
, the above evaluates to
; if
, it's
; if
, it's
; finally, if
it's
. Can we capture this behavior by a purely combinatorial argument, without dealing with roots of unity at all?
First we might as well give the expressions for the number of subsets with size congruent to
(which we will call
- and
-sets respectively). While the above technique is well-known, I believe less people are aware of the following two sums, which can be determined by guesswork, by looking at the coefficients in a DFT matrix, or by using Dirichlet characters:

.
There is another way to derive the above sums which is perhaps more natural in that we are naturally led to the use of third roots of unity, and it generalizes readily to higher moduli. Rename the above sums
. Call a subset of a set with
elements with cardinality congruent to
a
-subset and similarly for
and
. We can determine these sums by working inductively: starting with a set with
elements and adding a new element, a
-subset is either a
-subset without the new element or a
-subset with a new element, and similarly; in other words, we have the system



which, letting
, we can write as the matrix recurrence
.
We dealt with companion matrices earlier in dealing with linear recurrences; this is another special type of matrix known as a circulant matrix, whose eigenvalues turn out to be determined by a discrete Fourier transform. Its eigenvalues are
, which explains the patterns we observed directly (and this observation generalizes), but eigendecomposition is not in the purely combinatorial spirit of this post.
The casework is too tedious to go through a second or third time. What we're basically seeing here is that these three sums are nearly equal except for a tiny error term that behaves in a very annoying way. We should therefore expect this to manifest itself in "near-bijections" between the corresponding three sets.
Actually, half of the time there are just bijections. When
the complement of a
-set is a
-set, when
the complement of a
-set is a
-set, and when
the complement of a
-set is a
-set, so in any case at least two of these sums are equal at a time. Perhaps we should just stick to one case; let's try
. Our algebraic arguments tell us we should have
, which means we should try to find an injection from either
or
to
that misses exactly one set.
... and it is here that I am stuck! I have not been able to construct such an injection, and again I must end a post in disappointment. I will keep thinking about this problem, but in the meantime I suppose this is food for thought.
Practice Problem 1: Recall the combinatorial definition of
as the number of ways to tile a
rectangle using
and
tiles. Give combinatorial arguments for the following identities (not arranged in order of difficulty):
-
-
-
-
-
Practice Problem 2: Let
be a prime and let
be an integer not divisible by
. How many ways are there to arrange
beads of
different colors on a necklace? Conclude that
.
Practice Problem 3: Give a combinatorial interpretation for the equality of exponential generating functions
;
in other words, a combinatorial proof of the Pythagorean theorem.
Professor Stevens also told us to "think deeply of simple things." At PROMYS (as at Ross), that meant thinking very hard about the integers - about addition, about multiplication, about what it meant for an integer to be greater than another integer - but of course this principle applies everywhere. So I'd like to think deeply about this simple thing:
Why does


This is such a basic fact (it's certainly an indispensable bit of competition math) that it doesn't really appear to be worth paying attention to. The usual proofs boil down to manipulation or induction. At a later stage, one can also view this identity as the statement of the factorization

which is also basic enough. So why even bother thinking about this identity further?
Maybe an "algebraic" proof just isn't satisfying. Is there a combinatorial way to interpret the above identity? Let me give an example. In proving that

we could just cite the binomial theorem on





Bijection is the preferred proof technique of enumerative combinatorialists. "Manipulative" techniques - series, in other words - don't really tell you what's going on from this point of view. Let me illustrate this point by pushing the above identity further. We started with the generating function



(Sums like this are also discussed in Engel.) So what's the combinatorial interpretation here? Well, what we're doing on the LHS is essentially picking a subset and then distinguishing an element of a subset. We can think of this in terms of picking a committee out of


Differentiating again or picking a committee and electing a President and a Secretary immediately leads to the identity

and



which is a nice lemma in its own right, and summing over all possible values of


because we are splitting up our set into

Now that we've discussed at length the kind of argument we're looking for, what could be counted by



The RHS might now remind us of base-








Perhaps that was too simple of an answer. Here's another based on the observation that when









It turns out that we've already encountered a structure that behaves like this: it is the projective space












in the appropriate sense, for any field



so in fact the projective space





On the other hand, there are





Voila!
Of course, we haven't pushed far enough. The combinatorial interpretation in terms of projective space, while pretty, does not provide an obvious generalization to the identity

which we obtain by differentiating with respect to


Instead of starting with algebra, though, let's just push our combinatorial argument. Suppose we now want to exclude configurations where only one person is in a committee; that is, at least two people must be in a committee. This can clearly be done in








This is, in fact, the identity we get if we take







The polynomial that begins to appear on the RHS is precisely the expansion of



Back to the binomial theorem: here's another example of a simple problem, but this time I'd like to use it to introduce a more sophisticated technique in the guise of a simple one:
Why does


Remember, an answer like



Ideally we'd still like to produce a bijection. Here's an idea: if we add an element to a set, we change its parity! Well, unless that element's already there. But if we remove an element from a set, we change its parity! That's a remarkably simple idea, and it works beautifully. But what may not be obvious is that this strategy and the complement are really the same idea.
The operation we performed above was almost a union - except when our chosen element was in a set, in which case it became the union minus the intersection. The union minus the intersection of two sets is called their symmetric difference, a slightly less intuitive notion than union and intersection, but a very useful one. The complement of a subset is just its symmetric difference with the whole set, so our above two strategies are not really different.
But the story doesn't end here. Let me introduce a convenient notation for subsets of a set with







- The intersection of two subsets is the componentwise multiplication of their binary strings

- The symmetric difference of two subsets is the componentwise addition of their binary strings

Several properties I was too lazy to prove now follow immediately: intersection distributes over symmetric difference, symmetric difference is associative, etc. Intersection and symmetric difference together turn the power set of a set into a structure called a Boolean ring. This turns out to characterize all finite Boolean rings.
This is a lovely reformulation of the problem, but we still don't quite have a nice way of expressing "the parity of a subset." The next step is to think of an



This has been an extremely "simple" use of linear algebra in combinatorics, but my hope is that it has been simple enough that the ideas are well-motivated. Another simple example can be found here, and an alternate approach to a generalization of the above question can be found here.
The natural extension of the above question is to ask about the number of subsets with size congruent to





The standard ("manipulative") technique to answer this question is to apply the roots of unity filter to the coefficients of the polynomial





where













First we might as well give the expressions for the number of subsets with size congruent to





There is another way to derive the above sums which is perhaps more natural in that we are naturally led to the use of third roots of unity, and it generalizes readily to higher moduli. Rename the above sums













which, letting
![$ \mathbf{S}_n = \left[ \begin{array}{c} S_{0,n} \\
S_{1,n} \\
S_{2,n} \end{array} \right]$](http://latex.artofproblemsolving.com/7/2/f/72fa53b332e28747c8201abfc39463eba73e428e.png)
![$ \mathbf{S}_{n + 1} = \left[ \begin{array}{ccc} 1 & 0 & 1 \\
1 & 1 & 0 \\
0 & 1 & 1 \end{array} \right] \mathbf{S}_{n}$](http://latex.artofproblemsolving.com/9/8/0/980e63c646d4a79ab61eb05831d2c3ca36bed869.png)
We dealt with companion matrices earlier in dealing with linear recurrences; this is another special type of matrix known as a circulant matrix, whose eigenvalues turn out to be determined by a discrete Fourier transform. Its eigenvalues are

The casework is too tedious to go through a second or third time. What we're basically seeing here is that these three sums are nearly equal except for a tiny error term that behaves in a very annoying way. We should therefore expect this to manifest itself in "near-bijections" between the corresponding three sets.
Actually, half of the time there are just bijections. When














... and it is here that I am stuck! I have not been able to construct such an injection, and again I must end a post in disappointment. I will keep thinking about this problem, but in the meantime I suppose this is food for thought.
Practice Problem 1: Recall the combinatorial definition of




-

-

-

-

-

Practice Problem 2: Let






Practice Problem 3: Give a combinatorial interpretation for the equality of exponential generating functions

in other words, a combinatorial proof of the Pythagorean theorem.