Академический Документы
Профессиональный Документы
Культура Документы
S. E. Payne
0.1 Notation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
0.2 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
3
4 CONTENTS
3 Polya Theory 89
3.1 Group Actions . . . . . . . . . . . . . . . . . . . . . . . . . . . 89
3.2 Applications . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92
3.3 The Cycle Index: Polya’s Theorem . . . . . . . . . . . . . . . 96
3.4 Sylow Theory Via Group Actions . . . . . . . . . . . . . . . . 98
3.5 Patterns and Weights . . . . . . . . . . . . . . . . . . . . . . . 100
3.6 The Symmetric Group . . . . . . . . . . . . . . . . . . . . . . 106
3.7 Counting Graphs . . . . . . . . . . . . . . . . . . . . . . . . . 110
3.8 Solutions and/or Hints to Selected Exercises . . . . . . . . . . 111
0.1 Notation
Throughout these notes the following notation will be used.
bxc = The floor of x (i.e., the largest integer not larger than x)
P([n]) = {A : A ⊆ [n]}
(j)q = 1 + q + q 2 + · · · + q j−1
0.2 Introduction
The course at CU-Denver for which these notes were assembled, Math 6409
(Applied Combinatorics), deals more or less entirely with enumerative com-
binatorics. Other courses deal with combinatorial structures such as Latin
squares, designs of many types, finite geometries, etc. This course is a
one semester course, but as it has been taught different ways in different
semesters, the notes have grown to contain more than we are now able to
cover in one semester. On the other hand, these notes contain considerably
less material than the standard textbooks listed below. It is always difficult
to decide what to leave out, and the choices clearly are a reflection of the
likes and dislikes of the author. We have tried to include some truly tradi-
tional material and some truly nontrivial material, albeit with a treatment
that makes it accessible to the student.
Since the greater part of this course is, ultimately, devoted to developing
ever more sophisticated methods of counting, we begin with a brief discussion
of what it means to count something. As a first example, for n ∈ N , put
f (n) = |P([n])|. Then no one will argue that the formula f (n) = 2n is any-
thing but nice. As a second example, let d(n) be the number of derangements
i
of (1, . . . , n). Then (as we show at least twice later on) d(n) = n! ni=0 (−1)
P
i!
.
This is not so nice an answer as the first one, but there are very clear proofs.
Also, d(n) is the nearest integer to n!e . This is a convenient answer, but it
lacks combinatorial significance. Finally, let f (n) be the number of n × n
matrices of 0’s and 1’s such that each row and column has 3 1’s. It has been
shown that
X (−1)β (n!)2 (β + 3γ)!2α 3β
f (n) = 6−n ,
α!β!(γ!)2 6γ
where the sum is over all α, β, γ ∈ N for which α + β + γ = n. As far as we
know, this formula is not good for much of anything, but it is a very specific
answer that can be evaluated by computer for relatively small n.
As a different kind of example, suppose we want the Fibonacci numbers
F0 , F1 , F2 , . . . , and what we know about them is that they satisfy the recur-
rence relation
Fn+1 = Fn + Fn−1 (n ≥ 1; F0 = F1 = 1).
The sequence begins with 1, 1, 2, 3, 5, 8, 21, 34, 55, 89, . . .. There are ex-
act, not very complicated formulas for Fn , as we shall see later. But just to
0.2. INTRODUCTION 9
is difficult reading but gives many wonderful insights into the more advanced
aspects of the subject.
Chapter 1
(i) f is arbitrary
(iii) f is surjective
Case 1. Both the balls and the boxes are labeled (or distinguishable).
11
12 CHAPTER 1. BASIC COUNTING TECHNIQUES
f:
1
2
3
a b c d
g:
1
3
2
a b c d
h:
1
2
3
a b c d
i:
2
3
1
a b c d
Case 2. Balls unlabeled; boxes labeled.
f ∼ g:
a b c d
h:
a b c d
i:
a b c d
Case 3. Balls labeled; boxes unlabeled.
f ∼ h:
1
2
3
g: 1 3 2
i: 2 3 1
For the four different possibilities arising according as N and X are each
labeled or unlabeled there are different definitions describing when two func-
tions from N to X are equivalent.
Definition: Two functions f , g : N → X are equivalent
1. with N unlabeled provided there is a bijection π : N → N such that
f (π(a)) = g(a) for all a ∈ N . (In words: provided some relabeling of the
elements of N turns f into g.)
2. with X unlabeled provided there is a bijection σ : X → X such that
σ(f (a)) = g(a) for all a ∈ N . (In words: provided some relabeling of the
elements of X turns f into g.)
1.1. SETS AND FUNCTIONS: THE TWELVEFOLD WAY 13
f
N - X
π σ
?
g
N - X
As we develop notations, methods and results, the reader should fill in the
blanks in the following table with the number of functions of the indicated
type. Of course, sometimes the formula will be elegant and simple. Other
times it may be rather implicit, e.g., the coefficient of some term of a given
power series.
14 CHAPTER 1. BASIC COUNTING TECHNIQUES
Unlabeled Labeled 4 5 6
Labeled Unlabeled 7 8 9
Unlabeled Unlabeled 10 11 12
Let N = {a1 , . . . , an }, X = {0, 1}. Let P(N ) be the set of all subsets of
N . For A ⊆ N , define fA : N → X by
(
1, ai ∈ A
fA (ai ) =
0, ai ∈
6 A.
Then F : P(N ) → X N : A 7→ fA is a bijection, so
Exercise: 1.1.1 Generalize the result of Eq. 1.1 to provide an answer for
the first blank in the twelvefold way.
! !
N n
We define to be the set of all k-subsets of N , and put =
k k
!
N
# . Let N (n, k) be the number of ways to choose a k-subset T of N
k
!
n
and then linearly order the elements of T . Clearly N (n, k) = k!. On
k
the other hand, we could choose any element of N to be the first element
1.1. SETS AND FUNCTIONS: THE TWELVEFOLD WAY 15
nk
!
n n(n − 1) · · · (n − k + 1)
= = (1.2)
k k! k!
!
(n)k
= according to some authors .
k!
n n n+1
Exercise: 1.1.2 Prove: k−1
+ k
= k
.
!
n k
Note: := nk! makes sense for k ∈ N and n ∈ C. What if n ∈ Z,
k
k ∈ Z and k < 0 or k > n? The best thing to do is to define the value of
the binomial coefficient to be zero in these cases. Then we may write the
following
n
! !
n n k n
xn−k ,
X X
(1 + x) = x = (1.4)
k=0
k k
k
where in the second summation the index may be allowed to run over all
integers since the coefficient on xn−k is nonzero for at most a finite number
of values of k.
Put x = −1 in Eq. 1.4 to obtain
!
X
k n
(−1) = 0. (1.5)
k
k
16 CHAPTER 1. BASIC COUNTING TECHNIQUES
Pr j r+1
Exercise: 1.1.4 Prove that j=m m
= m+1
.
Exercise: 1.1.7
Evaluate the following two sums:
a. ni=1 ni i3i
P
Pn n
b. i=2 i
i(i − 1)mi
1.2. COMPOSITION OF POSITIVE INTEGERS 17
· | · · | · · · | · | · · ↔ 1 + 2 + 3 + 1 + 2 = 9.
Exercise: 1.2.3 a) Show that nk=1 m+k−1 = m n+k−1
P P
k k=1 k
. b) Derive a
closed form formula for the number of weak compositions of n into at most
m parts.
1.3. MULTISETS 19
Exercise: 1.2.4 Let S be a set of n elements. Count the ordered pairs (A, B)
of subsets of S such that ∅ ⊆ A ⊆ B ⊆ S. Let c(j, k) denote the number of
such ordered pairs for which |A| = j and |B| = k. Show that:
0≤j≤k≤n
xk
!
x
fk (x) = =
k k!
is a polynomial in x with rational coefficients (not all of which are integers)
and such that for each integer m (positive, negative or zero) fk (m) is also an
integer.
1.3 Multisets
A finite multiset M on a set S is a function ν : S → N such that x∈S ν(x) <
P
ν(xi )
(1 + x1 + x21 + · · ·)(1 + x2 + x22 + · · ·) · · · (1 + xn + x2n + · · ·) =
X Y
( xi ).
ν:S→N xi ∈S
!!
2 n
P
ν(xi ) #M n
xk .
X X X
(1 + x + x + · · ·) = x = x =
ν:S→N M on S k≥0
k
!
n
Theorem 1.4.1 The coefficient of xa11 xa22 · · · xamm in (x1 +· · ·+xm )n is .
a1 , . . . , a m
1.5 Permutations
There are several ways to approach the study of permutations. One of the
most basic is as a group of bijections. Let A be any nonempty set (finite or
infinite). Put S(A) = {π : A → A : π is a bijection}.
Notation: If π : a1 7→ a2 we write π(a1 ) = a2 (unless noted otherwise). If
π, σ ∈ S(A), define the composition π ◦ σ by (π ◦ σ)(a) = π(σ(a)).
Pn Pn−1
= k=1 b(n − 1, k − 1)xk + (n − 1) k=0 b(n − 1, k)xk
Pn
= k=0 [b(n − 1, k − 1) + (n − 1)b(n − 1, k)]xk .
24 CHAPTER 1. BASIC COUNTING TECHNIQUES
" #
Pn n
Corollary 1.5.6 xn = k=0 (−1)
n−k
· xk .
k
n!
.
1c1 c 1 !2c2 c cn
2 ! · · · n cn !
Insert parentheses in π so that the first c1 cycles have length 1, the next
c2 cycles have length 2, . . . , etc. This defines a map Φ : S([n]) → Sc ([n]),
where Sc ([n]) = {σ ∈ S([n]) : σ has type (c1 , . . . , cn )}. Clearly Φ is onto
Sc ([n]). We claim that if σ ∈ Sc ([n]), then the number of π mapped to σ
is 1c1 c1 !2c2 c2 ! · · · ncn cn !. This follows because in writing σ as a product of
disjoint cycles, we can order the cycles of length i (among themselves) in ci !
ways, and then choose the first elements of all these cycles in ici ways. These
choices for different i are all independent. So Φ : S([n]) → Sc ([n]) is a many
to one map onto Sc ([n]) mapping the same number of π to each σ. Since
#S([n]) = n!, we obtain the desired result.
b) λ1 ≥ · · · ≥ λk ≥ 0.
Two partitions of n are identical if they differ only in the number of
terminal 0’s. For example, (3, 3, 2, 1) ≡ (3, 3, 2, 1, 0, 0). The nonzero λi are
the parts of the partition λ. If λ = (λ1 , . . . , λk ) with λ1 ≥ · · · ≥ λk > 0,
we say that λ has k parts. If λ has αi parts equal to i, we may write
λ =< 1α1 , 2α2 , . . . > where terms with αi = 0 may be omitted, and the
superscript αi = 1 may be omitted.
Notation: “λ ` n” means λ is a partition of n. As an example we have
Put p(n) equal to the total number of partitions of n, and pk (n) equal to
the number of partitions of n with k parts.
Convention: p(0) = p0 (0) = 1.
pn (n) = 1.
pn−1 (n) = 1 if n > 1.
p1 (n) = 1 for n ≥ 1.
p2 (n) = bn/2c.
A great deal of time and effort has been spent studying the partitions of
n and much is known about them. However, most of the results concerning
the numbers pk (n) have been obtained via the use of generating functions.
Hence after we have studied formal power series it would be reasonable to
return to the topic of partitions. Unfortunately we probably will not have
time to do this, so this topic would be a great one for a term project.
(b) Bi ∩ Bj = ∅ if i 6= j;
(c) B1 ∪ B2 ∪ · · · ∪ Bk = N.
We
( call) Bi a block of π and say that π has k = |π| = #π blocks. Put
n
= S(n, k) = the number of partitions of an n-set into k-blocks.
k
S(n, k) is called a Stirling number of the second kind.
We immediately have the following list of Stirling numbers.
( )
0
= 1;
(
0 )
n
= 0 if k > n ≥ 1;
(
k )
n
= 0 if n > 0;
(
0 )
n
= 1;
(
1 )
n
= 2n−1 − 1;
(
2 )
n
= 1;
(
n ) !
n n
= .
n−1 2
( ) ( ) ( )
n n−1 n−1
Theorem 1.7.1 =k + .
k k k−1
Proof: To obtain a partition of [n] into k blocks, we can either
( (i) partition
) [n − 1] into k blocks and place n into any of these blocks in
n−1
k ways, or
k
( (ii) put) n into a block by itself and partition [n − 1] into k − 1 blocks in
n−1
ways.
k−1
Bell Number Let B(n) be the total number of partitions of an n-
set. Hence
1.8. TABLE ENTRIES IN THE TWELVEFOLD WAY 27
n n
( ) ( )
X n X n
B(n) = = ,
k=1
k k=0
k
for all n ≥ 1.
Theorem 1.7.2 ( )
n n
xk ,
X
x = n ∈ N.
k
k
( )
n
Corollary 1.7.3 xn = n−k
xk̄ .
P
k (−1)
k
1.9 Recapitulation
We have already
! established
! the following.
!
n n−1 n−1
1. = + “n choose k 00
k k k−1
" # " # " #
n n−1 n−1
2. = (n − 1) + “n cycle k 00
k k k−1
( ) ( ) ( )
n n−1 n−1
3. =k + “n subset k 00
k k k−1
" #
n̄ n k
xn = (−1)n (−x)n̄
P
4. x = k x
k
" #
n n−k n k
xn̄ = (−1)n (−x)n
P
5. x = k (−1) x
k
( )
n n
(−1)n−k xk̄
P
6. x = k
k
( )
n n
xk
P
7. x = k
k
It appears that 4. and 6. (resp., 5. and 7.) are some kind of inverses
of each other. Later we shall make this a little more formal as we study the
incidence algebra of a finite POSET.
Also in this section we want to recap certain results on composition of
integers, etc.
!
r r!
1. P (r; r1 , r2 , . . . , rn ) = =
r1 , r2 , . . . , rn r1 !r2 ! · · · rn !
= the number of ways to split up r people into n labeled committees with
ri people in committee Ci
= the number of words of length r with ri letters of type i, 1 ≤ i ≤ n,
where r1 + r2 + · · · + rn = r
30 CHAPTER 1. BASIC COUNTING TECHNIQUES
n
!
r− ai + n − 1
P
X
= |{(y1 , . . . , yn ) : yi = r and yi ≥ ai for 1 ≤ i ≤ n}|.
r − ai
P
i=0
(b) G is acyclic;
(c) b = n − 1.
A labeled tree on [n] is just a spanning tree of the complete graph Kn on
[n]. Hence we may state Cayley’s theorem as follows.
Proof #1. The first proof is due to H. Prüfer (1918). It uses an algorithm
that uniquely characterizes the tree.
Let T be a tree with V = [n], so the vertex set already has a natural
order. Let T1 := T . For i = 1, 2, . . . , n − 2, let bi denote the vertex of degree
1 with the smallest label in Ti , and let ai be the vertex adjacent to bi , and
let Ti+1 be the tree obtained by deleting the vertex bi and the edge {ai , bi }
from Ti . The “code” assigned to the tree T is [a1 , a2 , . . . , an−2 ].
32 CHAPTER 1. BASIC COUNTING TECHNIQUES
z6
A
A
A
4 x 5 x Ax7 z8
@ BB
@ B
x
@ B
@X
XXX
2 XX B
X XBz
x10
H H
3 x 1 HH
H
Hz9
HH
A Tree on 10 Points
Exercise: 1.10.2 With the sequence bi defined from the code as indicated in
the proof above, show that {{bi , ai } : i = 1, . . . , n − 1} will be the edge set of
a tree on [n]. Fill in the details of why the mapping associating a code to a
tree, and the mapping associating a tree to a code, are inverses.
Proof #2. This proof starts by showing that the number N (d1 , . . . , dn ) of
labeled trees on vertices {v1 , . . . , vn } in which vi has degree di + 1, 1 ≤ i ≤ n,
1.10. CAYLEY’S THEOREM: THE NUMBER OF LABELED TREES 33
is the multinomial coefficient d1n−2 ,...,dn
. As an inductive hypothesis we assume
that this result holds for trees with fewer than n vertices and leave to the
reader the task of checking that the result holds for n = 3. Since the degree
of each vertex is at least 1, we know that the d’s are all nonnegative integers.
The sum of the degrees of the vertices counts the n−1 edges twice, so we have
2(n − 1) = ni=1 (di + 1) = ( di ) + n ⇒ di = n − 2. Hence at least d1n−2
P P P
,...,dn
is in proper form. We also know that any tree has at least two vertices with
degree 1. We need to show that if (d1 , . . . , dn ) is a sequence of nonnegative
P
integers
with
di = n−2 then (d1 +1, . . . , dn +1) really is the degree sequence
n−2
of d1 ,...,dn labeled trees. Clearly if di = n − 2 then at least two of the di ’s
P
equal zero. The following argument would work with any particular dj = 0,
but for notational ease we suppose that dn = 0. If there is a labeled tree with
degree sequence (d1 + 1, . . . , 1), then the vertex vn is adjacent to a unique
vertex vj with degree at least 2. So the tree obtained by removing vn and
the edge {vj , vn } has degree sequence (d1 + 1, . . . , dj , . . . , dn−1 + 1). It follows
that N (d1 , . . . , dn−1 , 0) = N (d1 − 1, d2 , . . . , dn−1 ) + N (d1 , d2 − 1, . . . , dn−1 ) +
· · · + N (d1 , d2 , . . . , dn−1 − 1). By the induction hypothesis this is the sum of
the multinomial coefficients
! !
n−3 n−3
+ + ···+
d1 − 1, d2 , . . . , dn−1 d1 , d2 − 1, . . . , dn−1
! !
n−3 n−2
+ = .
d1 , d2 , . . . , dn−1 − 1 d1 , d2 , . . . , dn−1 , 0
Cayley’s Theorem now follows. For the number T (n) of labeled trees
on n vertices is the sum of all the terms N (d1, . . . , dn ) with di ≥ 0 and
Pn n−2
i=1 di = n − 2, which is the sum of all terms d1 ,d2 ,...,dn with di ≥ 0 and
Pn n−2
i=1 di = n − 2. Now in the multinomial expansion of (a1 + a2 + · · · + an )
n−2
set a1 = · · · = an = 1 to obtain the desired result T (n) = (1+1+. . .+1) =
n−2
n .
Proof #3. This proof establishes a bijection between the set of labeled
trees on n vertices and the set of mappings from the set {2, 3, . . . , n−1} to the
set [n] = {1, 2, . . . , n}. Clearly the number of such mappings is nn−2 . Suppose
f is such a mapping. Construct a functional digraph D on the vertices 1
through n by defining (i, f (i)), i = 2, . . . , n − 1 to be the arcs. Clearly 1 and
n have zero outdegrees in D, but each of them could have positive indegree.
In either case, the (weakly) connected component containing 1 (respectively,
34 CHAPTER 1. BASIC COUNTING TECHNIQUES
on [n] that together use each element in [n] exactly once. Each collection
of disjoint oriented cycles of rooted trees (using each element of [n] exactly
once) corresponds uniquely to a linear arrangement of a collection of rooted
trees (using each element of [n] exactly once). Hence nn is the number of
linear arrangements of rooted trees on [n] (by which we always mean that
the rooted trees in a given linear arrangement use each element of [n] exactly
once).
We claim now that nn = n2 tn , where tn is the number of (labeled) trees
on [n].
It is clear that n2 tn is the number of triples (x, y, T ), where x, y ∈ [n] and
T is a tree on [n]. Given such a triple, we obtain a linear arrangement of
rooted trees by removing all arcs on the unique path from x to y and taking
the nodes on this path to be the roots of the trees that remain, and ordering
these trees by the order of their roots in the original path from x to y. In
this way each labeled tree corresponds to n2 linear arrangements of rooted
trees on [n].
det(AS )det(B S )
X Y
det(A 4 B) = ei ,
S i∈S
det(AS )det(B S ).
0
if x is not incident with e, or e is a loop,
N (x, e) = 1 if x is the head of e,
−1
if x is the tail of e.
Lemma 1.11.4 Let A be a square matrix that has at most two nonzero en-
tries in each column, at most one 1 in each column, at most one -1 in each
column, and whose entries are all either 0, 1 or -1. Then det(A) is 0, 1 or
-1.
Proof: First let H be a connected digraph with n vertices and with inci-
dence matrix N . H must have at least n − 1 edges, because it is connected
and must have a spanning tree, so we may let S be a set of n−1 edges. Using
the notation of the Cauchy-Binet Theorem, consider the n by n−1 submatrix
NS of N whose columns are indexed by elements of S. By Lemma 1.11.3, NS
has rank n−1 iff the spanning subgraph of H with S as edge set is connected,
i.e., iff S is the edge set of a tree in H. Let N 0 be obtained by dropping any
single row of the incidence matrix N . Since the sum of all rows of N (or of
NS ) is zero, the rank of NS0 is the same as the rank of NS . Hence we have
the following:
38 CHAPTER 1. BASIC COUNTING TECHNIQUES
(
±1 if S is the edge set of a spanning tree in H,
det(NS0 ) = (1.14)
0 otherwise.
S S
where the sum is over all n − 1 subsets S of the edge set. By Eq. 1.14 this is
the number of spanning trees of G.
Example 1.12.4 The omega function: ω(n) is the number of distinct primes
dividing n.
Example 1.12.7 The Omega function: Ω(n) is the number of primes divid-
ing n counting multiplicity. So ω(n) = Ω(n) iff n is square-free.
Example 1.12.8 The tau function: τ (n) is the number of positive divisors
of n.
Example 1.12.9 The sigma function: σ(n) is the sum of the positive divi-
sors of n.
Obs. 1.12.11 f ∗ g = g ∗ f .
Obs. 1.12.12 If f, g, h are arithmetic functions, (f ∗ g) ∗ h = f ∗ (g ∗ h), and
[(f ∗ g) ∗ h)](n) = d1 d2 d3 =n f (d1 )g(d2 )h(d3 ).
P
= (f ∗ g)(m)(f ∗ g)(n).
Finally, we need to show that if f is multiplicative, in which case f −1
exists, then also f −1 is multiplicative. Define g as follows. Put g(1) = 1,
and for every prime p and every j > 0 put g(pj ) = f −1 (pj ). Then extend
g multiplicatively for all n ∈ P. Since f and g are both multiplicative, so
is f ∗ g. Then for any prime power pk , (f ∗ g)(pk ) = d1 d2 =pk f (d1 )g(d2 ) =
P
−1
(d2 ) = (f ∗ f −1 )(pk ) = I(pk ). So f ∗ g and I coincide on
P
d1 d2 =pk f (d1 )f
prime powers and are multiplicative. Hence f ∗ g = I, implying g = f −1 , i.e.,
f −1 is multiplicative.
Clearly µ is multiplicative, and d|n µ(d) = 1 if n = 1. For n = pe ,
P
Pe j
j=0 µ(p ) = 1 + (−1) + 0 · · · + 0 = 0. Hence µ ∗ U = I and we
P
d|n µ(d) =
have proved the following:
1.12. NUMBER THEORETIC FUNCTIONS 41
This follows from µ−1 = U and associativity. In its more usual form it
appears as:
X X
F (n) = f (d) ∀n ∈ P iff f (n) = µ(d)F (n/d) ∀n ∈ P.
d|n d|n
pα+1 −1
3. σ = U ∗ E is multiplicative and σ(n) =
Q
pα ||n p−1
.
4. φ ∗ τ = σ.
5. σ ∗ φ = E ∗ E.
6. E −1 (n) = nµ(n).
s0 = N ;
sm = N (a1 a2 · · · am ).
Also,
1.13. INCLUSION – EXCLUSION 43
e1 = N (a1 a02 a03 · · · a0m ) + N (a01 a2 a03 · · · a0m ) + · · · + N (a01 a02 · · · a0m−1 am );
..
.
em = N (a1 a2 · · · am ).
j=0 j r+j
Notice that
! !
r+j r+k (r + j)! (r + k)!
=
j r+j r!j! (r + j)!(k − j)!
! !
(r + k)! k! r+k k
= = .
r!k! j!(k − j)! r j
Thus the total count on the right hand side is
!
r + k j=k
!
X k
(−1)j = 0.
r j=0 j
44 CHAPTER 1. BASIC COUNTING TECHNIQUES
If we replace each property ai with its negation a0i , and use the notation
sˆi , Ŝ(x), eˆi , etc., for the corresponding analogues of si , ei , etc., we can see
how to write Ŝ(x) in terms of the si .
Theorem 1.13.5
m
! !
k m−k
sk x r .
X X
Ŝ(x) = (−1)
r=0 k m−r
1
Proof: It is clear that êi = em−i , i.e., Ê(x) is the reverse xm E x
of
E(x). Recall that S(x) = E(x + 1). Then
46 CHAPTER 1. BASIC COUNTING TECHNIQUES
1 −x
m
Ŝ(x) = Ê(x + 1) = (x + 1) E = (x + 1m E 1 +
x+1 s+1
m
−x
= (x + 1)m S (−x)k (x + 1)m−k sk .
X
=
x+1 k=0
r
The coefficient of x in this expression is
( )
n o
r k m−k
[xr−k ] (−1)k (x + 1)m−k sk
X X
[x ] (−x) (x + 1) sk =
k k
!
X
k m−k
= (−1) sk ,
k r−k
from which the theorem follows.
Application to derangements: Let Dn be the number of permutations
σ = b1 · · · bn of 1, 2, . . . , n for which bi 6= i for all i, i.e., Dn is the number
of derangements of n things. Here the N objects are the n! permutations.
The property ai is defined by: The permutation σ has property ! ai provided
n
bi = i. Then N (ai1 · · · air ) = (n − r)!, and sj = (n − j)! = n!
j!
. So
j
p
e0 = np=0 (−1)p n! = n! np=0 (−1)
P P
p! p!
.
Problèm des Rencontres: Let Dn,r be the number of permutations
σ = b1 · · · bn with exactly r fixed
! elements, i.e., bj = j for r values of j.
n
Choose r fixed symbols in ways, and multiply by the number Dn−r of
r
derangements on the remaining n − r elements.
h i
n!
Dn,r = r!(n−r)!
(n − r)!( 0!1 − 1
1!
+ 1
2!
+ ··· ± 1
(n−r)!
) . Or,
!
Pn−r p r+p n! n! Pn−r p1
Dn,r = p=0 (−1) (r+p)!
= r! p=0 (−1) p! .
r
Application to Euler’s Phi function: Let n = pa1 · · · par be the prime
power factorization of the positive integer n. Apply the Inclusion–Exclusion
Principle with E = [n] = {1, . . . , n}, and let ai be the property (of a positive
integer) that it is divisible by pi , 1 ≤ i ≤ r. This yields
r r
!
X n n X Y 1
φ(n) = n − + − ··· = n 1− .
i=1 pi 1≤i<j≤r pi pj i=1 pi
1.14. ROOK POLYNOMIALS 47
(m − j)!
sj = rj · (m − j)(m − j − 1) · · · (m − n + 1) = · rj . (1.20)
(m − n)!
X (m − k)!
r̂k xk = ŝk xk = Ŝ(x)
X
k (m − n)! k
1 −x
= Ê(x + 1) = (x + 1)n E = (x + 1)n E 1 +
x+1 x+1
(x + 1)n X
i
−x
= ri (m − i)!E
(m − n)! i x+1
48 CHAPTER 1. BASIC COUNTING TECHNIQUES
(m − i)!
(−x)i (x + 1)n−i .
X
= ri ·
i (m − n)!
k
The coefficient of x in the first term of this sequence of equal expressionsis
(m−k)!
clearly (m−n)! r̂k . The coefficient of xk in the last term is
( )
(m − i)!
[xk ] (−x)i (x + 1)n−i
X
ri
i (m − n)!
( )
k−i (m − i)!
(−1)i (x + 1)n−i
X
= [x ] ri
i (m − n)!
!
X (m − i)! n − i
i
= (−1) .
i (m − n)! k − i
Hence we have established the following theorem.
Theorem 1.14.1
!
X (m − i)! n − i
i
r̂k = (−1) ri .
i (m − k)! n − k
(−1)i (n − i)! · ri .
X
r̂n =
i
rk (C) = rk−1 (C 0 ) + rk (C 00 ).
It is now easy to see that we have
0 0 1 10
denote the rook polynomial of the displayed matrix and the 10 indicates the
entry about which the expansion formula is about to be applied. Then by
the expansion formula we have (since we write the matrix to mean its rook
polynomial)
! 1 1 0 0
1 1 0
RC (x) = x + 0 10 1 0
0 1 10
0 0 1 0
" !# ! 1 1 0 0
1 1 0 1 0 0
=x x 1 1 + + x
+ 0 0 1 0
0 10 0 0 1 0
0 0 1 0
50 CHAPTER 1. BASIC COUNTING TECHNIQUES
" !#
2 1 1 0
= x (1 + 2s) + x x(1 0) + + x(1 + x)2 + (1 + 2x)2
0 0 0
(Answer: 1 + 6x + 7x2 + x3 .)
1 1 1 1 1
1 0 1 1 1
c. .
0 0 1 1 0
0 0 1 1 1
a 0 1 0 0
b 0 0 0 1
.
c 1 0 0 0
d 0 0 1 0
1.15. PERMUTATIONS WITH FORBIDDEN POSITIONS 51
Since an object cannot be placed in more than one position and a position
cannot hold more than one object, in the matrix representation of an accept-
able distribution there will never be more than one 1 in a row or column.
Hence an acceptable distribution is equivalent to an independent 4-set of 1’s.
We can extend this notion to the case where there are forbidden positions
for each of the objects. For example, for the derangement of four objects,
the forbidden positions are
just those along the main diagonal. Also, it is
easy to see that rk (x) = k4 , so that RI (x) = (1 + x)4 . Hence the problem
of enumerating the number of derangements of four objects is equivalent to
the problem of finding the value of r̂4 for the complementary matrix
0 1 1 1
1 0 1 1
.
1 1 0 1
1 1 1 0
P4 i (4−i)! 4−i P4 i
By Theorem 1.14.1, r̂4 = i=0 (−1) (4−4)! 4−4
ri = i=0 (−1) (4 − i)! ×
4 P4 i1
i
= i=0 (−1) i! . Of course, this agrees with the usual formula.
Nontaking Rooks A chess piece called a rook can capture any oppo-
nent’s piece in the same row or column of the given rook (provided there are
no intervening pieces). Instead of using a normal 8 × 8 chessboard, suppose
we “play chess” on the “board” consisting solely of those positions of an n×m
(0,1)-matrix where the 1’s appear. Counting the number of ways to place k
mutually nontaking rooks on this board of entries equal to 1 is equivalent to
our earlier problem of counting the number of independent k-sets of 1’s in
the matrix. Consider the example represented by the following matrix:
1 1 0 0 1
1 1 1 0 1
B=
1 0 1 1 0
1 1 1 1 1
1 1 1 1 0
Thus rk (B) counts the number of ways k nontaking rooks can be placed
in those entries of B equal to 1. The 5 × 5 matrix B could be consideredto
have arisen from a job assignment problem. The rows correspond to workers,
the columns to jobs, and the (i, j) entry is a 1 provided worker i is suitable
52 CHAPTER 1. BASIC COUNTING TECHNIQUES
for job j. We wish to determine the number of ways in which each worker can
be assigned to one job, no more than one worker per job, so that a worker
only gets a job to which he or she is suited. It is easy to see that this is
equivalent to the problem of computing r5 (B). Since there are several more
1’s than 0’s, it might be easier to deal with the complementary matrix
0 0 1 1 0
0 0 0 1 0
B0 =
0 1 0 0 1
0 0 0 0 0
0 0 0 0 1
Easy Exercise: Show that if the matrix C 0 is obtained from matrix C
by deleting rows or columns with no entries equal to 1, then rk (C 0 ) = rk (C).
Let B 00 be the matrix obtained by deleting column 1 and row 4 from B 0 .
So
0 1 1 0
0 0 1 0
B 00 =
1 0 0 1
0 0 0 1
By Equation 1.21, we see that rB 00 (x) = rC1 (x) × rC2 (x),where
!
1 1
C1 = C2 = .
0 1
We easily compute rC1 (x) = 1 + 3x + x2 , so
The next example is a 5×7 matrix B that arises from a problem of storing
computer programs. The (i, j) position is a 1 provided storage location j has
1.15. PERMUTATIONS WITH FORBIDDEN POSITIONS 53
∗
Mn0 (x) = x · Mn−1
0
(x) + Mn−1 (x). (1.23)
Mn∗ has one 1 in its bottom row, in position (n, n). Expand about this 1.
Deleting the row and column of this 1 gives m∗n−1 . Changing this 1 to a 0
and deleting the bottom row of zeros gives the transpose of M̄n0 . Hence
Mn = 2n!Un .
Consequently we may concentrate our attention on the ménage numbers
Un . The formula we derived using rook polynomials will now be obtained
using recursion techniques.
Lemma 1.16.1 Let f (n, k) denote the number of ways of selecting k objects,
no two consecutive, from n objects arranged in a row. Then
!
n−k+1
f (n, k) = . (1.26)
k
56 CHAPTER 1. BASIC COUNTING TECHNIQUES
Proof: Clearly !
n
f (n, 1) = = n,
1
and for n > 1, !
1
f (n, n) = = 0.
n
Now let 1 < k < n. Split the selections into those that include the first
object and those that do not. Those that include the first object cannot
include the second object and are enumerated by f (n − 2, k − 1). Those that
do not include the first object are enumerated by f (n − 1, k). Hence we have
the recurrence
Lemma 1.16.2 Let g(n, k) denote the number of ways of selecting k objects,
no two consecutive, from n objects arranged in a circle. Then
!
n n−k
g(n, k) = )n > k).
n−k k
Proof: As before, split the selections into those that inlcude the first
object and those that do not. the selections that include the first object
cannot include the second object or the last object and are enumerated by
f (n − 3, k − 1).
The selections that do not include the first object are enumerated by
f (n − 1, k).
Hence
g(n, k) = f (n − 1, k) + f (n − 3, k − 1),
1.16. RECURRENCE RELATIONS: MÈNAGE NUMBERS AGAIN 57
a1 , b1 , a2 , b2 , . . . , an , bn .
Select k of these properties and ask for the number of permutations that
satisfy each of the k properties. The answer is 0 if the k properties are not
compatible. If they are compatible, then k images under the permutation
are fixed and there are (n − k)! ways to complete the permutation. Let vk
denote the number of ways of selecting k compatible properties from the 2n
properties. Then by the classical inclusion-exclusion principle,
n
(−1)i vi (n − i)!.
X
Un = (1.28)
i=0
67
68 CHAPTER 2. SYSTEMS OF REPRESENTATIVES AND MATROIDS
Lemma 2.1.2 If A satisfies Condition (H), then the union and intersection
of critical blocks are themselves critical blocks.
Proof of Lemma 2.1.2. Let Br,r and Bt,t be given critical blocks. Say Br,r ∩
Bt,t = Bu,v ; Br,r ∪ Bt,t = By,z . The z elements of the union will be the r + t
elements of Br,r and Bt,t reduced by the number of elements in both blocks,
and this latter number includes at least the v elements in the intersection:
z ≤ r + t − v. Also v ≥ u and z ≥ y by Condition (H). Note: y + u = r + t.
Hence r + t − v ≥ z ≥ y = r + t − u ≥ r + t − v, implying that equality holds
throughout. Hence u = v and y = z as desired for the proof of Lemma 2.1.2
.
as blocks of the union By,z , contain at least z − k elements not in Bk,k . Thus
s0 ≥ v + (z − k) ≥ u + y − k = u + (r + k − u) − k = r. Hence s0 ≥ r, as
desired for the proof of Lemma 2.1.3.
As indicated above, for the proof of the main theorem we now use induc-
tion on n. For n = 1 the theorem is obviously true.
Induction Hypothesis: Suppose the theorem holds (Condition (H) implies
that there is an SDR) for any family of m sets, 1 ≤ m < n.
We need to show the theorem holds for a system of n sets. So let 1 <
n, assume the induction hypothesis, and let A = (S1 , . . . , Sn ) be a given
collection of subsets of S satisfying Condition (H).
First Case: There is some critical block Bk,k with 1 ≤ k < n. Delete
the elements in the members of Bk,k from the remaining subsets, to obtain
a new family A0 = Bk,k ∪ Bn−k,v
0 0
, where Bk,k and Bn−k,v have no common
elements in their members. By Lemma 2.1.3, Condition (H) holds in A0 , and
0
hence holds separately in Bk,k and in Bn−k,v viewed as families of sets. By
0
the induction hypothesis, Bk,k and Bn−k,v have (disjoint) SDR’s whose union
is an SDR for A.
Remaining Case: There is no critical block for A except possibly the
entire system. Select any Sj of A and then select any element of Sj as its
representative. Delete this element from all remaining sets to obtain a family
A0 . Hence a block Br,s with r < n becomes a block Br,s
0 0
0 with s ∈ {s, s − 1}.
0
By hypothesis Br,s was not critical, so s ≥ r + 1 and s ≥ r. So Condition
(H) holds for the family A0 \ {Sj }, which by induction has an SDR. Add to
this SDR the element selected as a representative for Sj to obtain an SDR
for A.
In the text by van Lint and Wilson, Theorem 5.3 gives a lower bound on
the number of SDR’s for a family of sets that depends only on the sizes of
the the sets. It is as follows.
Theorem 5.3 of van Lint and Wilson: Let A = (S0 , S1 , . . . , Sn−1 ) be
a family of n sets that does have an SDR. Put mi = |Si | and suppose that
m0 ≤ m1 ≤ · · · ≤ mn−1 . Then the number of SDR’s for A is greater than or
equal to
70 CHAPTER 2. SYSTEMS OF REPRESENTATIVES AND MATROIDS
n−1
Y
Fn (m0 , m1 , . . . , mn−1 ) := (mi − i)∗ ,
i=0
Exercise: 2.1.6 Let G be a (finite, undirected, simple) graph with vertex set
V . Let C = {Cx : x ∈ V } be a family of sets indexed by the vertices of
G. For X ⊆ V , let CX = ∪x∈X Cx . A set X ⊆ V is C-colorable if one can
assign to each vertex x ∈ X a “color” cx ∈ Cx so that cx 6= cy whenever x
and y are adjacent in G. Prove that if |CX | ≥ |X| whenever X induces a
connected subgraph of G, then V is C-colorable. (In the current literature of
graph theory, the sets assigned to the vertices are called lists, and the desired
proper coloring of G chosen from the lists is a list coloring of G. When G is
a complete graph, this exercise gives precisely Hall’s Theorem on SDR’s. A
current research topic in graph theory is the investigation of modifications of
this condition that suffice for the existence of list colorings.
Exercise: 2.1.7 With the same notation of the previous exercise, prove that
if every proper subset of V is C-colorable and |CV | ≥ |V |, then V is C-
colorable.
2.1. THE THEOREM OF PHILIP HALL 71
Proof: First note that Hall’s theorem says that G has a matching of size
t = |X| if and only if δ(S) ≤ 0 for all S ⊆ X iff |X| ≤ |X| − δ(S) for
all S ⊆ X. So our theorem is true in case t = |X|. Now suppose that
t < |X|. Form a new graph G0 = (X, Y ∪ Z, E 0 ) by adding new vertices
Z = {z1 , . . . , z|X|−t } to Y , and join each zi to each element of X by an edge
of G0 .
If G has a matching of size t, then G0 has a matching of size |X|, implying
that for all S ⊆ X,
implying
Corollary 2.1.9 The largest matching of G has size |X| − δ(G) = m(G),
i.e., m(G) + δ(G) = |X|.
72 CHAPTER 2. SYSTEMS OF REPRESENTATIVES AND MATROIDS
T2 = {b̂1 , b2 , . . . , bt ; bt+1 , . . . , bs },
where bt+1 , . . . , bs are the elements in S(b1 ) not already in T1 .
If some one of bt+1 , . . . , bs is not in Ar , use it to represent S(b1 ) and use
b1 to represent Sr+1 . Leave the other ai ’s as before.
Each list Tj looks like Tj = {b̂1 , b̂2 , . . . , b̂k , bk+1 , . . . , bm }. If all members
of Tj are in Ar , construct
Tj+1 = {b̂1 , . . . , bˆk , b̂k+1 , bk+2 , . . . , bm , (list here any members of S(b̂k+1 )
Exercise: 2.3.4 Sixteen (male - female) couples and a caller attend a square
dance. At the door each dancer selects a name-tag of one of the colors red,
blue, green, white. There are four tags of each color for males, and the same
for females. As the tags are selected, each dancer fails to notice what color
her/his partner selects. The caller is then given the job of constructing four
squares with four (original!) couples each in such a way that in each square
no two dancers of the same sex have tags of the same color. Show that this
is possible no matter how the dancers select their name tags.
1 ≤ i, j ≤ n − 1;
a if
Pij
1 − n−1
k=1 aik if i 6= n and j = n;
Pn−1
1 − k=1 akj if j 6= n and i = n;
2 − n + n−1
P Pn−1
l=1 akl if i = j = n.
k=1
Theorem 2.4.1 Suppose that for each i in some index set I there is a finite
subset Ai of a set S. The system A = (Ai )i∈I has an SDR if and only if
the following Condition (H’) holds: For each finite subset I 0 of I the system
A0 = (Ai )i∈I 0 satisfies Condition (H).
a maximal deletion each element left is in some critical block. And if Bk,k is
a critical block, we may delete elements of Bk,k from all sets not in Bk,k and
still preserve Condition (H) by Lemma 2.1.3 (since it needs to apply only
to finitely many sets at a time). By Theorem 2.1.1 each critical block Bk,k
(being finite) possesses an SDR when Condition (H) holds. Hence we may
perform an additional deletion leaving Bk,k as a collection of singleton sets
and with Condition (H) still holding for the entire remaining sets. It is now
clear that after a maximal deletion D̄ preserving Condition (H), each element
78 CHAPTER 2. SYSTEMS OF REPRESENTATIVES AND MATROIDS
is in a critical block, and each critical block consists of singleton sets. Hence
after a maximal deletion D̄ preserving Condition (H), each set consists of a
single element, and these elements form an SDR for A.
The following theorem, sometimes called the Cantor–Schroeder–Bernstein
Theorem, will be used with the theorem of M. Hall, Jr. to show that any
two bases of a vector space V over a field F must have the same cardinality.
X1 = {x ∈ X : x has no eve};
X2 = {x ∈ X : x has an eve};
Y1 = {y ∈ Y : y has no eve};
Y2 = {y ∈ Y : y has an eve}.
Now a little thought shows that θ : X1 → Y1 is a bijection, and ψ −1 :
X2 → Y2 is a a bijection. So
φ = θ|X1 ∪ ψ −1 |X2
is a bijection from X to Y .
1. Let I = ∅.
4. Delete x from X.
80 CHAPTER 2. SYSTEMS OF REPRESENTATIVES AND MATROIDS
We do the case for matchings in bipartite graphs, and we note that if the
empty set (as a subset of X in G = (X, Y, E)) is defined to be independent,
then clearly the set I of independent subsets of X satisfies the Subset Rule.
So we consider the Expansion Rule. However, before proceding into the
proof we need to be sure that our terminology is clear and we need to prove
a couple preliminary lemmas.
A matching M of size m in a graph G is a set of m edges, no two of which
have a vertex in common. A vertex is said to be matched (to another vertex)
by M if it lies in an edge of M . We defined a bipartite graph G with parts
X and Y to be a graph whose vertex set is the union of the two disjoint sets
X and Y and whose edges all connect a vertex in X with a vertex in Y . A
complete matching of X int Y is a matching of G with X edges. Our first
lemma describes the interaction of two matchings.
Lemma 2.5.3 Let M1 and M2 be matchings of the graph G = (V, E). Let
G0 = (V, E 0 ) be the subgraph with E 0 = (M1 ∪ M2 ) \ (M1 ∩ M2 ) = (M1 \ M2 ) ∪
(M2 \ M1 ). Then each connected component of G0 is one of the following
three types:
(i) a single vertex
(ii) a cycle with an even number of edges and whose edges are
alternately in M1 and M2 .
(iii) a chain whose edges are alternately in M1 and M2 ,
and whose two end vertices are each matched by one of
M1 , M2 but not both.
Moreover, if |M1 | < |M2 |, there is a component of G0 of type (iii) with
first and last edges in M2 and whose endpoints are not M1 -matched.
Each vertex of G1 has degree 1 or 2 (on at most one edge of M1 and at most
one edge of M2 ). Hence
X
2(n − 1) ≤ 2(number of edges of G1 ) = deg(x) ≤ 2n.
x∈V (G1 )
(M \ M 0 ) ∪ (E 0 \ M 0 ) = (M \ E 0 ) ∪ (E 0 \ M )
is a matching with one more edge than M has.
Proof: If for some F ⊆ E it is true that (V, F ) has no cycles, then (V, F 0 )
has no cycles for any subset F 0 of F . This says that the Subset Rule is
satisfied.
84 CHAPTER 2. SYSTEMS OF REPRESENTATIVES AND MATROIDS
Polya Theory
89
90 CHAPTER 3. POLYA THEORY
Proof: These results are special cases of results usually proved for homo-
morphisms in general. If you don’t remember them, you should work out the
proofs in this special case.
Let G act on X. For x, y ∈ X, define x ∼ y iff there is some g ∈ G for
which g(x) = y.
f (gGx ) = g(x).
First we show that f is well-defined. If g1 Gx = g2 Gx , then g2−1 g1 ∈ Gx ,
so that (g2−1 · g1 )(x) = x, which implies g1 (x) = g2 (x). Hence f (g1 Gx ) =
f (g2 Gx ). So f is well-defined. Now we claim f is a bijection. Suppose
f (g1 Gx ) = f (g2 Gx ), so by definition g1 (x) = g2 (x) and
(g2−1 · g1 )(x) = x. Hence g2−1 · g1 ∈ Gx , implying g1 Gx = g2 Gx , so f is one-to-
one. If y ∈ [x], then there is a g ∈ G with g(x) = y. So f (gGx ) = g(x) = y,
implying f is onto [x].
Hence f is a bijection from the set of left cosets of Gx in G to [x], i.e.,
|G|/|Gx | = |[x]| as claimed.
µ(g)(f ) = f ◦ ν(g −1 ).
1 X 1 X
|Fµ(g) | = |Y |c(g) ,
|G| g∈G |G| g∈G
3.2 Applications
1 2
Example 3.2.1 Let G be the group of symmetries of the square writ-
4 3
ten as permutations of [4] = {1, 2, 3, 4}. The convention here is that if a
symmetry σ of the square moves a corner labeled i to the corner previously
labeled j, then σ(i) = j. We want to paint the corners with W and R (white
and red) and then determine how many essentially different paintings there
are.
1 X
|Y |c(g) .
|G| g∈G
1 4
(2 + 21 + 22 + 21 + 23 + 22 + 23 + 22 ) = 6.
8
Example 3.2.2 How many necklaces are there with n beads of m colors if
two are the same provided one can be rotated into the other?
n
.
gcd(n, i)
So the number of cycles in σ i is gcd(n, i). For each d such that d|n, there
are φ(n/d) integers i with 1 ≤ i ≤ n and d = gcd(n, i). So the number of
σ i ∈ G with d cycles is φ(n/d), for each d with d|n. If Y is the set of m
colors to be used, the number of distinct necklace patterns under the action
of G is
1 X c(g) 1X 1X
m = φ(n/d)md = φ(d)mn/d .
n g∈G n d|n n d|n
with rational coefficients which lie between 0 and 1, but f (m) is a positive
integer for each positive integer m.
Exercise: 3.2.3 If G also has “flips,” so |G| = 2n, i.e., G is dihedral, how
many necklaces of n beads in m colors are there? (Hint: Do the cases n odd,
n even separately.)
94 CHAPTER 3. POLYA THEORY
Answer to Exercise
(i) For n odd, the number of necklaces is:
1 X n+1
φ(n/d)md + nm 2 .
2n d|n
[n]
Z2 = {x : [n] → Z2 = {0, 1}}
[n]
according to the following: For x ∈ Z2 (write x = (x1 , . . . , xn ): xi =0 or 1,
and where xi is the image of i under x), and g ∈ G, put
3.2. APPLICATIONS 95
Now let
n
Fn = {f : Z2n → Z2 } = Z (Z2 ) .
(µ(g))(f ) = f ◦ ν(g −1 ),
i.e.,
This last equation says that g ∈ Sn acts on Fn by: (g(f ))(x) = f (x ◦ g),
i.e.,
We say that f1 and f2 are equivalent if they are in the same G-orbit, and
we would like to determine how many inequivalent switching functions in n
variables there are. This is too difficult for us to do for general n, so we put
n = 3. But first we rename the elements of Z23 essentially by listing them in
a natural order so that we can simplify notation in what follows.
(000) ↔ 0; (001) ↔ 1; (010) ↔ 2; (011) ↔ 3;
F3 = {f : Z23 → Z2 } = {f : {0, . . . , 7} → Z2 }.
1 X c(g) 1 8
2 = [2 + 24 + 26 + 26 + 24 + 26 ] = 80
|G| g∈G 6
3
(whereas |F3 | = 2(2 ) = 28 = 256).
1 X λ1 (g)
PG (x1 , . . . , xn ) = x · · · xλnn (g) .
|G| g∈G 1
3.3. THE CYCLE INDEX: POLYA’S THEOREM 97
1 X (Pn λt (g))
PG (m, . . . , m) = m t=1 .
|G| g∈G
(pα m)!
!
pα m
= =
pα (pα )!(pα m − pα )!
Looking at this expression written out, one can see that except for the
factor m in the numerator, the power of p dividing (pα m − i) = pα − i +
pα (m − 1) is the same as that dividing pα − i, since 0 ≤ i ≤ pα − 1, so all
powers of p cancel out except the power which divides m.
there must be some t for which p does not divide |Qt |. Choose any T ∈ Qt ,
α
so Qt = {hT : h ∈ H}. Then |HT | · |Qt | = |H| = pα and |Qt | = |Hp T | . Hence
pα = |HT |, since p does not divide |Qt |. Since HT ≤ H and |HT | = |H|, we
have H = HT . Also T ∈ O with p not dividing |O|, so |GT | = pα . Then
HT ≤ GT and |HT | = |GT | imply that HT = GT . Hence H = GT .
(b) tp |q.
is then defined by
Y
W (f ) = w(f (x)).
x∈X
since ν(g −1 )(x) varies over all elements of X as x varies over all elements of
X. So the weight of a pattern may be defined as the weight of any function in
P |X|
that pattern. The inventory f ∈Y X W (f ) of Y X is equal to
P
y∈Y w(y) .
This is a special case (with each |Xi | = 1) of the following result, whose proof
is given.
3.5. PATTERNS AND WEIGHTS 101
w(φ(i))|Xi | =
Y Y Y
w(φ(i)) = w((φ ◦ ψ)(x)) = w(f (x)),
x∈Xi x∈Xi x∈Xi
k
[w(φ(i))]|Xi | =
Y Y
w(f (x)) = W (f ).
i=1 x∈X
k X k
!
|Xi | |Xi |
Y X Y X
(w(y)) = w(φ(i)) = W (f ).
i=1 y∈Y φ:{1,...,k}→Y i=1 f ∈S
P |X|
If each |Xi | = 1, we have S = Y X and
P
f ∈Y X W (f ) = y∈Y w(y) .
[w(y)]2 , . . . , [w(y)]m ),
X X X X
W (F ) = PG ( w(y),
F y∈Y y∈Y y∈Y
where the summation is over all patterns F , and PG is the cycle index. In
particular, if all weights are chosen equal to 1, the number of patterns is
102 CHAPTER 3. POLYA THEORY
Proof: Let w be one of the possible values that the weight of a function
may have. Put S = {f ∈ Y X : W (f ) = w}. If g ∈ G, then W (f ◦ ν(g −1 )) =
w. Hence for each g ∈ G, µ(g) : f → f ◦ ν(g −1 ) maps S into S. (And it is
easy to see from earlier results that µ is an action of G on S.) Clearly, for
f1 , f2 ∈ S, f1 and f2 belong to the same pattern (in the sense mentioned at
the beginning of this section) if and only if they are equivalent relative to
the action µ of G on S. Not Burnside’s Lemma applied to µ : G → SS says
1 P
that the number of patterns contained in S is equal to |G| g∈G ψw (g), where
ψw (g) denotes the number of functions f with W (f ) = w and f = µ(g)(f ) =
f ◦ ν(g −1 ).
The patterns contained in S all have weight w. So if we multiply by w
and sum over all possible values of w, we obtain the patern inventory
X 1 XX
W (F ) = ψw (g) · w.
F |G| w g∈G
Also,
(g)
X X
ψw (g) · w = W (f ),
w f
where the right hand side is summed over all f ∈ Y X with f = f ◦ ν(g −1 ).
It follows that
(g)
X 1 XX
(W (F )) = W (f ).
|G| g∈G f
i.e., f is constant on each cycle of ν(g −1 ), and hence on each cycle of ν(g).
Conversely, each f constant on each cycle of ν(g) automatically satisfies
f = f ◦ ν(g −1 ), since ν(g −1 )(x) always belongs to the same cycle as x itself.
P(g)
Thus if the cycles are X1 , . . . , Xk , then f W (f ) is the inventory calculated
by Theorem 3.5.1 to be
(g) k
[w(y)]|Xi | .
X Y X
W (f ) =
f i=1 y∈Y
Let (b1 , . . . , bm ) be the cycle type of ν(g). This means that among the
numbers |X1 |, . . . , |Xk |, the number 1 occurs b1 times, 2 occurs b2 times, . . . ,
etc. Hence
P 1 P P(g)
Finally, F W (F ) = |G| g∈G f W (f ) is obtained by putting
i 1 P b1 bm
g∈G x1 · · · xm .
P
xi = y∈Y (w(y)) in PG (x1 , . . . , xm ) = |G|
We close this section with some examples that partly duplicate some of
those given earlier.
2
w(y)|Xi | = (1 + x2 + x4 + · · · + x2m )(1 + x + x2 + · · · + xm ).
Y X
i=1 y∈Y
104 CHAPTER 3. POLYA THEORY
1 1 1
(1 − x2 )−1 (1 − x)−1 = (1 + x)−1 + (1 − x)−2 + (1 − x)−1 ,
4 2 4
1
(1 − x + x2 − x3 + · · · + (−1)m xm + · · ·)+
4
∞
!
1 2+i−1 1
xi + (1 + x + x2 + · · · + xm + · · ·),
X
+
2 i=0 i 4
which is equal to
(
1 1 1
m + 1, m even,
(m + 1) + ((−1)m + 1) = 2
1
2 4 2
(m + 1), m odd.
For the next few examples let G be the group of rigid motions of a cube.
The elements of G were given earlier, but this time we want to include the
details giving the cycle indices. Recall the elements of G from Example 3.3.2.
Example 3.5.4 Let X be the set of vertices of the cube. The cycle types are
indicated as follows:
Example 3.5.5 Let X be the set of edges of the cube. The cycle types are
indicated as follows:
(a) x12
1 (b) x62 (c) x34
1 6
PG = (x + 3x21 x22 + 6x21 x4 + 6x32 + 8x23 ).
24 1
Example 3.5.7 Determine how many ways a cube can be painted so that
each face is red or blue? In other words, how many patterns are there?
Example 3.5.8 In the preceding example, how many color patterns show
four red faces, two blue?
1
[(x + y)6 + 3(x + y)2 (x2 + y 2 )2 +
X
W (F ) =
F 24
Example 3.5.9 In how many ways can the eight vertices be painted with n
colors?
1
PI = [(x1 + · · · + xn )8 + 9(x21 + · · · + x2n )4 + 6(x41 + · · · + x4n )2 +
24
Let k(a) be the order of a for each a ∈ G. Then λa splits G into m/k(a)
cycles of length k(a). So
1 X m/k(a) 1 X m/d
PG = xk(a) = ν(d)xd ,
m a∈G m d|m
1 X m/d
PG = φ(d)xd .
m d:d|m
m!
#(b̄) = .
b1 !1b1 b 2 !2b2 b b3
3 !3 · · · bm !m
bm
!
m
Proof: There are ways to form the 1-cycles. Suppose we have
b1
taken care of the 1-cycles, 2-cycles, . . . , (k − 1)-cycles and are about to form
the k-cycles. We have ηk−1 = m − b1 − 2b2 − · · · − (k − 1)bk−1 elements at !
ηk−1
our disposal (η0 = m). The first k-cycle can be formed (k − 1)!
k
!
ηk−1 − k
ways, the second k-cycle in (k − 1)! ways, . . . , the bk th k-cycle
k
!
ηk−1 − (bk − 1)k
in (k − 1)! ways. Hence the k-cycles can be formed in
k
( ! ! !
1 ηk−1 ηk−1 − k ηk−1 − 2k
(k − 1)! (k − 1)! (k − 1)! ···
bk ! k k k
!)
ηk−1 − (bk − 1)k
· · · (k − 1)! =
k
(
1 (k − 1)!ηk−1 ! (k − 1)!(ηk−1 − k)! (k − 1)!(ηk−1 − 2k)!
= · · ···
bk ! k!(ηk−1 − k)! k!(ηk−1 − 2k)! k!(ηk−1 − 3k)!
X xb11 · · · xbmm
PS m = ,
b̄
b1 !1b1 b2 !2b2 · · · bm !mbm
Now let M = {1, 2, . . . , m}, and let X be the set of unordered pairs {i, j}
of distinct elements of M . The symmetric group Sm acting on M also has a
natural action on X. For each σ ∈ Sm , define σ ∗ ∈ SX by
∗
µ : Sm → Sm = {σ ∗ : σ ∈ Sm }
∗
is an isomorphism. We need to calculate the cycle index of Sm .
This is feasible because the cycle type of each σ determines the cycle type
of the corresponding σ ∗ . Specifically, each factor xbt t in a term xb11 · · · xbmm of
PSm corresponding to a σ of cycle type (b1 , . . . , bm ) yields specific factors in
the term corresponding to σ ∗ .
Let σ ∈ Sm have cycle type (b1 , . . . , bm ), m i=1 ibi = m. Then for {i, j} ∈
P
(r,t)br bt
If r 6= t, xbrr xbt t ∈ PSm 7→ x[r,t] ∈ PSm∗ .
!
bk
k
2
If r = t = k, xbkk ∈ PSm 7→ xk .
PSm∗ (x1 , . . . , xm ) =
1 X m!
·
m! b̄ b1 ! · · · bm !1b1 2b2 · · · mbm
!
bk
[ m−1
2 ]
m
[Y
2 ]
m
[Y
2 ]
k
kb2k+1 2 (r,t)br bt
(xk xk−1 b2k
Y Y
· x2k+1 · 2k ) · xk · x[r,t]
k=0 k=1 k=1 1≤r<t≤m−1
1 4
PS 4 = (x + 6x21 x2 + 8x1 x3 + 3x22 + 6x4 ).
4! 1
1 6
PS4∗ = (x1 + 6x21 x22 + 8x23 + 3x21 x22 + 6x2 x4 ).
4!
110 CHAPTER 3. POLYA THEORY
PS4∗ (1 + x, 1 + x2 , 1 + x3 , 1 + x4 ) =
Cn (x) = (1 + O1 x)(1 + O2 x) · · · (1 + On x) = 1 + a1 x + · · · an xn ,
where ak is the “kth elementary symmetric function” of the n variables O1 to
On . Cn (x), after multiplication of it factors, contains the actual exhibition
of combinations. If only the number of combinations is of interest, the object
labels may be ignored and the generating function becomes an enumerating
generating function (sometimes just called an enumerator), i.e.,
∞
!
n n
xk .
X
Cn (x) = (1 + x) =
k=0
k
113
114CHAPTER 4. FORMAL POWER SERIES AS GENERATING FUNCTIONS
n−1
!
n n−1
xk
X
Cn (x) = (1 + x) = (1 + x)Cn−1 (x) = (1 + x)
k=0
k
! " ! !#
n−1 n−1 n−1
= + + x + ···+
0 0 1
" ! !# !
n−1 n−1 n−1 n−1
+ + x + xn ,
n−2 n−1 n−1
where for 1 ≤ k ≤ n − 1, the coefficient on xk is
! !
n−1 n−1
+ .
k−1 k
But
n
!
n
xk ,
X
Cn (x) =
k=0
k
so
! ! !
n n−1 n−1
= + ,
k k−1 k
a basic relation for binomial coefficients.
Now look again at the generating function for combinations of n distinct
objects. In its factored form, each object is represented by a binomial and
each binomial spells out the fact that its object has two possibilities in any
combination: either it is absent (the term 1) or it is present (the term Oi x for
the object Oi ). So Cn (x) = (1 + O1 x) · · · (1 + On x) is the generating function
for combinations without repetition. For repetitions of a certain kind, we
use a special generating function. For example, if an object may appear in
a combination zero, one or two times, then the function is the polynomial
1 + Ox + O2 x2 . If the number of repetitions is unlimited, it is the function
1 + Ox + O2 x2 + O3 x3 · · · = (1 − Ox)−1 . If the number of repetitions is even,
the generating function is 1 + O2 x2 + O4 x4 + · · · + O2k x2k + · · ·. Moreover,
the specification of repetitions may be made arbitrarily for each object. The
generating function is a representation of this specification in its factored
4.1. USING POWER SERIES TO COUNT OBJECTS 115
∞ ∞
!
k n+k−1
xk .
X X
= (−n)(−n − 1) · · · (−n − k + 1)(−x) /(k!) =
k=0 k=0
k
∞
!
2 n −n
X n+k−1 k
(1 + x + x + · · ·) = (1 − x) = x . (4.1)
k=0 k
For the problem in the above example with the added specification that
2 n
each object must appear at least once,
! ! + x + · · ·) =
the enumerator is (x
n+k−1 k−1
xn (1 − x)−n = ∞ xk+n = ∞ xk . Hence the
P P
k=0 k=n
k k−n
number of combinations of n objects taken k at a time with the restriction
that each object appear at least once is the same as the number of combina-
tions without repetition of k − 1 objects taken k − n at a time.
In this section we have seen power series manipulated without any con-
cern for whether or not they converge in an analytic sense. We want to
establish a firm foundation for the theory of formal power series so that we
can continue to work “magic” with series, so a significant part of this chap-
ter will be devoted to developing those properties of formal power series that
have proved to be most useful. Before starting this development we give one
more example to illustrate the power of these methods.
116CHAPTER 4. FORMAL POWER SERIES AS GENERATING FUNCTIONS
x x2
Bk (x) = Bk−1 (x) = Bk−2 (x).
1 − kx (1 − kx)(1 − (k − 1)x)
x x x2 x
So B1 (x) = 1−x
·1; B2 (x) = B (x)
1−2x 1
= (1−x)(1−2x)
; B3 (x) = B (x)
1−3x 2
=
x3
(1−x)(1−2x)(1−3x)
. And in general,
xk
( )
n
xn =
X
Bk (x) = , k ≥ 0. (4.5)
n k (1 − x)(1 − 2x) · · · (1 − kx)
1
αr =
(1 − 1/r)(1 − 2/r) · · · (1 − (r − 1)/r)(1 − (r + 1)/r) · · · (1 − k/r)
rk−1 rk−1 (−1)k−r
= = ,
(r − 1)(r − 2) · · · (1)(−1) · · · (−(k − r)) (r − 1)!(k − r)!
which implies
(−1)k−r rk−1
αr = , 1 ≤ r ≤ k.
(r − 1)!(k − r)!
Notation: If f (x) = ∞ n n
n=0 an x , then [x ]{f } = an . Clearly in general
P
n n+r r
we have [x ]{f } = [x ]{x f }.
We now have for k ≥ 1:
xk
( ) ( ) ( )
n 1
= [xn ] = [xn−k ]
k (1 − x) · · · (1 − kx) (1 − x) · · · (1 − kx)
k k
αr 1
= [xn−k ] αr [xn−k ]
X X
=
r=1 1 − rx r=1 1 − rx
k k
X
n−k
X (−1)k−r rn−1
= αr r = =
r=1 r=1 (r − 1)!(k − r)!
k
!
1 k−1
(−1)k−r rn−1
X
= .
(k − 1)! r=1 r−1
( )
n
This gives a closed form formula for :
k
k−1
( ) !
n 1 k−1
(−1)k−r−1 (r + 1)n−1
X
= . (4.6)
k (k − 1)! r=0 r
118CHAPTER 4. FORMAL POWER SERIES AS GENERATING FUNCTIONS
f − a0 − a1 x − · · · − ah−1 xh−1
↔ {an+h }∞
n=0 .
xh
ah xh +ah+1 xh+1 +···
This follows immediately, since the L.H.S. equals xh
= ah +
ah+1 x + · · · + ah+n xn + · · ·.
an xn is f 0 = nan xn−1 .
P P
Def. The derivative of f =
Proposition 1. f 0 = 0 iff f = a0 , i.e., ai = 0 for ı ≥ 1.
Proposition 2. If f 0 = f , then f = cex .
Proof: f = f 0 iff ∞
P∞
n n−1
= ∞ n
P P
n=0 an x = n=1 nan x n=0 (n + 1)an+1 x iff
an
an = (n + 1)an+1 for all n ≥ 0, i.e., an+1 n+1 for n ≥ 0. By induction on n,
P xn
an = A0 /n! for all n ≥ 0. So f = a0 n! = a0 ex .
P∞
Starting with f = n=0 an xn , we clearly have
∞
0
(n + 1)an+1 xn .
X
f =
n=0
n=0 n=0
where the sum is over all (an1 , . . . , ank ) for which n1 + n2 + · · · + nk = n and
ni ≥ 0.
Example: Lef f (n, k) be the number of weak k-compositions of n. Since
1 ops 1 ops ∞
(1−x)
↔ {1}, by Rule 4, (1−x) k ↔ {f (n, k)}n=0 . And as we have already seen,
! !
−k −k
(1 − x)−k = ∞ (−1)n xn implies that f (n, k) = (−1)n
P
n=0 =
n n
!
n+k−1
.
n
1
We now ask what happens when we multiply by 1−x :
2 2
f (x)/(1 − x) = (a0 + a1 x + a2 x + · · ·)(1 + x + x + · · ·) = a0 + (a0 + a1 )x +
(a0 + a1 + a2 )x2 + · · ·.
This leads to Rule 5:
ops ops Pn
Rule 5: If f ↔ {an }∞
0 , then f /(1 − x) ↔ { j=0 aj }n≥0 .
1 (m+n)!
Exercise: 4.3.1 Dn (1−x)m+1
= m!(1−x)m+n+1
; m, n ≥ 0.
Recall that there is a “formal” Taylor’s formula. If you did not prove it
when we met it earlier, do it now.
120CHAPTER 4. FORMAL POWER SERIES AS GENERATING FUNCTIONS
P∞ 1
Exercise: 4.3.2 If f (x) = n=0 an xn , then an = n!
Dn (f (x))|x=0 .
1 ops ops
1
Example 4.3.3 Start with f = 1−x
↔ {1}n≥0 . By Rule 2 we have (xD)2 1−x ↔
ops P
{n2 }n≥0 . Then by Rule 5, 1−x1
(xD)2 1−x 1
↔ { nj=0 j 2 }n≥0 . This implies
= [xn ] x(1+x)
Pn 2 n 1 2 1
j=0 j = [x ] 1−x (xD) 1−x (1−x)4
, after some calculation. From
!
1 (n+3)! n+3 Pn
Eq.4.7 with m = 3, [xn ] (1−x) 4 = 3!n!
= . Hence j=0 j2 =
3
! !
n
x x2
n−1
1
n−2
1
n+2 n+1
[x ] (1−x)4
+ (1−x)4
= [x ] (1−x)4
+[x ] (1−x)4
= + =
3 3
n(n+1)(2n+1)
6
.
Example 4.3.4 (Harmonic numbers) Put Hn = 1 + 12 + 13 + · · · + n1 , n ≥ 1.
n
Start with f = n≥1 xn = −log(1−x). (Check: f 0 = n≥1 xn−1 = n≥0 xn =
P P P
ops
1
1−x
, and (−log(1 − x))0 = (−1) · 1−x
1 1
· (−1) = 1−x ). So f ↔ { n1 }n≥1 . By Rule
ops
1
5, 1−x f ↔ {Hn }∞n=1 . So
1 1
ops
log ↔ {Hn }∞ n=1 .
1−x 1−x
In the first three sections of this chapter we gave examples of how to use
“ordinary power series” as generating functions. There are other ways to
associate a possibly infinite sequence with a power series. In later sections
we explore a couple more types, the more important of which are the exp-
ponential generating functions. But first we begin our exposition of formal
power series.
Exercise: 4.3.5 Given three types of objects O1 , O2 and O3 , let an be the
number of combinations of these objects, n at a time, so that O1 is selected at
most 2 times, O2 is selected an odd number of times, and O3 is selected at least
once. Determine the ordinary generating function of the sequence {an }∞ n=0
and determine a recursion that is satisfied by the sequence. Compute a closed
form formula for an . As always, be sure to include enough details so that I
can see a proof that your answers are correct.
4.4. FORMAL POWER SERIES 121
Exercise: 4.3.6 If O1 can appear any number of times and O2 can appear
any multiple of k times, use Rule 5 to show that the number of combinations
of n objects must be an = 1 + b nk c.
i=0
ai = bi for all i = 0, 1, 2, . . . .
122CHAPTER 4. FORMAL POWER SERIES AS GENERATING FUNCTIONS
ai xi )( bi x i ) = ci xi ,
P P P
(ii) (
Pi
where ci = j=0 aj bi−j .
P∞
As an example of multiplication, consider the product (1−x) n=0 xn = 1.
Be sure to check this out! It says that (1 − x)−1 = ∞ n
P
n=0 x .
The ring of all formal power series in x over the commutative ring A is
denoted by A[[x]].
Proof: All the details are fairly routine. We include here only the details
for associativity of multiplication. So, to show
hX X i X X hX X i
ai x i · bi x i ci xi = ai x i bi x i ci xi
we must show that the coefficient of xj on on the left hand side is equal to the
coefficient of xj on the right hand side. The coefficientof xj on the L.H.S.
Pj Pk Pj Pk
is d c , where dk = i=0 ai bk−i . This is k=0 i=0 ai bk−i cj−k =
P k=0 k j−k
ai bk cl , where this last sum is over all i, k, l for which i + k + l = j, 0 ≤
i, k, l ≤ j. And the last equality is obtained by observing that each term
of the R.H.S. of this last equality appears exactly once on the left, and vice
versa. Similarly, the coefficient of xj on the R.H.S. of the main equality to
P
be established is also equal to ai bk cl , where this sum is over all i, j, k for
which i + k + l = j, 0 ≤ i, k, l ≤ j.
We have established the following:
n X o
[xj ] ( ai xi )( bi xi )( ci xi ) =
X X X
ai b k c l (4.8)
!k
∞
X
[xn ] ai x i
X
= an1 an2 · · · ank (4.9)
i=0
Theorem 4.4.4 If K is a field, then the units of K[[x]] are the power series
of order 0, i.e., with nonzero constant term.
The next result is inserted here for the algebraists. It will not be needed
in this course.
124CHAPTER 4. FORMAL POWER SERIES AS GENERATING FUNCTIONS
Theorem 4.4.5 Let K be a field. Then the ring K[[x]] has a unique maximal
ideal (generated by x), and the nontrivial ideals are all powers of this one.
Theorem 4.4.6 Let K be a field, and let f (x), g(x) ∈ K[[x]] with f (x) 6= 0.
Then there is a unique power series q(x) ∈ K[[x]] and there is a unique poly-
nomial r(x) ∈ K[x] such that g(x) = q(x)f (x) + r(x) and either deg(r(x)) <
o(f (x)) or r(x) = 0.
to show that !2
2n
!
X
k 2n n 2n
(−1) = (−1) .
k=0 k n
If b(x) = ∞ n
P
n=0 bn x is a power series with o(b(x)) = 1, i.e., b0 = 0
and b1 6= 0, we can also find the (unique!) inverse function as a power
series. We “solve” the equation b(a(x)) = x by substitution, assuming that
a(x) = ∞ n
P
n=1 an x with a0 = 0. Then
∞ ∞ ∞
an x n ) + b 2 ( an xn )2 + b3 ( an xn )3 + · · ·
X X X
x = b1 (
n=1 n=1 n=1
f 0 (x) = (i + 1)ai+1 xi .
X
(4.10)
i≥0
and
(f (x)g(x))0 = + j)ci dj xi+j−1
P
i,j≥0 (i
X f (n) (0)xn
f (x) = . (4.15)
n≥0 n!
In order to define the integral of f (x) we need to assume that the char-
acteristic of A is zero. In that case define the formal integral Ix f (x) by
Z x
i−1 ci−1 xi .
X
Ix f (x) = f (x)dx = (4.17)
0 i≥1
Z x
0 0
(Ix f (x)) = f (x); If F (x) = f (x), then f (x) = F (x) − F (0). (4.18)
0
4.7. LOG, EXP AND BINOMIAL POWER SERIES 129
X xj
ex = exp(x) = . (4.19)
j≥0 j!
X xj
log((1 − x)−1 ) = . (4.20)
j≥1 j
xj
(1 + x)y = 1 +
X
y(y − 1) · · · (y − j + 1) =
j≥1 j!
!
y
xj ∈ (R[y])[[x]].
X
= (4.21)
j≥0
j
log(ex ) = x (4.23)
so that
(1 − x)exp(log((1 − x)−1 )) = 1,
and
This is the binomial expansion for positive integers. Thus for positive
integers m and n, [xk ] can be applied to the binomial series expansion of
the identity (1 + x)m (1 + x)n = (1 + x)m+n , giving the Vandermonde
Convolution
k
! ! !
X n m m+n
= . (4.26)
i=0
i k−i k
4.7. LOG, EXP AND BINOMIAL POWER SERIES 131
so
X (x + y)n n
XX xi y n−i
exp(x + y) = = ,
n≥0 n! n≥0 i=0 i! (n − i)!
so
But we can’t find a nice closed form for G(x). On the other hand,
P (n, r) = C(n, r)r!, so the equation for C(x) can be written
P (n, 0)x0 /0! + P (n, 1)x1 /1! + P (n, 2)x2 /2! + · · · + P (n, n)xn /n! = (1 + x)n ,
i.e.,
n
P (n, k)xk /k! = (1 + x)n .
X
k=0
= O1i1 · · · Okik (xi1 +···+ik )/(i1 ! · · · ik !) = O1i1 · · · Okii (xn /n!)(n!/(i1 ! · · · ik !))
n=0 n=0
4.9. FAMOUS EXAMPLE: BERNOULLI NUMBERS 135
k k ∞
! !
x k k jx k−j k
(−1)k−j (jx)n /n!
X X X
(e − 1) = e (−1) =
j=0
j j=0
j n=0
∞ k
!
k
(−1)k−j j n xn /n!
X X
=
n=0 j=0
j
is
k
!
k
(−1)k−j j n .
X
j=0
j
This proves the strange result:
The number of permutations of n objects of k types, each type appearing
at least once is
k
!
k
(−1)k−j j n .
X
j=0
j
egf
Def. The symbol f ↔ {an }∞ 0 means that the power series f is the expo-
n
∞
nential generating function of the sequence {an }0 , i.e., that f = n≥0 an xn! .
P
egf P∞ n−1 P∞ n
So suppose f ↔ {an }∞ 0
0 . Then f = n=1
x
an (n−1)! = n=0 an+1 xn! , i.e.,
egf
f 0 ↔ {an+1 }∞
0 . By induction we have an analogue to Rule 1:
egf egf
Rule 10 : If f ↔ {an }∞ h ∞
0 , then for h ∈ N , D f ↔ {an+h }n=0 .
∞ ∞
xk xk
! !
X X
1= Bk .
k=0 (k + 1)! k=0 k!
Recursively we can solve for the Bk using this equation. But first notice
the following: Replace x by −x in the egf for Bn :
∞
X (−x)k −x xex
Bk = −x = x .
k=0 k! e −1 e −1
So
∞
xex 1 − (−1)k k
" #
x X
− = −x = Bk x .
ex − 1 ex − 1 k=0 k!
This implies that
2 2
−x = B0 · 0 + B1 · x + B2 · 0 · x2 + B3 · x3 + B4 · 0 · x4 + · · ·
1 3!
which implies that
1
B1 = − and B2k+1 = 0 for k ≥ 1.
2
Then recursively from above we find B0 = 1; B1 = − 12 ; B2 = 16 ; B4 = − 30
1
;
1
B6 = 42 , . . . ,.
A famous result of Euler is the following:
∞
1 (−1)k π 2k · 22k−1 −B2k
X
ζ(2k) = = , k = 1, 2, . . . .
n=1 n2k (2k − 1)! 2k
In particular,
π2
ζ(2) =
.
6
Bernoulli originally introduced the Bn to give a closed form formula for
Sn (m) = 1n + 2n + 3n + · · · + mn .
On the one hand
4.10. FAMOUS EXAMPLE: FIBONACCI NUMBERS 137
∞ ∞
!
x(emx − 1) x xk X mj xj
mx
X
= (e − 1 = Bk =
ex − 1 ex − 1 k=0 k! j=1 j!
∞ X ∞
"m n
mi n X n
# ! !
X Bn−i X n i x
= · x = n=0 Bn−i m .
n=0 i=1 (n − i)! i! i=1 i n!
0
(The coefficient on x0! is 0.)
On the other hand:
x(emx − 1) emx − 1
x
= x x
= x(e(m−1)x + e(m−2)x + · · · + ex + 1) =
e −1 e −1
m−1 ∞ r r ∞
j x xr+1 m−1
jr =
X X X X
=x = r=0
j=0 r−0 r! r! j=0
∞ ∞
X xr+1 X nxn
= Sr (m − 1) = Sn−1 (m − 1) .
r=0 r! n=1 n!
xn
Equating the coefficients of n!
we get:
n
!
n
Bn−i mi = Sn−1 (m − 1)n, n ≥ 1,
X
i=1 i
or
n+1
!
n+1
Bn+1−i (m + 1)i = Sn (m) · (n + 1), n ≥ 0.
X
i=1 i
So Bernoulli’s formula is:
n+1
(m + 1)i
!
n n n
X n+1
Sn (m) = 1 + 2 + · · · + m = i=1 Bn+1−i .
i (n + 1)
egf
f 00 ↔ {Fn+2 }∞ n=0 . Use the recursion given in the form Fn+2 = Fn+1 + Fn ,
n ≥ 0. So by Rule 10 we have f 00 = f 0 + f . From the theory√of differential
equations we see that f (x) = c1 er+ x + c2 er− x , where r± = 1±2 5 , and where
c1 and c2 are to be determined by the initial conditions: f (0) = F0 = 1, and
f 0 (0) = 1!F1 = 1. Then f (0) = c1 + c2 = 1 and f 0 (0) = r+ c1 + r− c2 = 1. So
! ! !
1 1 1 c1
= ⇒
1 r+ r − c2
1 1
√
−1− 5
√
1 r− r− − 1
2 1 + 5 r+
c1 = = = √ = √ =√ .
1 1 r− − r + − 5 2 5 5
r+ r −
Similarly,
11
r+
1 −r−
c2 = = √ .
r − − r+ 5
P∞ n P∞ n
√1 (r+ er+ x − r− er− x ) = √1 nx nx
So f = 5 5
r+ n=0 (r+ ) n! − r− n=0 (r− ) n! .
Then
xn 1
Fn = f = √ (r+ n+1 − r− n+1 ). (4.33)
n! 5
since (α−n g(x) − 1)i ∈ R[[x]] with o((α−n g(x) − 1)i ) ≥ 1. This is a solution
since
from Eq. 4.30 and since f (0) = α. To extablish uniqueness, suppose that f
and h are both solutions to Eq. 4.34, so that
The quotient field of R[[x]] may be identified with the set R((x)) of so-
called Laurent series f (x) = ∞ n
n=k an x , where k ∈ Z is the order o(f (x))
P
f −1 = x−k g −1 .
Since the usual quotient formula for derivatives of power series holds, it is
straightforward to carry the theory of the derivative over to Laurent series.
The following facts are then easily proved (Exercises!):
!
1
cn = Res . (4.36)
nW n (x)
Proof: From our computations above we see that c1 = w1−1 . Now apply
formal derivation to Z(W (x)) = x. This yields:
∞
kck W k−1 (x)W 0 (x).
X
1= (4.37)
k=1
4.12. LAURENT SERIES AND LAGRANGE INVERSION 141
If n 6= k, then the term W k−1−n (x)W 0 (x) is a derivative by the chain rule
and hence has residue 0 by (R1). Now apply (R2) to the term with n = k
to see that the residue of the R.H.S. is equal to cn · o(W (x)) = cn , proving
the theroem.
Practice using the previous theorem on the following exercises:
Before proving our next result we need to recall the so-called Rule of
Leibniz:
Exercise: 4.12.4 Let D denote the derivative operator, and for a function
f (in our case a formal power series or Laurent series) let f (j) denote the
jth derivative of f , i.e., Dj (f ) = f (j) .
(i) Prove that
n
!
n n
f (i) g (n−i) .
X
D (f · g) =
i=0 i
142CHAPTER 4. FORMAL POWER SERIES AS GENERATING FUNCTIONS
f n (z)
!
1 n−1 n
cn = Res = D f (0).
nz n n!
P∞
Proof: Since f (z) = i=0 fi z i with f0 6= 0, in C[[z]], W (z) = P∞z fi zi =
i=0
1
= (f0 z −1 + f1 + f2 z + f3 z 2 + · · ·)−1 = ∞ n
n=1 wn z . Here f0 6= 0
P
P∞
f z i−1
i=0 i P∞
implies that w1 6= 0. By Theorem 4.12.2, z = n=1 cn wn , where
f n (z)
!
1 1 n−1 n 1X
cn = Res zn = Res = [z ]f (z) = fi1 · · · fin .
n f n (z) nz n n n
i=1 i≥1
Similarly, let
j
b−i x−i + a0 + bi x i ;
X X
g(x) =
i=1 i≥1
j
0
−ib−i x−i−1 + 0 + ibi xi−1 .
X X
g (x) =
i=1 i≥1
j k
−1 0
X X
[x ] {f (x)g (x)} = ia−i bi + −iai b−i ;
i=1 i=1
j k
[x−1 ] {f 0 (x)g(x)} = iai b−i = −[x−1 ] {f (x)g 0 (x)} .
X X
−ia−i bi +
i=1 i=1
Note that neither a0 nor b0 affects this value. Hence we may write for
f, g ∈ R((x)),
( " !#)
−1
n
0 −1
o φ(w)
−1 0
[w ] f (w) · φ (w) · φ (w) = −[w ] f (w) log . (4.39)
φ(0)
Theorem 4.12.6 (Residue Composition) Let f (x), r(x) ∈ C((x)) and sup-
pose that α = o(r(x)) > 0. We want to make the substitution x = r(z).
n −1 n 0
Proof: First consider
n f (x) = o x , −1 6= n ∈ Z. Then [z ]{r (z)r (z)}
= (n + 1)−1 [z −1 ] dz d
rn+1 (z) = 0 by (R1), since rn+1 (z) ∈ C((z)). Also,
α[x−1 ]xn = 0.
On the other hand, if n = −1, then [z −1 ]r0 r−1 = o(r(z)) = α > 0. It
follows that for all integers n, [z −1 ]{rn (z)r0 (z)} = αδn,−1 = α[x−1 ]{xn }. Now
let f (x) = n≥k an xn (o(f (x)) = k < ∞). Since o(r(z)) > 0, f (r(z)) exists,
P
and we have
α[x−1 ]{f (x)} = [z −1 ]{ an rn (z)r0 (z)} = [z −1 ]{f (r(z))r0 (z)}.
P
n≥k
Pn
2. f ((1 + y)1/2 ) = k=0 (1 + y)k [x2k ]{f (x)}.
4.12. LAURENT SERIES AND LAGRANGE INVERSION 145
( !)
2n j Pn k 2n + 1
3. [y ] (1 + y) k=0 (1 + y) =
2k + 1
! !
P 2n + 1 j+k
k = S. (Hint: At one stage you will have to use
2k + 1 2n
! ! !
P a b a+b
the fact that m = for appropriate choices of
m n−m n
a, b, m, n. You might want to prove this as a separate step if you have not
already done so.)
So at this stage we have
n
( )
2n j k 2k
X
S = [y ] (1 + y) (1 + y) [x ]f (x) =
k=0
n o
= [y −1 ] y −(2n+1) (1 + y)j f ((1 + y)1/2 ) .
At this point we want to use Residue Composition using the substitution
y = y(z) = z 2 (z 2 −2), so o(y(z)) = 2, and y 0 (z) = 4z(z 2 −1). Also, (1+y)1/2 =
1
(1 − z 2 ). Now use f ((1 + y)1/2 ) = f (1 − z 2 ) = 2(1−z 2 2n+1
2 ) {(2 − z ) − z 4n+2 }
and Residue Composition to obtain S =
which simplifies to
( ! )
1 1
[z −1 ] (z 2 − 1)2j 2 2n+1
+ 4n+2 z =
(z − 2) z
n o
[z −1 ] (z 2 − 1)2j z −(4n+1) + 0,
1 2 2j
since (z2 −2) 2n+1 is a power series, so when multiplied by (z −1) it contributes
−1
nothing to [z ]. Hence
!
4n
n
2 2j
o 2j
S = [z ] (z − 1) = .
2n
146CHAPTER 4. FORMAL POWER SERIES AS GENERATING FUNCTIONS
)−1
w · φ0 (w)
(
cn tn , where cn = [λn ]{F (λ)φn (λ)}.
X
F (w) 1 − =
φ(w) n≥0
w
Proof: Let Φ(w) = φ(w) , so t = Φ(w) and o(Φ(w)) = 1, which implies that
Φ[−1] (λ) exists. Here w = Φ[−1] (t) is the unique solution w of w = tφ(w). For
any integer n : [tn ]{f (W (t))} = [t−1 ]{t−(n+1) f (Φ[−1] (t))}. Now use Residue
Composition to substitute t = Φ(w) with α = 1 = o(Φ), and f (x) of the
Residue Composition theorem is now
t−(n+1) f (Φ[−1] (t)). Hence
1
[tn ]{f (W (t))} = − [w−1 ]{f (w)(Φ−n (w))0 } =
n
φn (w)
( )
1 1
= [w−1 ]{f 0 (w)Φ−n (w)} = [w−1 ] f 0 (w) · .
n n wn
φ(w)−wφ0 (
n
If n = 0, [tn ]{f (w)} = [t0 ]f (w) = [w−1 ]{Φ−1 (w)f (w)Φ0 (w)} = [w−1 ] f (w) φ(w)
w φ2 (w)
φ0 (w)
[w−1 ]{ f (w)
w
− f (w) }=
nφ(w) o
−1 0 φ(w)
[w0 ]{f (w)} + [w ] f (w)log φ(0)
.
This completes the proof of 1. Now let F (λ)
Rw
∈ C[[λ]]. It follows that
F (λ)φ (λ) ∈ C[[λ]]. Hence we may put f (w) = 0 F (λ)φ−1 (λ)dλ and know
−1
4.12. LAURENT SERIES AND LAGRANGE INVERSION 147
that f (w) ∈ C[[w]]. Also, since f 0 (λ) = F (λ)φ−1 (λ), we see that F (w) =
f 0 (w)φ(w). By 1., f (w) = f (0) + n≥1 n1 [λn−1 ]{φn (λ)}f 0 (λ)}tn . Differentiate
P
−1 −1
f 0 (w)φ(w)[1 − tφ0 (w)] = F (w)[1 − tφ0 (w)] =
[λn ]{φn+1 (λ)f 0 (λ)}tn = [λn ]{φn (λ)F (λ)}tn .
X X
=
n≥0 n≥0
F (W (t))
[λn ]{φn (λ)F (λ)}tn .
X
0
=
1 − tφ (W (t)) n≥0
The following example illustrates the above ideas and gives some idea of
the power of the method.
Example of Inversion Formula Suppose that for all n ∈ Z we have
the following relation:
!
X k
bn = ak . (4.40)
k
n−k
The latter formula says nothing for n = 0, but the former says that
a0 = b0 . Multiply Eq. 4.40 by wn and sum over n:
148CHAPTER 4. FORMAL POWER SERIES AS GENERATING FUNCTIONS
! !
n k
ak w n =
X X X
B(w) = bn w =
n n k
n−k
! !
k n k k
wn−k =
X X X X
ak w = ak w
k n n−k k n n−k
k
!
k k
wn = ak wk (1 + w)k =
X X X
= ak w
k n=0
n k
2
= A(w + w ) = A(t),
1
where we have put t = w + w2 = w(1 + w), or w = t · 1+w . So in the
1
notation of the Theorem of Lagrange, φ(w) = 1+w . So if w = W (t) we want
to find A(t) = B(W (t)), i.e., we want the coefficients of A in terms of the
coefficients of B: an tn = bk (W (t))k . At this stage we can say:
P P
k k
n≥0
cn tn = (1 − 2t − 3t2 )−1/2 .
X
n≥0
P 1·3·5···(2i−1)3 n−i
Exercise: 4.12.8 Show that cn = n2 ≤i≤n (n−i)!(2i−n)!2 n−i =
! !
P n i
n
≤i≤n . (Hint: Remember that you now have cn described
2 i n−i
in two different ways as a coefficient of a certain term in a power series
expansion of some ordinary generating function.)
∞ n
xn
! !
X X n
(A · B)(x) = ak bn−k = A(x) · B(x).
n=0 k=0 k n!
∞ n ∞ n
(−1)k (−1)k xn
! !
· 1 xn =
X X X X
= n! .
n=0 k=0 k! n=0 k=0 k! n!
n
X (−1)k
dn = n! .
k=0 k!
∞ ∞
X x2k X x2k
= (2k − 1)!! =
k=0 (2k)! k=0 2k k!
k
∞ x2
X 2 x2
= =e2.
k=0 k!
1 1
(P · S)(x) = P (x) · S(x) = exp( x2 ) · exp(x) = exp(x + x2 ).
2 2
We can also obtain the same result by using a recursion relation. Denote
the structure P ·S by B. In the set {1, . . . , n} we can either let n be a singleton
or make a pair {x, n} with 1 ≤ x ≤ n − 1. So bn = bn−1 + (n − 1)bn−2 , n ≥ 1.
As b1 = 1 by definition, and b1 = b0 according to the recursion, it must be
xn−1
that b0 = 1. Multiply the recursion by (n−1)! for n ≥ 1 and sum.
∞ ∞ ∞
X xn−1 X xn−1 X xn−1
bn = bn−1 + (n − 1)bn−2 .
n=1 (n − 1)! n=1 (n − 1)! n=1 (n − 1)!
P∞ xn
Also B(x) = n=0 bn n! implies
∞
xn−1
B 0 (x) =
X
bn .
n=1 (n − 1)!
xn−1
Multiply Eq. 4.42 by (n−1)!
and sum over n ≥ k:
n n−1
S(n, k) xn! . Then Fk0 (x) = x
P P
Put Fk (x) = n≥k n≥k S(n, k) (n−1)! and
X xn−1 X xn−1 X xn
k(S(n − 1, k)) =k· S(n − 1, k) =k S(n, k) .
n≥k (n − 1)! n≥k+1 (n − 1)! n≥k n!
Also
X xn−1 X xn−1
S(n − 1, k − 1) = S(n − 1, k − 1) = Fk−1 (x).
n≥k (n − 1)! n−1≥k−1 (n − 1)!
Theorem 4.13.1
xn 1
= (ex − 1)k .
X
S(n, k)
n≥k n! k!
n
Proof: For n ≥ 1, S(n, 1) = 1. And n≥1 1 · xn! = 1!1 (ex − 1)1 . So the
P
1
(ex − 1)k−1 = (k−1)!
(k−1)!
1
(ex − 1k−1 [ex − 1 + 1] = G0k (x). This is enough
to guarantee that Fk (x) = Gk (x).
Let n be a positive integer. For each k-tuple (b1 , . . . , bk ) of nonnegative
integers for which b1 + 2b2 + · · · + kbk = n, we count how many ways there
are to partition an n-set into b1 parts of size 1, b2 parts of size 2, . . . , bk parts
of size k. Imagine the elements of the n-set are to be placed in n positions.
The positions are grouped from left to right in bunches. The first b1 bunches
have one position each; the second group of b2 bunches have b2 blocks of size
2, etc. There are n! ways to order the integers in the positions. Within each
grouping of k positions there are k! ways to permute the integers within those
positions. So we divide by (1!)b1 (2!)b2 · · · (k!)bk . But the groups of the same
cardinality can be permuted without affecting the partition. So we divide by
b1 !b2 ! · · · bk !. Hence the number of partitions is:
n!
.
b1 ! · · · bk !(1!)b1 · · · (k!)bk
Now suppose that each j-set can have nj “structures” of type N on it. So
each partition gives (n1 )b1 · · · (nk )bk configurations. Hence the total number
of such configurations is
b1 bk
n! n1 nk
· ··· .
b1 ! · · · bk ! 1! k!
b1 bk
n! n1 nk
X
an = · ··· ,
b1 ! · · · bk ! 1! k!
b1 bk
1 n1 nk
X
··· ,
b1 ! · · · bk ! 1! k!
154CHAPTER 4. FORMAL POWER SERIES AS GENERATING FUNCTIONS
The sum is over all k ≥ 0, and over all (b1 , . . . , bk ) with bi ≥ 0, ibi = n,
P
P
bi = m. Now sum over all m. (Of course the contribution is zero unless
xn
m ≤ n.) It is clear that A(x) = ∞
P
n=0 an n! = exp(N (x)), and we have proved
the following theorem.
Exercise: 4.13.3 A directed tree with all edges pointing toward one vertex
xn
called the root is called an arborescence. Let T (x) = ∞
P
n=1 tn n! , where tn
4.14. DIRICHLET SERIES - THE FORMAL THEORY 155
n
is the number of labeled trees on n vertices. And let A(x) = ∞ x P
n=1 an n! ,
where an is the number of arborescences on n vertices. Since a labeled tree
on n vertices can be rooted in n ways and turned into an arborescence, and
the process is reversible, clearly an = ntn , i.e., A(x) = xT 0 (x). Consider a
labeled tree on n + 1 vertices as an arborescence with vertex n + 1 as its root.
Then delete the root and all incident edges. The result is a “rooted forest”
on n vertices, with the roots of the individual trees being exactly the vertices
xn
that were originally adjacent to the root n + 1. If F (x) = ∞
P
n=1 fn n! , where
fn is the number of rooted forests on n vertices (and f0 = 1 by convention),
then by Theorem 4.13.2, exp(A(x)) = F (x). Hence we have
∞ ∞
xn xn
= T 0 (x) = x−1 A(x).
X X
exp(A(x)) = fn = tn+1
n=0 n! n=0 n!
P∞ n
Use the special case of Lagrange Inversion to find cn if A(x) = n=1 cn x =
P∞ xn
n=1 an n! , and complete another proof of Cayley’s Theorem.
Dsf g
nP o∞
Rule 100 A(s)B(s) ↔ d|n ad bn/d .
n=1
Dsf g
nP o∞
Rule 200 A(s)k ↔ (n1 ,...,nk ):n1 ···nk =n an1 an2 · · · ank .
n=1
A most famous example is given by the Riemann zeta function
156CHAPTER 4. FORMAL POWER SERIES AS GENERATING FUNCTIONS
∞
1 Dsf g
↔ {1}∞
X
ζ(s) = n=1 .
n=1 ns
∞ ∞
µ(pi )
!
µ(n) Y
1 − p−s = ζ −1 (s).
X Y X
L(µ, s) = = = (4.45)
n=1 ns p
i=0 pis p
In other words,
1 Dsf g
↔ {µ(n)}∞
n=1 . (4.46)
ζ(s)
In the present context we give another proof of the usual Möbius Inversion
Formula.
Dsf g Dsf g
Proof: suppose F (s) ↔ {F (n)}, f (s) ↔ {f (n)}. Then
1, so τ = U ∗ U is multiplicative, and
P
τ (n) = d|n
ζ (s) = n ( d|n 1 · 1) n1s = n τn(n)
2 P P P
s .
Similarly,
X nq q P q
X 1 X X n 1 X d|n d
ζ(s) · ζ(s − q) = · = 1· = .
ns ns n d|n
d ns n ns
Example 4.14.3
∞
1 0 X −(ns )0 −ns log(n)
0
X
ζ (s) = = = =
n=1 ns n2s n2s
X log(n) Dsf g
− ⇒ ζ 0 (s) ↔ {−log(n)}.
ns
X
φ(d) = n
d|n
n X n
X
φ(n) = µ(d) · = d·µ ,
d|n
d d|n
d
ζ(s − 1)
= L(φ, s).
ζ(s)
f (p) f (p2 )
! !
X f (n) Y Y 1
= 1 + s + 2s + · · · = p 1 + s =
ns p p p p
1
1−
!
Y p2s Y 1 Y 1
= 1 − 2s · p1 − 1
.
p
1 − p1s p p ps
Also,
4.15. RATIONAL GENERATING FUNCTIONS 159
!
−1
Y 1
(ζ(2s)) = 1 − 2s .
p p
Hence,
X |µ(n)| ζ(s)
= .
ns ζ(2s)
Example 4.14.6
!
1 1 µ(n)
X X
1 = ζ(s) · = s
=
ζ(s) n ns
(
n 1 1, n = 1;
X X X
1·µ · ⇒ µ(d) =
n d|n
d s n d|n
0, n > 1.
Example 4.14.7
1 n
· ζ 2 (s) ⇒ 1 =
X
ζ(s) = µ(d) · τ .
ζ(s) d|n
d
P
This also follows from doing Möbius inversion on τ (n) = d|n 1.
un xn
X
U (x) =
n≥0
p(x)
U (x) = .
q(x)
Here we assume q(0) 6= 0, so q(x)−1 exists in C[[x]]. Before considering the
connection between rational generating functions and homogeneous linear
recursions, we recall the notion of reverse of a polynomial.
Let f (x) = an + an−1 x + an−2 x2 + · · · + a0 xn ∈ C[x]. The reverse fˆ(x) of
f (x) is defined by
1
fˆ(x) = xn f ( ) = a0 + a1 x + · · · + an xn .
x
If n0 is the multiplicity of 0 as a zero of f (x), i.e., an = an−1 = · · · =
an−n0 +1 = 0, but an−n0 6= 0, and if w1 , . . . , wq are the nonzero roots of
f (x) = 0, then w11 , . . . , w1q are the roots of fˆ(x) = 0, and fˆ(x) = a0 (1 −
w1 x) · · · (1 − wq x). So deg(fˆ(x)) = n − n0 .
Alternatively, if f (x) = (x−α1 )m1 · · · (x−αs )ms , where m1 +· · ·+ms = n
and α1 , . . . , αs are distinct, then
1
fˆ(x) = xn f ( ) = (1 − α1 x)m1 · · · (1 − αs x)ms .
x
ˆ
If a0 · an 6= 0, so neither f (x) nor fˆ(x) has x = 0 as a zero, thenfˆ = f ,
and f (α) = 0 if and only if fˆ(α−1 ) = 0.
Suppose that U (x) = n≥0 un xn = p(x)
P
q(x)
, where deg(p(x)) < deg(q(x)),
is a rational generating function. We assume q(0) 6= 0 in order that q(x)−1
exist in C[[x]], so we may assume without loss of generality that q(0) = 1.
Hence q(x) = 1 + a1 + a2 x2 + · · · + ak xk , p(x) = p0 + p1 x + · · · + pd xd , d < k.
From this it follows that
p0 + · · · pd xd = (1 + a1 x + · · · + ak xk )(u0 + u1 x1 + · · · + un xn + · · ·).
Theorem 4.15.2 The ordinary generating function for the sequence{un } de-
fined by [HLR] is
∞
un xn = R(x)/(1 + a1 x + · · · + ak xk ),
X
U (x) =
n=0
Proof:
Consider the product:
(1 + a1 x + · · · + ak xk )(u0 + u1 x + · · ·).
The coefficient on xn+k is
And this equals 0 for n ≥ 0 by [HLR], so the only coefficients that are possibly
nonzero in the product are those on 1, x, . . . , xk−1 .
Note that the coefficients of R(x) may be obtained from multiplying out
the two factors (just as we did above):
Theorem 4.15.3 Suppose (un ) is given by [HLR] and the auxiliary polyno-
mial has the form
Proof: By the theory of partial fractions, U (x) can be written as the sum
of s expressions of the form:
! ! !
1+k−1 k 2+k−1 k m+k−1
γ1 α + γ2 α + · · · + γm αk
k k k
" ! ! !#
k k+1 m+k−1
= γ1 + γ2 + · · · + γm αk = P (k)αk .
0 1 m−1
4.15. RATIONAL GENERATING FUNCTIONS 163
The formula
!
k+l
= (k + l)(k + l − 1) · · · (k + 1)/l(l − 1) · · · 1
l
!
k+l
shows that is a polynomial in k with degree l. Hence P (k) is a
l
polynomial in k with degree at most m − 1. The theorem follows.
In practice we assume the form of the result for un and obtain the co-
efficients of the polynomials Pi (n) by substituting in the initial values of
u0 , u1 , . . . , uk−1 and solving k equations in k unknowns.
2 2
(α1 )i xi + (α2 )i xi .
X X
F (x) = √ √
5− 5 i 5+ 5 i
Hence
√ n √ n
2(1 + 5) 2(1 − 5)
Fn = [xn ]{F (x)} = √
n
+ √ .
(5 − 5)2 (5 + 5)2n
u0 = −1, u1 = +1, u2 = 0, u3 = 1.
Find a formula for un . Also find the generating function for the sequence
{un }∞
n=0 .
164CHAPTER 4. FORMAL POWER SERIES AS GENERATING FUNCTIONS
P k
P n−k n
1
o
Theorem 4.16.2 k n−k
xn−k = k k
xk = [y n ] 1−y−xy 2
.
P n−k
Proof: For n ≥ 0, put fn (x) = k k
xk (0 ≤ k ≤ n2 ). We claim that
fn+2 (x) = xfn (x) + fn+1 (x). For,
! !
X n−k k X n+1−k k
x x + x =
0≤k≤ n
k 0≤k≤ n+1 k
2 2
! ! !
X n − k k+1 n+1 X n+1−k k
= x + + x =
0≤k≤ n
k 0 1≤k≤ n+1
k
2 2
! ! !
X n−t+1 t n+2 X n+1−t t
= x + x
1≤t≤ n+1
t−1 0 1≤t≤ n+1
t
2 2
n+2
! ! !
n+2 X n+2−t t n− 2
+1 n+2
= + x n x 2
0 1≤t≤ n+1
t 2
2
4.16. MORE PRACTICE WITH GENERATING FUNCTIONS 165
!
X n+2−t t
= x = fn+2 (x).
0≤t≤ n+2
t
2
Exercise: 4.16.3 (Ex. 10F, p.77 of van Lint & Wilson) Show that
n
!
X
k 2n − k 2n−2k
(−1) 2 = 2n + 1.
k=0 k
P n−k
Exercise: 4.16.4 Evaluate the sum k k
(−1)k .
P n−k
Exercise: 4.16.5 Evaluate k k
.
Put
X bkc X X bkc
! ! !
∗ k n n
T (x, y) = 2 x y = 2 y xk
k,n n k n n
k
(1 + y)b 2 c xk = (1 + x) + (1 + y)(x2 + x3 ) + (1 + y)2 (x4 + x5 ) + · · ·
X
=
k
∞
1+x
(1 + y)i x2i =
X
= (1 + x) .
i=0 1 − (1 + y)x2
0 ∗ 1
Note that [y ]T (x, y) = 1−x
.
1−x2 −1+x2 +yx2
Now put T (x, y) = T ∗ (x, y) − 1
1−x
= 1+x
1−(1+y)x2
− 1
1−x
= (1−x)(1−(1+y)x2 )
=
x2 y
(1−x)(1−(1+y)x2 )
.
k
b2c
n
, n ≥ 1;
x2 y
n o
So [xk y n ] (1−x)(1−(1+y)x2 )
=
0, n = 0.
x2 y
Hence T (x, y) = is the generating function for the doubly-
(1−x)(1−(1+y)x2 )
infinite sequence of terms on the R.H.S. of Eq. 4.48.
Put ! !
X k j−n−1
S(x, y) = (−2)j−2n xk y n .
k,n,j j j − 2n
Then [xk y n ] {S(x, y)} is the desired sum (on the L.H.S. of Eq. 4.49).
Hence our task is equivalent to showing that S(x, y) = T (x, y).
Make!the invertible !substitution
! (change of variables):
t 1 −2 j
= , i.e., t = j − 2n, s = j − n, with inverse
s 1 −1 n
j = 2s − t, n = s − t. Hence we have
! !
k s−1
(−2)t xk y s−t
X
S(x, y) =
k,s,t 2s − t t
4.17. THE TRANSFER MATRIX METHOD 167
! ! !!
k s−1
xk y s−t (−2)t
X X X
=
s t k 2s − t t
(now use Theorem 4.16.1)
x2s−t
! !!
s−1
y s−t (−2)t
X X
= 2s−t+1
s t (1 − x) t
!t
x2s y s
!
X X s−1 (1 − x)(−2)
=
s t t xy (1 − x)2s+1
!s−1 !s
X
1−
2(1 − x) x2 y 1
=
s≥1 xy (1 − x)2 1−x
!j !j+1
X xy − 2(1 − x) x2 y 1
j≥0 xy (1 − x)2 1−x
!j
x2 y X (xy − 2(1 − x))x x2 y 1
= = · xy−2(1−x))x
(1 − x)3 j≥0 (1 − x)2 (1 − x)3 1 − (1−x)2
x2 y x2 y
= =
(1 − x)(1 − 2x + x2 − x2 y + 2x − 2x2 ) (1 − x)(1 − x2 − x2 y)
= T (x, y).
Lemma 4.17.1 If f (λ) = det(λI − A), then fˆ(λ) = det(I − λA). Moreover,
ˆ
if A is invertible, so n0 = 0, then fˆ = f , and f (λ) = det(λI − A) iff
fˆ(λ) = det(I − λA).
(An )ij λn .
X
Fij (A, λ) = (4.50)
n≥0
w1 λ w2 λ wq λ
= + +
1 − w1 λ 1 − w2 λ 1 − wq λ
4.17. THE TRANSFER MATRIX METHOD 169
q X
∞ ∞ q ! ∞
win λn win λn = tr(An )λn .
X X X X
= =
i=1 n=1 n=1 i=1 n=1
We have proved the following corollary:
P∞ −λq 0 (λ)
Corollary 4.17.4 If q(λ) = det(I − λA), then n=1 tr(An )λn = q(λ)
.
Let D = (V, E, φ) be a finite digraph, where V = {v1 , . . . , vp } is the set of
vertices, E is a set of (directed) edges or arcs, and φ : E → V × V determines
the edges. If φ(e) = (u, v), then e is an edge from u to v, with initial vertex
int(e) = u and final vertex f in(e) = v. If u = v, then e is a loop. A walk
Γ in D of length n from u to v is a sequence e1 e2 · · · en of n edges such that
int(e1 ) = u, f in(en ) = v, and f in(ei ) = int(ei+1 ) for 1 ≤ i < n. If also
u = v, then Γ is called a closed walk based at u. (Note: If Γ is a closed walk,
then ei ei+1 · · · en e1 · · · ei−1 is in general a different closed walk.)
Now let w : E → R be a weight function on E (R is some commutative
ring; usually R = C or R = C[x].) If Γ = e1 e2 · · · en is a walk, then the weight
of Γ is defined by w(Γ) = w(e1 )w(e2 ) · · · w(en ). Fix i and j, 1 ≤ i, j ≤ p.
P
Put Aij (n) = Γ w(Γ), where the sum is over all walks Γ in D of length n
from vi to vj . In particular, Aij (0) = δij . The fundamental problem treated
by the transfer matrix method (TMM) is the evaluation of Aij (n), or at least
the determination of some generating function for the Aij (n).
Define a p × p matrix A = (Aij ) by
X
Aij = w(e),
e
where the sum is over all edges with int(e) = vi and f in(e) = vj . So
Aij = Aij (1). A is the adjacency matrix of D with respect to the weight
function w.
Theorem 4.17.5 Let n ∈ N . Then the (i, j)-entry of An is equal to Aij (n).
(By convention, A0 = I even if A is not invertible.)
Proof: (An )ij = Aii1 Ai1 i2 · · · Ain−1 j , where the sum is over all sequences
P
−λq 0 (λ)
CD (n)λn = where q(λ) = det(I − λA).
P
Corollary 4.17.6 n≥1 q(λ)
,
subject to the restrictions used in defining D. Put q(λ) = det(I − λA) and
qij (λ) = det((I − λA) : j, i). Then by Theorem 4.17.2
p
f (n + 1)λn = Aij (n) λn
X X X
F (λ) := (4.53)
n≥0 n≥0 i,j=1
p X p p
(−1)i+j qij (λ)
Aij (n)λn =
X X X
= Fij (A, λ) = .
i,j=1 n≥0 i,j=1 i,j=1 q(λ)
We state this as a corollary.
Corollary 4.17.7 If w(e) = 1 for all edges in D and f (n) is the number
of sequences a1 a2 · · · an ∈ [p]n subject to the restrictions used in defining D,
then
p
(−1)i+j qij (λ)
f (n + 1)λn =
X X
. (4.54)
n≥0 i,j=1 q(λ)
1 1 0
4.17. THE TRANSFER MATRIX METHOD 171
if and only if
√ ! √ !n √ ! √ !n
5−2 1− 5 5+2 1+ 5
f (n + 1) = √ + √ .
5 2 5 2
172CHAPTER 4. FORMAL POWER SERIES AS GENERATING FUNCTIONS
Problem 2.2. Find the number of sequences of n + 1 0’s and 1’s with 11
never appearing as a subsequence ai ai+1 , i.e., x11 = 0 as above, but this time
consider only those sequences starting with a fixed i ∈ {0, 1} and ending
with a fixed j ∈ {0, 1}.
For this
! situation we need to find the ij entry of the nth power of A =
1 1
. Here we diagonalize the matrix A to find its powers
1 0
! 1+√5 n √ !
0
1
An = √ √ 2 √ −2 2 √ n 1 + √5 2
4 5 5−1 5+1 0 1− 5 1− 5 2
2
√ √ √ √
(1+ 5)n+1 (1− 5)n+1 (1+ 5)n (1− 5)n
1 n−1 − √ 2n−1 2n−2
− √2n−2
= √ √ 2√
( 5−1)(1+ 5)n+1
√
( 5+1)(1− 5)n+1
√ √
( 5−1)(1+ 5)n
√
( 5+1)(1− 5)n
.
4 5 2n
+ 2n 2n−1
+ 2n−1
For example the 12 entry of this matrix is the number of sequences of
n + 10’s and 1’s with 11 never appearing as a subsequence ai ai+1 and starting
with 0 and ending with 1. A little routine computation gives
√ √
n (1 + 5)n (1 − 5)n
tr(A ) = + .
2n 2n
0
We could also have used Cor. 4.17.6 and calculated −λq (λ)
q(λ)
= −2 +
2−λ 1 1
1−λ−λ2
= −2 + 1−αλ + 1−βλ . This agrees with the above for n ≥ 1, but
in the proof of Cor. 4.17.6 the term CD (0) is not accounted for.
Problem 2.3 Suppose we still require that two 1’s never appear together,
but now we want to count sequences with prescribed
! numbers of 0’s and 1’s.
x00 x01
Return to the situation where A = . Then
x10 x11
!
1 − x11 x01
∞
!−1
X
n −1 1 − x00 −x01 x10 1 − x00
A = (I − A) = =
n=0
−x10 1 − x11 (1 − x00 )(1 − x11 ) − x01 x10
!
1 − x11 x01 1 1
= · · x01 x10
x10 1 − x00 (1 − x00 )(1 − x11 1 − (1−x00 )(1−x11 )
∞
xi01 xi10
!
1 − x11 x01 X
=
x10 1 − x00 i=0 (1 − x00 )
i+1 (1 − x )i+1
11
4.17. THE TRANSFER MATRIX METHOD 173
∞ ∞ X
∞
! ! !
1 − x11 x01 i+j j i+k k
xi01 xi10
X X
= x00 x11 .
x10 1 − x00 i=0 j=0 k=0 j k
If we suppose that the pair 11 never appears, so x11 = 0, then xk11 = δk,0 .
And
∞ ∞
! !
X
n 1 x01 X i+j i i j
A = x01 x10 x00 .
n=0
x10 1 − x00 i,j=0 j
We now consider what this equation implies for the (i, j) position, 1 ≤
i, j ≤ 2.
P∞ n P∞ i+j i i j
Case 1. (1,1) position: n=0 (A )11 = i,j=0 j
x01 x10 x00 . So there
must be i+j j
ways of forming walks of length 2i + j using the edges x01 and
x10 each i times and the edge x00 j times. This corresponds to a sequence of
length 2i + j + 1 with exactly i 1’s (and i + j + 1 0’s), starting and ending
with a 0, and never having two 1’s next to each other. Another way to view
this is as needing to fill i + 1 boxes with 0’s (the boxes before and after each
1) so that each box has at least one 0. This is easily seen tobe the
same
as
i+j i+j
an (i + 1)-composition of i + j + 1, of which there are i = j . (See
pages 15-16 of our class notes.)
P∞ P∞ i+j j
Case 2. (1,2) position: n
n=0 (A )12 = i,j=0 j
xi+1 i
01 x10 x00 . Here there
must be i+j j
walks of length 2i + j + 1 using the edge x01 i + 1 times, the
edge x10 i times, and the edge x00 j times. This corresponds to a sequence
of length 2i + j + 2 with exactly i + 1 1’s (and i + j + 1 0’s), starting with
a 0 and ending with a 1, and never having two 1’s next to each other. It is
clear that this kind of sequence is just one from Case 1 with a 1 appended
at the end.
Case 3. (2,1) position: This is the same as Case 2., with the roles of x01
and x10 interchanged, and the 1 appended at the beginning of the sequence.
i+j
xi01 xi10 xj00
P∞ n P∞
Case 4. (2,2) position: n=0 (A )22 = i,j=0 j
i+j
xi01 xi10 xj+1
P∞
− i,j=0 j 00
∞ ∞
! !
i+j i+j
xi01 xi10 xj00 xi01 xi10 xj+1
X X
= − 00
i,j=0 j i,j=0 j
∞ ∞
" ! !#
i+j i+j−1
xi01 xi10 xi01 xi10 xj00
X X
= + −
i=0 i=0;j=1 j j−1
174CHAPTER 4. FORMAL POWER SERIES AS GENERATING FUNCTIONS
∞
!
i+j−1
xi01 xi10 xj00
X
=1+ ,
k,j=1 j
1 1 1
P3
i,j=1 Aij (n − 1). Put q(λ) = det(I − λA), and qij (λ) = det(I − λA : j, i).
By Theorem 4.17.2, F (λ) := n≥0 f (n + 1)λn = n≥0 ( 3i,j=1 A)ij (n)) =
P P P
P3
(−1)i+j qij (λ)
i,j=1
q(λ)
. It is easy to work out q(λ) = λ3 − λ2 − 2λ + 1. Then
det[(I − λA)−1 ] = [det(I − λA)]−1 . By Cor. 4.17.3, Fij (A, λ), and hence F (λ)
is a rational function of λ of degree less than the multiplicity n0 of 0 as an
eigenvalue of A. But q(λ) has degree 3, forcing A to have rank at least 3.
But A is 3 × 3, so n0 = 0. Since the denominator of F (λ) is q(λ), which has
degree 3, the numerator of F (λ) has degree at most 2, sois determined by
2 2 1
2
its values at three points. Note: we need A = 1 2 1 .
2 3 2
Then
3
X
f (1) = Aij (0) = tr(I) = 3.
i,j=1
3
X
f (2) = Aij (1) = 7.
i,j=1
4.17. THE TRANSFER MATRIX METHOD 175
3
X
f (3) = Aij (2) = 16.
i,j=1
2
Then for some a0 , a1 , a2 ∈ Q, F (λ) = a0 +a 1 λ+a2 λ
f (n + 1)λn =
P
1−2λ−λ2
= n≥0
2
f (1) + f (2)λ + f (3)λ + · · ·, which implies that
Hence
3 + λ − λ2
f (n + 1)λn =
X
F (λ) = ,
n≥0 1 − 2λ − λ2 + λ3
from which it follows that
X
n+1 3λ + λ2 − λ3
f (n + 1)λ = .
n≥0 1 − 2λ − λ2 + λ3
f (n + 3) = 2f (n + 2) + f (n + 1) − f (n). (4.55)
|{x1 (x2 (x3 x4 )), x1 ((x2 x3 )x4 ), (x1 x2 )(x3 x4 ), (x1 (x2 x3 ))x4 , ((x1 x2 )x3 )x4 }|.
By convention, u1 = u2 = 1.
A given associated product always looks like (x1 . . . xr )(xr+1 . . . xn ), where
1 ≤ r ≤ n − 1. So un = u1 un−1 + u2 un−2 + · · · + un−1 u1 , n ≥ 2. Hence
n−1
X
un = ui un−i .
i=1
P∞
Put f (x) = n=1 un xn . Then
∞ n−1 ∞
!
2
ui un−i xn = un xn = f (x) − x.
X X X
(f (x)) =
n=2 i=1 n=2
1
It follows that [f (x)]2 − f (x) + x = 0, so f (x) = 12 [1 ± (1 − 4x) 2 ]. We
must use the minus sign, since the constant term of f (x) is f (0) = 0. This
leads to
∞
!
1 1 1 1 1X 1
f (x) = − (1 − 4x) 2 = − 2 (−4x)n .
2 2 2 2 n=0 n
Then a little computation shows that
1 (2n − 2)!
un = (− )(−1)n−1 (1 · 3 · 5 · · · (2n − 3))(−1)n 4n ÷ 2n · n! = =
2 n!(n − 1)!
!
1 2(n − 1)
= = Cn−1 .
n n−1
1 2n
These numbers Cn = n+1 n
are the famous Catalan numbers. See
Chapter 14 of Wilson and van Lint for a great deal more about them.
4.19. MACMAHON’S MASTER THEOREM 177
recreate the digraph D along with the linear order on its arcs. To see this,
start with y1 y3 y3 y2 y1 y1 y3 and work from the left. For each j, 1 ≤ j ≤ 3, the
number of yj appearing in the word is the indegree mj of j. Since m1 = 3,
m2 = 1, and m3 = 3, the first 3 arcs have initial vertex 1, the 4th arc has
initial vertex 2, the last 3 arcs have initial vertex 3.
As another example, consider the word y2 y1 y2 y32 y12 , and let D be the
associated digraph. Here m1 = 3, m2 = 2, m3 = 2. It follows that
wt(D) = a112 a211 a312 a123 a223 a131 a231 y2 y1 y2 y32 y12 .
Two digraphs D1 and D2 in D are considered the same if and only if for
each i, 1 ≤ i ≤ n, and for each t, 1 ≤ t ≤ mi , the tth arc of D1 having initial
vertex i, and the tth arc of D2 having initial vertex i, both have the same
terminal vertex.
Consider the product
Yn
i=1
(ai1 y1 + · · · + ain yn )mi . (4.66)
Clearly,
i=1
along the arcs of D, always choosing the topmost arc( arc atij from i with t
the largest available), until one of the following occurs:
(i) We encounter a previously visited vertex (and have thus located a
directed cycle γ of D).
(ii) We encounter a vertex which has a positive outdegree in H (and thus
is a vertex on a directed cycle δ of H).
We note that if u is a vertex with positive outdegree in H then we are
immediately in case (ii). We also note that cases (i) and (ii) cannot occur
simultaneously. If case (i) occurs, we form a new element of G by removoing γ
form D and putting it in H. If case (ii) occurs, we remove δ from H and put
it in D in such a way that each arc of γ is put in front of (in the linear order)
those with the same initial vertex. Let (D0 , H 0 ) be the pair obtained in this
way. Then D0 is in D and H 0 is in H, and hence (D0 , H 0 ) is in G. Moreover,
since the number of directed cycles in H 0 differs from the number in H by
one, it follows that wt(D0 , H 0 ) = −wt(D, H). Define σ(D, H) = (D0 , H 0 ) and
note that σ(D0 , H 0 ) = (D, H). Thus σ is an involution on G \ (∅, ∅) which is
sign-reversing on weights. It follows that wt(G) = wt(∅, ∅) = 1. Hence the
proof is complete.
We give two examples to help the reader be sure that the above proof is
understood. Let D be the general digraph with arcs
det(I − AY )−1 = (1 − a11 y1 − a22 y2 + (a11 a22 − a12 a21 )y1 y2 )−1 =
! ! !
X m1 m2
ai am1 −i am 1 −i m2 −m1 +i
y1m1 y2m2 .
X
a22 (4.70)
(m1 ,m2 )≥(0,0) i i m1 − i 11 12 21
∞
[a11 y1 + a22 y2 − (a11 a22 − a12 a21 )y1 y2 ]k .
X
k=0
Then computing the coefficient of y1m1 y2m2 in this sum (and writing ∆ in
place of a11 a22 − a12 a21 ) we get
∞
!
k
ak−m a22 ∆m1 +m2 −k (−1)m1 +m2 −k .
2 k−m1
X
11
k=0 k − m 2 , k − m 1 , m 1 + m 2 − k
(4.71)
This gives a variety of equalities. In particular, suppose each aij = 1.
Hence ∆ = 0 so k = m1 + m2 for a nonzero contribution. Then the Master
Theorem yields the familiar equality:
! ! ! !
X m1 m2 m1 + m2 m1 + m2
= = . (4.72)
i i m1 − i m 1 , m2 , 0 m1
P k
Exercise: 4.19.7 Prove that k k−m,k−n,m+n−k
(−1)m+n−k = 1.
1 −1 0
x 0 0
Y = 0 y 0 . A simple calculation shows that
0 0 z
1 −y z
I − AY = x , and det(I − AY )−1 =
1 −z
−x y 1
!
i+j+k
= (1 + xy + yz + zx)−1 = (−1)i+j+k (xy)i (yz)j (zx)k . (4.73)
X
i,j,k≥0 i, j, k
4.19. MACMAHON’S MASTER THEOREM 185
This chapter deals with locally finite partially ordered sets (posets), their
incidence algebras, and Möbius inversion on these algebras.
5.1 Introduction
Recall first that we have proved the following (see Theorem 1.5.5):
n
(x)n = c(n, k)xk ,
X
(5.1)
k=0
" #
n
where c(n, k) = is the number of σ ∈ Sn with k cycles. Replacing x
k
with −x and observing that (−x)k = (−1)k (x)k , we obtained
n
s(n, k)xk ,
X
(x)k = (5.2)
k=0
203
204 CHAPTER 5. MÖBIUS INVERSION ON POSETS
n
xn =
X
S(n, k)(x)k , n ≥ 0, (5.3)
k=0
where S(n, k) is a Stirling number of the second kind. If we use the same
trick of replacing x with −x again, we get
n
xn = (−1)n−k S(n, k)(x)k .
X
(5.4)
k=0
Here we can see that Eq. 5.1 and Eq. 5.4 are “inverses” of each other, and
Eq. 5.2 and Eq. 5.3 are “inverses” of each other. We proceed to make this a
little more formal.
Let Pn be the set of all polynomials of degree k, 0 ≤ k ≤ n, (along with
the zero polynomial), with coefficients in C. Then Pn is an (n+1)-dimensional
vector space.
B1 = {1, x, x2 , . . . , xn },
B2 = {(x)0 = 1, (x)1 , . . . , (x)n }
and
B3 = {(x)0 = 1, (x)2 , . . . , (x)n }
are three ordered bases of Pn . Recall that if B = {v1 , . . . , vm } and B 0 =
{w1 , . . . , wm } are two bases of the same vector space over C (or over any field
K), then there are unique scalars aij , 1 ≤ i, j ≤ m for which wj = m
P
i=1 aij vi
and unique scalars a0ij , 1 ≤ i, j ≤ m for which vj = m 0
P
a w
i=1 ij i . And the
0 0
matrices A = (aij ) and A = (aij ) are inverses of each other.
So put:
A = (aij ), 0 ≤ i, j ≤ n aij = c(j, i);
Then A and D are inverses of each other, and B and C are inverses. So
n
X n
X
S(j, k)s(k, i) = bik ckj = (BC)ij = δij , (5.5)
k=0 k=0
n n
j−k
X X
(−1) S(j, k)c(k, i) = aik dkj = (AD)ij = δij . (5.6)
k=0 k=0
We want to see Eq. 5.5 expressed in the context of “Möbius inversion over
a finite partially ordered set.” Also, when two matrices, such as A and D
above, are recognized as being inverses of each other, b̄ = āA iff ā = b̄D.
Consider a second example. Let A, B, C be three subsets of a universal
set E. Then |E \ (A ∪ B ∪ C)| = |E| − (|A| + |B| + |C|) + (|A ∩ B| + |A ∩ C| +
|B ∩ C|) − |A ∩ B ∩ C|. This is a very special case of the general principle
of inclusion - exclusion that we met much earlier and which we now want to
view as Möbius inversion over a certain finite partially ordered set.
As a third example, recall “Möbius inversion” as we studied it earlier:
f (n) = d|n g(d) for all n ∈ N iff g(n) = d|n µ(d)f (n/d) for all n ∈ N ,
P P
5.2 POSETS
A partially ordered set P (i.e., a poset P ) is a set P together with a
relation “ ≤ ” on P for which (P, ≤) satisfies the following:
206 CHAPTER 5. MÖBIUS INVERSION ON POSETS
[x, y] = {z ∈ P : x ≤ z ≤ y},
where x ≤ y. So [x, x] = {x}, but ∅ is NOT an interval. P is called locally
finite provided |[x, y]| < ∞ whenver x, y ∈ P , x ≤ y. An element of P is
called a zero (resp., one) of P and denoted 0̂ (resp., 1̂) provided 0̂ ≤ x for
all x ∈ P (resp., x ≤ 1̂ for all x ∈ P ). Finally, we write x < y provided x ≤ y
but x 6= y.
EXAMPLES OF LOCALLY FINITE POSETS
Example 5.2.5 Let n ∈ P. The set Πn of all partitions of [n] is made into
a poset by defining π ≤ σ (for π, σ ∈ Πn ) iff each block of π is contained in
some block of σ. In that case we say π is a refinement of σ.
For example, (12)(3) ≤ (123), (1)(23) ≤ (123), but (13)(2) 6≤ (123). The
0̂ of Sn is 0̂ = (1)(2) · · · (n). As an example, for σ = (12435) ∈ S5 , we give
the Hasse diagram of the interval [0̂, σ]. Note, for example, that (12)(435) is
not in the interval since it would appear as (12)(354).
208 CHAPTER 5. MÖBIUS INVERSION ON POSETS
(12435)
h
H HH
H
HH
H
HH
H
HH
H (1)(2435)
h
(124)(35) h(1243)(5) HHh
PP
@ PP J
B
@ PP J
B
PP
@ PP J
B
PP
@ PP J
B
@ PP J
B
@ PP J
B
PP
@ J PP
B
PP
B
PP (1)(24)(35)
@ J PP
@ J
B
(12)(35)(4) @ J
PP
PP B
h
P @h(124)(3)(5) h(1)(243)(5)
J
PBh
B PPP B
"
PP "
B PP B
"
B "
B PP
"
PP "
B PP B
"
B PPB
"
"
B P
B PP
"
B B PP
"
P "
B B
P "
"P
P
B B
" PP
PP
B B
"" PP
Bh (12)(3)(4)(5) h
B
" PPh
HHH (1)(24)(3)(5)
HH
H (1)(2)(35)(4)
HH
H
HH
H
HH
H h
0̂ = (1)(2)(3)(4)(5)
Hasse diagram of interval [0̂, σ] in Sn , σ = (12435)
then it is an easy exercise to show that conditions (1) and (2) hold so that
V with this product is an algebra.
An algebra V over K is said to be associative provided its multiplication
satisfies the associative law (xy)z = x(yz) for all x, y, z ∈ V .
The algebras we study here are finite dimensional linear associative alge-
bras.
210 CHAPTER 5. MÖBIUS INVERSION ON POSETS
X
(f ∗ g)(x, y) = f (x, z)g(z, y), (5.7)
z:x≤z≤y
for all intervals [x, y] of P and all f, g ∈ I(P, K). The sum in Eq. 5.7 is finite
(so f ∗ g really is defined), since P is locally finite.
Proof:
This is a straightforward and worthwhile (but tedious!) exercise. It is
probably easier to establish associativity by showing (f ∗ g) ∗ h = f ∗ (g ∗ h)
in general than it is to establish associativity for some specific basis and
then to use Theorem 5.3.1. And to establish that δ ∗ f = f ∗ δ = f for
all f ∈ I(P, K) is almost trivial. The bilinearity conditions are also easily
established.
Note: It is quite helpful to think of I(P, K) as the set of all formal
expressions
X
f= f (x, y)[x, y] (allowing infinite linear combinations).
[x,y]∈Int(P )
5.4. THE INCIDENCE ALGEBRA I(P, K) 211
for all [x, y], [z, w] ∈ Int(P ), and then extending to all of I(P, K) by bilin-
earity. This shows that when Int(P ) is finite, one basis of I(P, K) may be
obtained by setting 1[x,y] equal to the function from Int(P ) to K defined by
(
1, if [z, w] = [x, y];
1[x,y] (z, w) =
0, if [z, w] 6= [x, y] (but [x, y], [z, w] ∈ Int(P )).
Then the set {1[x,y] : [x, y] ∈ Int(P )} is a basis for I(P, K) and the
multiplication constants are given by
(
1, y = z;
1[x,y] ∗ 1[z,w] = δy,z 1[x,w] , where δy,z = (5.8)
0, otherwise.
Exercise: 5.4.2 Show that if P is any finite poset, its elements can be labeled
as x1 , x2 , . . . , xn so that xi < xj in P implies that i < j.
So
X
(f ∗ g)([xi , xl ]) = f (xi , xj )g(xj , xl ).
xj :xi ≤xj ≤xl
x5
h
@
@
x3 h @hx4
@
@
@
@
x1 h @hx2
Theorem 5.4.3 Let f ∈ I(P, K). Then the following are equivalent:
(i) f has a left inverse;
(iv) f (x, x) 6= 0 ∀x ∈ P.
(" k # )
−1
X
g(x, y) = −(f (x, x)) f (x, zi )g(zi , y) + f (x, y)g(y, y) .
i=1
214 CHAPTER 5. MÖBIUS INVERSION ON POSETS
Now suppose that the maximum length of any chain in [x, y] is 4, say
x < w < z < y is a maximal chain in [x, y]. For each u ∈ [x, y] \ {x}, either
[u, y] = {y}, or [u, y] = {u, y}, or [u, y] = {u, w, y} for some w ∈ [x, y]. In
any case, g(u, y) is already defined for all u ∈ [x, y] \ {x}. So we may define
The zeta function is of interest in its own right (we include an optional
page dealing with ζ), but for us the main interest in ζ is that by Theorem 5.4.3
it has an inverse µ called the Möbius function of the poset P .
One can define µ inductively without reference to the incidence algebra
I(P, K). Namely, µ ∗ ζ = δ is equivalent to
and
X
µ(x, y) = − µ(x, z) whenever x < y. (5.10)
z:x≤z<y
Similarly, ζ ∗ µ = δ is equivalent to
5.5. OPTIONAL SECTION ON ζ 215
and
X
µ(x, y) = − µ(z, y) whenever x < y. (5.11)
z:x<z≤y
ζ k (x, y) =
X
1
(x0 ,...,xk ):x=x0 ≤x1 ≤···≤xk =y
Proof: Clear.
Hence for k ∈ P, (ζ − δ)k (x, y) is the number of chains x = x0 < x1 <
· · · < xk = y of length k from x to y.
(
1, if x = y;
Theorem 5.5.2 (2δ −ζ)(x, y) = So (2δ −ζ)−1 exists and
−1, if x < y.
(2δ − ζ)−1 (x, y) is equal to the total number of chains x = x0 < x1 < · · · <
xk = y from x to y.
Proof: Let l be the length of the longest chain in the interval [x, y].
Then (ζ − δ)l+1 (u, v) = 0 whenever x ≤ u ≤ v ≤ y. Thus for x ≤ u ≤
v ≤ y, (2δ − ζ)[1 + (ζ − 1) + (ζ − 1)2 + · · · + (ζ − 1)l ](u, v) = (1 − (ζ −
1))[1 + (ζ − 1) + (ζ − 1)2 + · · · + (ζ − 1)l ](u, v) = (1 − (ζ − 1)l+1 )(u, v) =
δ(u, v). Hence (2δ − ζ)−1 = 1 + (ζ − 1) + · · · + (ζ − 1)l , when restricted to the
interval [x, y]. But by the definition of l and theorem 5.5.1 it is clear that
(1 + (ζ − 1) + · · · + (ζ − 1)l )(x, y) is the total number of chains from x to y.
216 CHAPTER 5. MÖBIUS INVERSION ON POSETS
Similarly, δ · f = f .
For these to be actions in the usual sense, it must be true that the fol-
lowing hold:
and
We verify Eq. 5.14 and leave Eq. 5.15 as a similar exercise. So for each
x ∈ P,
X X
((f · ξ1 ) · ξ2 )(x) = (f · ξ1 )(y)ξ2 (y, x) = f (w)ξ1 (w, y)ξ2 (y, x) =
y≤x w,y:w≤y≤x
X X X
= f (w)( ξ1 (w, y)ξ2 (y, x)) = f (w)(ξ1 ∗ ξ2 )(w, x) =
w:w≤x y:w≤y≤x w:w≤x
= (f · (ξ1 ∗ ξ2 ))(x).
Example 5.6.2 Consider the chain P = N with the usual linear ordering,
so 0̂ = 0. Here µ(x, x) = 1 for all x ∈ P , and for x < y, µ(x, y) =
− z:x≤z<y µ(x, z). So if y covers x, µ(x, y) = −1. If y covers z and z
P
Example 5.7.2 Recall that for a positive integer k, k denotes [k] with the
usual linear order. Then let n1 , . . . , nk ∈ N and put P = 1 + n1 ×· · ·×1 + nk .
We may identify the elements of P with the set of k-tuples (a1 , . . . , ak ) ∈ N k
with 0 ≤ ai ≤ ni ordered componentwise, i.e., (a1 , . . . , ak ) ≤ (b1 , . . . , bk )
iff ai ≤ bi for i = 1, 2, . . . , k. Then if this relation holds, the interval
[(a1 , . . . , ak ), (b1 , . . . , bk )] is isomorphic to b1 − a1 + 1 × b2 − a2 + 1 × · · · ×
bk − ak + 1.
5.7. EVALUATING µ: THE PRODUCT THEOREM 219
Q
Example 5.7.3 Recall that n is the poset of partitions of [n], where two
partitions σ, π ∈ n satisfy σ ≤ π iff each block of σ is contained in a single
Q
λ2 × · · · ×
Q Q
λk .
Note that 2 ' 2 and ( 2 )k ' (2)k . Hence if σ and π are partitions of
Q Q
[n] for which π = {A1 , . . . , Ak } has k parts, each of which breaks into two
Q
parts in σ, then the interval [σ, π] in n is isomorphic to Bk .
Proof: X
µP (x, u)µQ (y, v) =
(u,v):(x,y)≤(u,v)≤(x0 ,y 0 )
X X
= µP (x, u) µQ (y, v) =
u:x≤u≤x0 v:y≤v≤y 0
Also
X
µP ×Q ((x, y), (u, v)) = δ(x,y),(x0 ,y0 ) .
(u,v):(x,y)≤(u,v)≤(x0 ,y 0 )
P
But since z:x≤z≤y µ(x, z) = δx,y in some poset with Möbius function µ
determines µ uniquely, it must be that µP ×Q = µP · µQ .
Using the product theorem we can say a great deal about the Möbius
functions of the three examples above.
220 CHAPTER 5. MÖBIUS INVERSION ON POSETS
Let f, g : Bn → K. Then
Theorem 5.7.6 P
(i) g(S) = T :T ⊆S f (T ) ∀S ∈ Bn
|S−T |
∀S ∈ Bn ;
P
iff f (S) = T :T ⊆S (−1) g(T )
and
∀S ∈ Bn
P
(ii) g(S) = T :T ⊇S f (T )
|T −S|
∀S ∈ Bn .
P
iff f (S) = T :T ⊇S (−1) g(T )
(−1)|S−T | g(S).
X
f (T ) = (5.16)
S:S⊇T
5.7. EVALUATING µ: THE PRODUCT THEOREM 221
n
X X
|E \ (A1 ∪ · · · ∪ An )| = |E| − |Ai | + |Ai ∩ Aj |+
i=1 1≤i<j≤n
n
!
X
|T |
X
i n
D(n) = (−1) g(T ) = (−1) (n − i)! =
T :T ⊆{A1 ,...,An } i=0
i
n
X (−1)i
= (n!) . (5.19)
i=0 i!
m m
! !
X m X m
b(m) = a(m − k) = a(i). (5.20)
k=0
k i=0
i
P |S−T |
And f (T ) = S:S⊇T (−1) !
f (S) becomes
Pn j−i n−i
a(n − i) = j=i (−1) b(n − j), which we rewrite as
j−i
m m
! !
X
k m X
m−k m
a(m) = (−1) b(m − k) = (−1) b(k). (5.21)
k=0
k k=0
k
m
!
X
m−i m
iff a(m) = (−1) b(i), 0 ≤ m ≤ n. (5.22)
i=0
i
!
j
Exercise: 5.7.7 If A is the matrix whose (i, j) entry is , 0 ≤ i, j ≤ n,
i
then A−1 is the matrix whose (i, j) entry is (−1)j−i ji . (Hint: Try putting
b(m) = (x + 1)m and a(m) = xm for 0 ≤ m ≤ n.)
−1
1 1 1 1 1 1 −1 1 −1 1
0 1 2 3 4 0 1 −2 3 −4
0 0 1 3 6
=
0 0 1 −3 6
.
0 0 0 1 4
0 0 0 1 −4
0 0 0 0 1 0 0 0 0 1
Application 3 (Euler’s phi-function φ again). Let n ∈ P and suppose
p1 , . . . , pk are the distinct prime divisors of n. E = [n]. Ai = {x ∈ E : pi |x},
5.7. EVALUATING µ: THE PRODUCT THEOREM 223
n
i = 1, . . . , k. First note that |Ai1 ∩ Ai2 ∩ · · · ∩ Aik | = pi1 ···pik
. Then the
principle of inclusion-exclusion gives
k
!
X n n n 1
− · · · + (−1)K
X Y
φ(n) = n − + =n 1− .
1≤i≤k pi 1≤i<j≤k pi pj p1 · · · pk i=1 pi
It is easy to show that this agrees with our formula developed quite some
time ago.
Note:
X 1 1 1 X µ(d)
− · · · + (−1)k
X
1− + = ,
pi pi pj p1 · · · pk d|n
d
where µ is the classical Möbius function. So φ(n) = d|n µ(d) nd . Now using
P
P
classical Möbius inversion (in reverse), n = d|n φ(d), a familiar equality.
We now propose to illustrate the connection between classical Möbius
inversion and our new version over posets. Recall Example 5.7.2 from the
beginning of this section: P = (n1 + 1)×· · ·×(nk + 1), as well as the Möbius
function for chains as worked out in Section 5.6. By the product theorem, if
[(a1 , . . . , ak ), (b1 , . . . , bk )] is an interval in P ,
( P
(bi −ai )
(−1) , if each bi − ai = 0 or 1;
µ((a1 , . . . , ak ), (b1 , . . . , bk )) =
0, otherwise.
(5.23)
Now suppose n is a positive integer of the form n = pn1 1 · · · pnk k , where
p1 , . . . , pk are distinct primes. Let Dn be the poset of positive integral divisors
of n, ordered by division (i.e., i ≤ j in Dn iff i|j.) But we identify P above
with Dn according to the following scheme: (a1 , . . . , ak ) ∈ P corresponds
to pa11 · · · pakk in Dn . (Here it is convenient to let the elements of nk + 1
be {0, 1, . . . , nk }.) Then Eq. 5.23, when interpreted for Dn , becomes: For
r, s ∈ Dn ,
(
(−1)t , if s/r is a product of t distinct primes,
µ(r, s) = (5.24)
0, otherwise.
224 CHAPTER 5. MÖBIUS INVERSION ON POSETS
In other words, µ(r, s) is just the classical Möbius function µ(s/r). Then
our new Möbius inversion formula in Dn looks like:
X X
g(m) = f (d), ∀m|n, iff f (m) = g(d)µ(d, m), ∀m|n.
d|m d|m
parts of σ. Example: If σ = {{1, 3}, {5, 7, 8}, {2, 4, 6}}, then ν(σ) = 3. The
Q
goal of this exercise is to compute the Möbius function of n . The underlying
field of coefficients is denoted by K. The exercise is broken into ten small
steps.
Q
Step 1. n has a 0̂ and a 1̂, with ν(σ) = n iff σ = 0̂ and ν(σ) = 1 iff
σ = 1̂.
Step 2. Let σ ≤ π = {B1 , . . . , Bk } ∈ n . Suppose that Bi is partitioned
Q
Q
into λi blocks in σ. The interval [σ, π] in n is isomorphic to the direct prod-
uct λ1 × λ2 × · · · × λk . Illustrate this with π = {{1, 2, 3, 4, 5}, {6, 7, 8, 9}},
Q Q Q
σ = {{1, 2}, {3, 4, 5}, {6}, {7}, {8, 9}}. As a special case, for σ ∈ n , ,
Q
Step 3. For each positive integer n, let µn = µ(0̂, 1̂), where µ is the
Möbius function of n . Then using the notation of Step 2, i.e., [σ, π] '
Q
mn = mν(σ) .
X
Q
σ∈ n
g(σ) = mν(σ)
mν(π)
P
= π∈
Q
ν(σ)
mν(π)
P
= π∈
Q
:π≥σ
n
P
= π∈
Q
:π≥σ f (σ).
n
to obtain
X
f (σ) = µ(σ, π)g(π).
π≥σ
µ(σ, π)mν(σ) .
X
f (σ) =
Q
π∈ n
:π≥σ
n
µ(0̂, π) xk .
X X
(x)n =
Q
k=0 π∈ :ν(π)=k
n
Recall (from where?) that (x)n = nk=0 s(n, k)xk , where s(n, k) = (−1)n−k c(n, k)
P
and
" #
n
(−1)n−k , 1 ≤ k ≤ n.
X
µ(0̂, π) =
Q k
π∈ n
:ν(π)=k
Yn
µ(σ, π) = i=1
(−1)λi −1 [(λi − 1)!].
Proof: The first two identities need no further explanation, but the third
one probably does. A disposition may be visualized as a placing of n distin-
guishable flags on m distinguishable flagpoles. The poles are not ordred, but
the flags on each pole are ordered. For the first flag there are m choices of
flagpole. For the second flag there are m − 1 choices of pole other than the
one flag 1 is on. On that pole there are two choices, giving a total of m + 1
choices for flag 2. Similarly, it is easy to see that there is one more choice for
flag k + 1 than there was for flag k. Hence the number of ways to assign all
n flags is m(m + 1) · · · (m + n − 1) = (m)n .
(iii)0 xn = n−k
S(n, k)(x)k .
P
k (−1) This is Corollary 1.7.3.
228 CHAPTER 5. MÖBIUS INVERSION ON POSETS
Proof: The only two parts that need proving are (ii) and (ii)0 , and we
now establish (ii).
A linear partition λ is a partition of [n] together with a total order on
the numbers in each part of λ. The parts themselves are unordered. Let Ln
denote the collection of all linear partitions of [n] , and let ν(λ) denote the
number of blocks of λ. Each disposition from [n] to [m] may be thought of
as a pair consisting of a linear partition λ of [n] into k blocks and a one-to-
one function g mapping [k] to [m]. Since (m)n counts the total number of
dispositions of [n] and (m)k counts the one-to-one functions from [k] to [m],
we have
(m)n =
X
(m)ν(λ) . (5.25)
λ∈Ln
To obtain
the number of linear partitions of [n] into k blocks, note that
n−1
there are n! k−1 linear partitions of [n] with k ordered blocks. Visualize this
as a placing of k − 1 slashes into the n − 1 interior spaces of a permutation
(ordered array) of [n], at most one slash per space. Then divide by k! to get
unordered blocks:
!
n! n − 1
= #{ linear partitions of [n] with k blocks.} (5.26)
k! k − 1
n−1
These numbers n!
k! k−1
are called Lah numbers. Then (ii) follows im-
mediately from Eqs. 5.25 and 5.26. Now replacing x with −x interchanges
(ii) and (ii)0 .
We now do Möbius inversion on each of three carefully chosen posets to
explore the relationship between (a) and (a)0 , for a = i, ii, and iii.
Let n be the set of all partitions of [n] made into a poset by: for σ, π ∈
Q
mn =
X
(m)ν(σ) . (5.27)
Q
σ∈ n
g(π) = mν(π) . Since 0̂ in n is 0̂ = {{1}, {2}, {3}, . . . , {n}}, and ν(σ) = n iff
Q
σ = 0̂, we have
5.8. MORE APPLICATIONS OF MÖBIUS INVERSION 229
g(0̂) = mn =
X X
(m)ν(σ) = f (σ). (5.28)
Q
σ∈ n
σ≥0̂
ν(σ) . So σ ∈
Q Q
n is 0̂ in Pσ . And Eq. 5.28 applied to Pσ says:
X Y
g(σ) = f (π), for all σ ∈ n
. (5.29)
π≥σ
µ(σ, π)mν(π) .
X X
f (σ) = µ(σ, π)g(σ) = (5.30)
π≥σ π≥σ
Putting σ = 0̂ yields
n
µ(0̂, π)mν(π) µ(0̂, π) mk .
X X X
(m)n = f (0̂) = =
Q
k=1 and
Q
π∈ n π∈ ν(π)=k
n
(5.31)
As this holds for infinitely many m, we have a polynomial identity:
n
k
X X
(x)n =
µ(0̂, π)
x . (5.32)
k=1 and
Q
π∈ n
ν(π)=k
X
s(n, k) = µ(0̂, π) = wn−k , (5.33)
and
Q
π∈ n
ν(π)=k
the (n − k)th Whitney number of n of the first kind. This shows that (i)
Q
If π has type (a1 , . . . , ak ), i.e., π has ai parts of size i, then [0̂, π] ' ( 1 )a1 ×
Q
a
( 2 ) × · · · × ( k )ak . Hence µ(0̂, π) = ki=1 [(−1)i−1 (i − 1)!] i . Putting this
Q a2 Q Q
Define g : Ln → K by
Note that for λ ∈ Ln , λ = 0̂ iff ν(λ) = n. Then Eq. 5.25 implies that
(m)n = g(0̂) =
X X
(m)ν(λ) = f (λ). (5.37)
λ≥0̂ λ≥0̂
X
g(η) = f (λ) for all η ∈ Ln . (5.38)
λ≥η
X
f (η) = g(λ)µ(η, λ) for all η ∈ Ln . (5.39)
λ≥η
µ(0̂, λ)(m)ν(λ) .
X
(m)n = (5.40)
λ∈Ln
µ(0̂, λ)(m)ν(λ) =
X
(m)n =
λ∈Ln
n
!
n−ν(λ) ν(λ) n−k n!ν−1
(m)k .
X X
= (−1) (m) = (−1) (5.41)
λ∈Ln k=1 k! k − 1
n
!
n−k n! n−1
(x)k .
X
(x)n = (−1) (5.42)
k=1 k! k − 1
(m)c(τ ) = mc(σ) .
X
(5.44)
σ≥τ
Define f : Sn → K and g : Sn → K by
X
f (τ ) = g(σ), for all τ ∈ Sn . (5.46)
σ≥τ
X
g(τ ) = µ(τ, σ)f (σ), for all τ ∈ Sn . (5.47)
σ≥τ
mn = µ(0̂, σ)(m)c(σ) .
X
(5.48)
σ∈Sn
(
(−1)n−c(σ) , if σ is increasing;
For each σ ∈ Sn , µ(0̂, σ) = (5.49)
0, otherwise.
5.9. LATTICES AND GAUSSIAN COEFFICIENTS 233
Proof: Given σ ∈ Sn , consider the interval [0̂, σ]. The atoms of [0̂, σ]
correspond to transpositions (ir , ir+1 ) where (ir , ir+1 ) is a substring of a cycle
of σ and ir < ir+1 . Thus if σ is increasing, the atoms of Iσ = [0̂, σ] correspond
to all of the possible n − c(σ) transpositions. In that case Iσ is Boolean, and
µ(0̂, σ) = (−1)n−c(s) . So suppose σ is not increasing. Then some cycle
of σ has a consecutive pair (· · · , ir , ir+1 · · ·) with ir > ir+1 . Form a new
permutation σ ∗ from σ by inserting a pair )( of parentheses between ir , ir+1
for every consecutive pair (· · · ir , ir+1 · · ·) of all cycles where ir > ir+1 . Then
σ ∗ ≥ τ for every atom τ of [0̂, σ], and σ ∗ < σ. It follows that in the upper
Möbius algebra AV ([0̂, σ], K), if X is the set of atoms of [0̂, σ],
Y X
x= σt has σσ∗ as a summand. (5.50)
x∈X t≥x∀x∈X
mn = (−1)n−c(σ) (m)c(σ) =
X
σ∈Sn ;σ increasing
(−1)n−k (m)k
X X
1 . (5.51)
k σ increasing and c(σ)=k
and (ii)0 are related by Möbius inversion on Ln ; and (iii) and (iii)0 are
related by Möbius inversion on Sn .
Theorem 5.9.1 (L. Weisner 1935) Let µ be the Möbius function of a finite
lattice L. And let a ∈ L with a > 0̂. Then
X
µ(0̂, x) = 0.
x:x∨a=1̂
P
Proof: Fix a. Put S := x,y∈L µ(0̂, x)ζ(x, y)ζ(a, y)µ(y, 1̂). Now compute
S in two different ways.
X X X X
S= µ(0̂, x)µ(y, 1̂) = µ(0̂, x) · µ(y, 1)
x∈L x
y≥x y≥x
y≥a y≥a
X X
= µ(0̂, x) µ(y, 1̂)
x y≥x∨a
(
X X X 1 x ∨ a = 1;
= µ(0̂, x) · µ(y, 1̂) = µ(0̂, x) ·
x x 0, otherwise.
x∨a≤y≤1̂
5.9. LATTICES AND GAUSSIAN COEFFICIENTS 235
X
= µ(0̂, x), which is the sum in the theorem.
x:x∨a=1
Also,
X X X
S= µ(y, 1̂) µ(0̂, x) = µ(y, 1̂) · 0,
y≥a 0≤x≤y y≥a
since y > 0.
Let Vn (q) denote an n-dimensional vector space over Fq = GF (q). The
term k-subspace will denote a k-dimensional subspace. It is fairly easy to
see that the poset Ln (q) of all subspaces of Vn (q) is a lattice with 0̂ = {0}
and 1̂ = Vn (q). We begin with some counting.
Exercise: 5.9.2 The number of ordered bases for a k-subspace of Vn (q) is:
(q k −1)(q k −q)(q k −q 2 ) · · · (q k −q k−1 ). How many ordered, linearly independent
subsets of size k are there in Vn (q)?
(q n − q 0 ) (q n − q 1 ) (q n − q n−1 )
M (n, q) = · · · · =
(q 1 − q 0 ) (q 2 − q 1 ) (q n − q n−1 )
(q n − 1)q(q n−1 − 1)q 2 (q n−2 − 1) · · · q n−1 (q − 1)
= =
(q − 1)q(q − 1)q 2 (q − 1) · · · q n−1 (q − 1)
(q n − 1)(q n−1 − 1) · · · (q − 1)
= .
(q − 1)n
This implies that
" #
n
M (n, q) = · M (k, q) · M (n − k, q),
k q
" # " #
n M (n, q) n
= =
k q
M (k, q)M (n − k, q) n−k q
In fact there is a satisfactory way to generalize the notion and the notation
of Gaussian coefficient to the multinomial case. (See the book by R. Stanley
for this.) However, for our present purposes it suffices to consider just the
binomial case. Define (0)q = 1, and for a positive integer j put (j)q =
1 + q + q 2 + · · · + q j−1 . Then put (0)!q = 1 and for a positive integer k, put
(k)!q = (1)q (2)q · · · (k)q . So (n)!q = M (n, q). With this notation we have
" #
n (n)!q
= . (5.52)
k q
(k)!q (n − k)!q
" #
n
For some purposes it is better to think of as a polynomial in an
k q " #
n
indeterminate q rather than as a of function of a prime power q. That
k q
is a polynomial in q is an easy corollary of the following exercise.
(Hint: There is a completely elementary proof just using the formulas for
the symbols.)
Note that the relation of the previous exercise reduces to the binomial
recurrence when q = 1. However, unlike the binomial recurrence, it is not
‘symmetric’.
" #
n
= l≥0 αl q l , where αl is the number of
P
Exercise: 5.9.5 Show that
k q
partitions of l into at most k parts, each of which is at most n − k.
238 CHAPTER 5. MÖBIUS INVERSION ON POSETS
" #
n
If we regard a Gaussian coefficient as a function of the real variable
k q
q (where n and k are fixed integers), then we find that the limit as q goes to
1 of a Gaussian coefficient is a binomial coefficient.
n
" #
n−1
i(i−1) n
xi , for n ≥ 1.
X
(1 + x)(1 + qx) · · · (1 + q x) = q 2
i=0
i q
k
" # " # " #
n+m n m
q (n−i)(k−i) .
X
=
k q i=0
i q
k−i q
Proof: (i) is a special case of the binomial theorem. And (ii) becomes (i)
if q = 1. To prove (ii), suppose V, W are vector spaces over F = GF (q) with
dim(V ) = n and |W | = r. Here r = q t is any power of q with t ≥ n. Then
|HomF (V, W )| = rn .
Now classify f ∈ HomF (V, W ) according to the kernel subspace f −1 (0) ⊆
V . Given some subspace U ⊆ V , let {u1 , . . . , uk } be an ordered basis of
U and extend it to an ordered basis {u1 , . . . , uk , uk+1 , . . . , un } of V . Then
f −1 (0) = U iff f (ui ) = 0, 1 ≤ i ≤ k, and f (uk+1 ), . . . , f (un ) are linearly
independent vectors in W . Now
U ⊆V
n
" #
n
(r − 1)(r − q) · · · (r − q n−k−1 )
X
=
k=0
k q
n
" #
n
(r − 1)(r − q) · · · (r − q k−1 )
X
=
k=0
k q
" # " #
n n
(Use the fact that = )
k q
n−k q
n
" #
X n
= gk (r).
k=0
k q
i=0
i q
240 CHAPTER 5. MÖBIUS INVERSION ON POSETS
(Hint: In Ex. 5.9.7 first replace x with −x and then replace q with q −1
and simplify.)
If {an }∞n=0 is a given Psequence of numbers we have considered its ordi-
nary generating function n≥0 an xn and its exponential generating function
P xn
n≥0 an n! . (Also considered in Chapter 4 was the Dirichlet generating series
function.) There is a vast theory of Eulerian generating series functions de-
P xn
fined by n≥0 an n! q
. (See the book by R. Stanley for an introduction to this
subject with several references.) The next exercise shows that two specific
Eulerian generating functions are inverses of each other.
k
!
(−t)k q (2)
P
P tk
Exercise: 5.9.11 k≥0 k!q k≥0 k!q = 1.
n n
" # " #
n n−i n
n−i
q( 2 )
X X
vn = ui (n ≥ 0) ⇔ un = (−1) vi (n ≥ 0).
i=0
i q i=0
i q
k
(−1)k q (2) ,
(
µ(U, W ) = if U ⊆ W and k = dim(W ) − dim(U );
0, if U ⊆
6 W.
5.9. LATTICES AND GAUSSIAN COEFFICIENTS 241
Proof: The idea is to use Weisner’s theorem on the interval [U, W ] viewed
as isomorphic to the lattice of subspaces of the quotient space W/U . This
means that we need only compute µ(0̂, 1̂), where V = 1̂ is a space of dimen-
sion n and 0̂ = {0}. If V has dimension 1, then L1 (q) is a chain with two
1
elements and µ(0̂, 1̂) = −1 = (−1)1 · q (2) . Now suppose n = 2. Let a be a
point. By Weisner’s theorem
X
µ(0̂, 1̂) = − µ(0̂, p) = |{p : p ∨ a = 1̂ and p 6= 1̂}|
p : p ∨ a = 1̂
p 6= 1̂
2
= q = (−1)2 q (2) .
k=0
k q
Corollary 5.9.15 The number of n×m matrices over F = GF (q) with rank
r is
r
" # " #
m r r−k
(−1)r−k q nk+( ).
X
2
r q k=0
k q
n
" #
n n−k
n n n n n−1 n−k
q nk+( ).
X
gn (q ) = (q − 1)(q − q) · · · (q − q )= (−1) 2
k=0
k q
X X X
= µ(s, t)s = µ(s, t)s . (5.53)
(s,t):s≤t≤x and s≤t≤y t∈Λx ∩Λy s:s≤t
P
So if we put δt = s:s≤t µ(s, t)s, we have
X
x·y = δt (5.54)
t∈Λx ∩Λy
fix z ∈ P , g:g≤z
P P
If we
P
δt = (s,t):s≤t≤z µ(s, t)s =
s:s≤z δs,z · s = z. Hence:
P P
s:s≤z t:s≤t≤z µ(s, t) s =
X X
z= δt , and x · x = δt = x. (5.55)
t∈Λz t∈Λx
Moreover,
x · y = y iff y ≤ x. (5.56)
Note: In the above we are thinking of f ∈ K P as a formal (possibly
P
infinite) linear combination of the elements of P : f = x∈P f (x)x. So the
P
element x of P is identified with the element 1x = y∈P (δx,y )y = x. Then
the above discussion shows that {δt : t ∈ P } (as well as {x : x ∈ P }) is a
basis for AΛ (P, K).
Let A0Λ (P, K) be the abstract algebra x∈P Kx with Kx isomorphic to
Q
X X X X
a · δx = a · µ(t, x)t = µ(t, x)(a · t) = µ(t, x) d.
t≤x t≤x d∈P t:t≤x and a·t=d
(
X X δx , if x ≤ a;
a · δx = δt · δx = δt,x δx =
t:t≤a t:t≤a
0, if x ≤
6 a.
(a) If d ≤ x ≤ a, then
P
t:t≤x and a·t=d
µ(t, x) = µ(d, x),
and
(b) If d 6=≤ x ≤ a, then
P
t:t≤x and a·t=d
µ(t, x) = 0.
Theorem 5.10.4 Let P be a finite poset with 0̂ and 1̂, 0̂ 6= 1̂. And let
X be the set of coatoms of P (i.e., elements covered by 1̂). Then µ(0̂, 1̂) =
P|X| k
k=1 (−1) Nk , where Nk is the number of subsets of X of size k whose product
is 0̂.
exactly k (−1)k Nk .
P
X X X
σy = φ(p) = φ σx φ(σx ). (5.57)
y≥φ(p) x≥p x≥p
X0
φ∗ (σp ) =
X
σq = σq
q∈Q:φ(ψ(q))=φ(p)
where the sum 0 is over all σq for which q satisfies the following: q ≥ φ(p)and
P
the largest x with φ(x) ≤ q has φ(x) = φ(p). In this set of q’s, φ(p) is the
only one in the image of φ. The other q’s are only “slightly” larger than φ(p).
(This set of q’s is the set {q : [φ(p), q] ∩ φ(P ) = {φ(p)}}.)
5.10. POSETS WITH FINITE ORDER IDEALS 247
φ∗ (p) = φ∗ φ∗ (σx ) =
X X
σx =
x:x≥p x:x≥p
φ∗ (σx )
X
=
x:x≥p and x=max{y:φ(x)=φ(y)}
X X
= σq =
x:x≥p and x=max{y:φ(x)=φ(y)} q:q≥φ(x) and x=max{t∈P :φ(t)≤q}
X
= σq = φ(p).
q:q≥φ(p)
Corollary 5.10.6 Let (P, ≤) be a finite poset, and let P0 ⊆ P . Then the
injection φ : P0 → P : p 7→ p “extends to ” an algebra homomorphism of
AV (P, K) into AV (P, K) iff the restriction of each principal order ideal of P
to P0 is either empty or principal.