Академический Документы
Профессиональный Документы
Культура Документы
An introduction to the
theory of determinants
7.1
7.1.1
170
toward studying the eigentheory of square matrices over F. We should point
out that there are now many powerful readily available tools such as Maple
for computing determinants and for approximating eigenvalues.
7.1.2
The 2 2 case
Lets begin with the 2 2 case. Suppose A = ac db . Define the determinant
of A to be
|A| = det(A) = ad bc.
Clearly, ad bc = 0 if and only if the rows of A are proportional. Thus,
A has rank 2 if and only if ad bc 6= 0. Since A has maximal rank, it is
invertible, and (as we have shown elsewhere),
1
d b
1
A =
.
(ad bc) c a
Proposition 7.1. The determinant function det : M2 (F) F has the
following properties:
(1) |A| 6= 0 if and only if A is invertible;
(2) |I2 | = 1; and
(3) |AB| = |A||B| for any A, B M2 (F).
The proof is an easy calculation, so we will omit it. The remarkable
property is (3). We will see that it also holds for n n matrices.
7.1.3
Unlike many situations we have seen, the definition of the determinant for
2 2 matrices gives only the slightest hint of how to extend the definition
to the n n matrices so that the two properties of Proposition 7.1 continue
to hold. One hint is that the determinant of a diagonal matrix should be
the product of the diagonals. The general definition involves permuting the
rows of A and then taking the product of the new diagonals. Thus we will
need permutation matrices again.
Recall that a permutation matrix is a matrix P of zeros and ones that can
be expressed as a product of row swap matrices, that is, elementary matrices
of type I. P A is then A after a rearangement of its rows. If S1 , . . . , Sk are
row swap matrices such that P = S1 Sk , then we define |P | = (1)k .
The only thing wrong with this definition is that since we can express P as
171
a product of row swaps in many ways, we need to know that this definition
of |P | is independent of the representation P = S1 Sk . Thus we need to
0 , then m k is even.
know that if we also write P = S10 Sm
Proposition 7.2. For any two representations of a permutation matrix P ,
the number of row swap matrices involved is either always even or always
odd. Hence, det(P ) is well defined.
We willpostpone the proof of Proposition 7.2 until Chapter ?? as it
requires some familiarity with group theory, especially that of the symmetric
group.
Example 7.1. The 3 3 permutation matrices can be listed as follows.
I3 =
1 0 0
010 ,
001
0 1 0
100 ,
001
0 1 0
001 ,
100
0 0 1
010 ,
100
0 0 1
100 ,
010
1 0 0
001 .
010
(7.1)
c d
a b
172
Example 7.2. Lets work out the 3 3 case. We listed the 6 permutation
matrices in Example 7.1. Let Pijk be the 3 3 permutation matrix in which
the first row of I3 is in the ith row, the second in the jth row and the third
in the kth row. Then Pijk A is the same rearrangent of the rows of A. For
example,
a3
P312 A = a1 .
a2
Thus (P312 A) = a31 a12 a23 , and more generally, (Pijk A) = ai1 aj2 ak3 .
Working from left to right in (7.1), we have
|P123 | = 1,
|P213 | = 1,
|P231 | = 1,
|P312 | = 1,
|P321 | = 1,
and |P132 | = 1.
173
7.2
Calculating determinants
So far, we havent given any insight as to how determinants might be computed, but its a good bet that we arent going to compute them from the
definition.
7.2.1
Its obvious that if P is a permutation matrix and S is any row swap matrix,
then Q = P S is also a permutation matrix, and, moreover, Q = P S varies
over the set of all permutation matrices as P does. Therefore, we can rewrite
the sum as
X
X
|SA| =
|P S|(P SSA) =
(|P |)(P A).
P
The first equality is due to the fact that every permutation can be written
P S for exactly one P , the second is from the fact that |P S| = |P |, there
174
being one more row swap in P S than in P , and the fact that S 2 = In .
Therefore |SA| = |A|.
(Det II) If B is obtained from A by multiplying a row of A by a (possibly zero) scalar r, then |B| = r|A|.
1 if E is of type I,
|E| = r
if E is of type II,
1
if E is of type III
Therefore we can summarize Det I-Det III in the following way.
Proposition 7.4. If A is n n and E is an elementary n n matrix, then
|EA| = |E||A|
(7.4)
175
7.2.2
The method for evaluating |A|, based on Proposition 7.4, can be described
as follows. First find elementary matrices E1 , . . . , Ek such that U =
Ek E1 A is upper triangular. By Proposition 7.3 and repeated application
of Proposition 7.4,
|U | = |E1 | |Ek ||A| = u11 u22 unn ,
where the uii are the diagonal entries of U . Since no |Ei | = 0,
u11 u22 unn
|A| =
.
|E1 | |Ek |
(7.5)
1
0
A=
1
1
where
0
1
1
1
1
0
1
0
1
1
,
1
1
1 0 1 1
1 0 1 1
1 0 1
1
0 1 0 1
0 1 0 1
0 1 0
1
A
0 1 0 0 0 1 0 0 0 0 0 1
1 1 0 1
0 1 1 0
0 1 1 0
1 0 1
1
1 0 1
1
0 1 0
1
1
0 1 0
0 0 0 1
0 0 1 0
0 0 1 0
0 0 0 1
Thus |A| = 1.
176
Example 7.4. Let us next compute |A|, where A is the matrix of the
previous example, this time taking the field of coefficients to be Z2 . First
add the first row to the third and fourth rows successively. Then we get
1 0 1 1
0 1 0 1
|A| = det
0 1 0 0 .
0 1 1 0
Since the field is Z2 , row swaps also leave
1 0
0 1
|A| = det
0 1
0 1
1 1
0 0
.
0 1
1 0
Adding the second row to the third row and the fourth row successively, we
get
1 0 1 1
0 1 0 0
|A| = det
0 0 0 1 .
0 0 1 0
Finally, switching the last two rows,
1
0
|A| = det
0
0
we get
0
1
0
0
1
0
1
0
1
0
= 1.
0
1
One can simplify evaluating |A| even more in some special cases. For
example,
B C
det
= det(B) det(D),
(7.6)
O D
provided the submatrices B and C are square. The proof is similar to the
proof of (??).
7.2.3
We will now extend Proposition 7.1 to the n n case by showing the fundamental fact that |AB| = |A||B|. In fact, from what we have now established,
the product theorem is now transparent.
177
Theorem 7.6. If A and B are any nn matrices over F, then |AB| = |A||B|.
If both A and B both have rank n, then we can write each one of them
as a product of elementay matrices, say A = E1 Ek and B = Ek+1 Em .
Then AB = (E1 Ek )(Ek+1 Em ), so
|AB| = |E1 | |Ek ||Ek+1 | |Em | = |A||B|.
by Proposition 7.5. To finish the proof, we have to show that if either |A|
or |B| is zero, then |AB| = 0. However, if |AB| 6= 0, then AB is invertible,
and we know this implies both A and B are invertible. Hence if either |A|
and |B| is zero, then |AB| = 0, and the proof is complete.
Proposition 7.7. If A is invertible, then |A1 | = |A|1 .
Proof. This follows from the previous result since AA1 = In implies
|AA1 | = |A||A1 | = 1.
by the rule for transposing a product. Thus it suffices to show that for any
elementary matrix E, we have
|E T | = |E|.
This is clear if E is of type II or III, since elementary matrices of type II are
symmetric and the transpose of a type III elementary matrix is also of type
III, so in both cases the determinant is 1. If E is of type I, then E 1 = E T ,
so |E| = |E 1 | = |E T |, the common value being 1.
178
7.2.4
7.2.5
Laplaces Expansion
n
X
(1)i+j aij |Aij |.
(7.7)
i=1
Of course, one has to show that all the row and column expansions all
give the same result, namely |A|. The proof isnt hard, but its messy, so we
179
wont attempt it. For example, if A is 3 3, expanding |A| along the first
column gives
|A| = a11 (a22 a33 a32 a23 ) a21 (a12 a23 a13 a32 ) + a31 (a12 a23 a13 a22 ).
This is simply the formula |A| = a1 (a2 a3 ) mentioned above.
Example 7.5. Here is an example of why the Laplace expansion can be
useful. Suppose we want to find all values of x for which the matrix
1x
2
0
Cx = 2
1 x 1
0
1 2 x
has rank less that 3, i.e. is singular. Hence we will set |Cx | = 0 and solve
for x. Clearly, row operations arent going to be of much help here, so we
will compute |Cx | via Laplace. If we expand along the first column, we get
|Cx | = (1 x) (1 x)(2 x) (1)(1) 2 2(2 x) 0(1)
= x3 + 4x 7
Hence Cx is singular at the three roots of x3 4x + 7 = 0.
A moments pause tells us that the Laplace expansion isnt even in the
same ballpark row ops when it comes to finding an efficient a procedure for
computing |A|. Indeed, after reading the definition of the Laplace expansion,
try to compare the number of operations needed to compute a 10 10 determinant using the Laplace expansion as compared to using row ops. In fact,
it isnt completely unreasonable that one might need to evaluate something
like a 2000 2000 determinant, as calculations involving the human genome
routinely require evaluating much larger determinants. Using Laplace, even
a 20 20 determinant is impractical, except possibly for a super computer
(note 20! = 2432902008176640000). To get an appreciation for what Im
claiming, try computing a Laplace expansion on a 5 5 determinant with
your calculator.
To sum up, the Laplace expansion is frequently useful for computing
examples, since one can combine it with row and column operations, but
I really would like to warn you against solely relying on it for exercises or
tests, as row operations involve much less calculation.
180
7.2.6
Cramers rule
Another reason the Laplace expansion is important, at least from a theoretical point of view, is that it gives a closed formula for the inverse of a
matrix. Recall that the inverse of inverse of a 2 2 matrix A = (aij ) is given
as follows:
1
a22 a12
1
A =
.
|A| a21 a11
Inspecting this formula may suggest the correct formula for A1 in the
general case.
Proposition 7.9. Suppose |A| 6= 0 and let bij = (1)i+j |Aji |, where, as
above, Aij denotes the (n 1) (n 1) submatrix of A obtained by deleting
As ith row and jth column. Then
A1 =
1
(bij ).
|A|
1 x 2
A(x) = x 1 x
2 3 1
is singular, that is, not invertible.
181
Exercise 7.8. Do the same as Problem
1 x
1 0
B(x) =
0 x
1 0
1 x
x 1
.
1 1
1 0
1 2 1
2 1
1
0 1 2
1 1 1
0
1
0
1
in two cases: first when the field is Q and secondly when the field is Z5 .
7.3
182
7.3.1
e1 e2 e3
x y = det x1 x2 x3 .
y1 y2 y3
A frequently asked question is why the cross product exists only in three
dimensions. In fact there is an n dimensional generalization of the cross
product called the vector product. The vector product assigns to any (n 1)
vectors x1 , x2 , . . . , xn1 Rn a vector
[x1 , x2 , . . . , xn1 ] Rn
orthogonal to each of the xi . It is defined by the follows determinantal
expression:
e2
...
en
e1
x11
x12
...
x1n
.
[x1 , x2 , . . . , xn1 ] = det
xn1,1 xn1,2 . . . xn1,n
The fact that
xi [x1 , x2 , . . . , xn1 ] = 0
for each i, 1 i n 1, is simply due to the fact that a determinant with
two equal rows is 0.
7.3.2
183
Note that here the vertical bars denote the absolute value. To connect this
with matrices, consider the linear transformation of A : Rn Rn such that
A(ei ) = wi . In other words, the ith column of the matrix of A is wi . Thus
|det(A)| is the volume of the image under A of the unit cube spanned by
the standard basis vectors e, . . . , en , i.e.
|det(A)| = Vol(< w1 , . . . , wn >).
Let us make a couple of comments about determinants and the geometry
of linear transformations. The upshot of the above remarks is that the linear
transformation associated to a real matrix having determinant of absolute
value 1 preserves the volume of a cube. A linear transformation T : Rn Rn
whose determinant is 1 is called unimodular. We say that the matrix of T
is unimodular. The set of all unimodular real n n matrices is denoted by
SL(n, R). It is called the special linear group.
Proposition 7.10. Products and inverses of unimodular real matrices are
also unimodular.
Unimodular matrices have another property that is a little more subtle.
For concreteness, lets assume for the moment that n = 3. If the matrix A
of a linear transformation T : R3 R3 has positive determinant, then the
linear transformation preserves right handed systems. Thus if det(T ) > 0,
then T is orientation preserving.
If A has determinant 1, then A preserves volumes but reverses orientation. The possibility of having volume preserving orientation reversing
transformations is what makes it necessary to put the absolute value in the
change of variables formula below.
7.3.3
f
f1
1
x1 . . .
xn
(f1 , . . . , fn )
JF =
.
= det
(x1 , . . . , xn )
fn
fn
x1 . . .
xn
184
It is an easy exercise in partial differentiation that if F is linear and A is its
associated matrix, then the Jacobian of F satisfies JF = |A|.
Now the change of variables formula says that if U is a closed n dimensional solid, then
Z
Z
g(y)dV =
g(F (x))|JF (x)|dV
F (U )
for all functions g that are continuous on F (U ). The intuition for this is
based on the observation that |JF (x)|dV is a good approximation of the
volume of the (curvilinear) solid parallelogram F (U ), where U denotes a
small cube in Rn and x U is arbitrary.
A specific application of this formula is that the volume of F (U ) is the
integral of the absolute value of the Jacobian of F over U . If F is a linear
transformation, then
Vol(F (U )) = |A|Vol(U ).
Using this identity, it is not hard to give a geometric proof that if A, B are
two n n matrices over R, then
|det(AB)| = |det(A)||det(B)|,
where we have used the det notation to avoid confusion with absolute values.
7.3.4
The notion of the determinant of a matrix was introduced in the 19th century. At that time, many if not most mathematicians were amateurs. In
other words, mathematics was a hobby, not a vocation. This may explain
why the determinant held so much fascination. It is complicated enough
to be a non trivial concept, yet it is also a concrete concept that one can
calculate. One 19th century mathematician who studied determinants was
Charles Dodgson, a professor of mathematics at Oxford who is better known
for being the author of Alice in Wonderland. Dodgson is one of the discoverers of an amusing identity, known as Lewis Carrolls Identity. Suppose
A is an n n matrix. Let AC be the (n 2) (n 2) submatrix in the
middle of A obtained by deleting the first and last rows and columns. Also,
let AN W the (n 1) (n 1) submatrix in the upper left hand corner of A,
and define AN E , ASW and ASE to be the (n 1) (n 1) submatrices in
the other three corners of A. If n = 2, put |AC | = 1. Then Lewis Carrolls
Identity says:
|A||AC | = |AN W ||ASE | |AN E ||ASW |
(7.8)
185
(see C.L. Dodgson, Proc. Royal Soc. London 17, 555-560(1860)). Interestingly, Lewis Carrolls Identity has recently been extended into the modern
setting of semi-simple Lie algebras.
Exercises
Exercise 7.14. Verify Lewis Carrolls Identity for the matrix
1 2 1 0
2 1
1 1
0 1 2 0 .
1 1 1 1
Exercise 7.15. Under what condition does Lewis Carrolls Identity make
it possible to evaluate |A|?
Exercise 7.16. Suppose A is an n n matrix over R.
(a) First show that the Jacobian of the linear transformation defined by A
is |A|.
(b) Use this to give a verification of the identity that says
|det(AB)| = |det(A)||det(B)|.
Exercise 7.17. List all the 3 3 permutation matrices that lie in SL(3, R).
Exercise 7.18. Prove Proposition 7.10.
Exercise 7.19. Prove that SL(3, R) is a subgroup of GL(3, R), where
GL(n, R) = {A Mn (R) | det(A) 6= 0}.
Exercise 7.20. *. If G and H are groups, then a mapping : G H is
called a group homomorphism if for any a, b G, (ab) = (a)(b). Explain
how the determinant can be viewed as a group homomorphism if we choose
the group G to be GL(n, F), where F is any field.