Академический Документы
Профессиональный Документы
Культура Документы
M. Anthony, M. Harvey
MT2175, 2790175
2012
Undergraduate study in
Economics, Management,
Finance and the Social Sciences
This is an extract from a subject guide for an undergraduate course offered as part of the
University of London International Programmes in Economics, Management, Finance and
the Social Sciences. Materials for these programmes are developed by academics at the
London School of Economics and Political Science (LSE).
For more information, see: www.londoninternational.ac.uk
This guide was prepared for the University of London International Programmes by:
Martin Anthony, Professor of Mathematics, and Michele Harvey, Course Leader, Department of
Mathematics, London School of Economics and Political Science.
This is one of a series of subject guides published by the University. We regret that due to
pressure of work the authors are unable to enter into any correspondence relating to, or
arising from, the guide. If you have any comments on this subject guide, favourable or
unfavourable, please use the form at the back of this guide.
Contents
Preface 1
1 Introduction 3
1.1 This subject . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.1.1 Aims of the course . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.1.2 Learning outcomes . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.1.3 Topics covered . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.2 Reading . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.2.1 Essential reading . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.2.2 Further reading . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.3 Online study resources . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
1.3.1 The VLE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
1.3.2 Making use of the Online Library . . . . . . . . . . . . . . . . . . 6
1.4 Using the subject guide . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
1.5 Examination . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
1.6 The use of calculators . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
i
Contents
ii
Contents
Reading . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59
Essential reading . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59
Further reading . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59
Aims of the chapter . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59
5.1 The direct sum of two subspaces . . . . . . . . . . . . . . . . . . . . . . . 59
5.1.1 The sum of two subspaces . . . . . . . . . . . . . . . . . . . . . . 59
5.1.2 Direct sums . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 60
5.2 Orthogonal complements . . . . . . . . . . . . . . . . . . . . . . . . . . . 61
5.2.1 The orthogonal complement of a subspace . . . . . . . . . . . . . 61
5.2.2 Orthogonal complements of null spaces and ranges . . . . . . . . . 62
5.3 Projections . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 63
5.3.1 The definition of a projection . . . . . . . . . . . . . . . . . . . . 63
5.3.2 An example . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 64
5.3.3 Orthogonal projections . . . . . . . . . . . . . . . . . . . . . . . . 65
5.4 Characterising projections and orthogonal projections . . . . . . . . . . . 65
5.4.1 Projections are idempotents . . . . . . . . . . . . . . . . . . . . . 65
5.5 Orthogonal projection onto the range of a matrix . . . . . . . . . . . . . 66
5.6 Minimising the distance to a subspace . . . . . . . . . . . . . . . . . . . 67
5.7 Fitting functions to data: least squares approximation . . . . . . . . . . . 68
5.7.1 A linear algebra view . . . . . . . . . . . . . . . . . . . . . . . . . 69
5.7.2 An example . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69
Learning outcomes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71
Test your knowledge and understanding . . . . . . . . . . . . . . . . . . . . . 72
Feedback on selected activities . . . . . . . . . . . . . . . . . . . . . . . . . . . 72
6 Generalised inverses 73
Reading . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73
Aims of the chapter . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73
6.1 Left and right inverses . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73
6.2 Weak generalised inverses . . . . . . . . . . . . . . . . . . . . . . . . . . 77
6.3 Strong generalised inverses . . . . . . . . . . . . . . . . . . . . . . . . . . 79
6.4 A method for calculating SGIs . . . . . . . . . . . . . . . . . . . . . . . . 81
6.5 Why are SGIs useful? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85
Learning outcomes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89
Test your knowledge and understanding . . . . . . . . . . . . . . . . . . . . . 90
Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 90
iii
Contents
Comments on exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92
Feedback on selected activities . . . . . . . . . . . . . . . . . . . . . . . . . . . 99
iv
Preface
This subject guide is not a course text. It sets out a logical sequence in which to study
the topics in this subject. Where coverage in the main texts is weak, it provides some
additional background material. Further reading is essential. We are grateful to
Dr James Ward and Dr Bob Simon for providing us with their materials on generalised
inverses and Jordan normal form.
1
Preface
2
1
Chapter 1
Introduction
In this very brief introduction, we aim to give you an idea of the nature of this subject
and to advise on how best to approach it. We give general information about the
contents and use of this subject guide, and on recommended reading and how to use the
textbooks.
enable students to acquire further skills in the techniques of linear algebra, as well
as understanding the principles underlying the subject
prepare students for further courses in mathematics and/or related disciplines (e.g.
economics, actuarial science).
As emphasised above, however, we do also want you to understand why certain
methods work: this is one of the ‘skills’ that you should aim to acquire. The examination
will test not simply your ability to perform routine calculations, but will probe your
knowledge and understanding of the fundamental principles underlying the area.
3
1. Introduction
1
expected to be able to use your knowledge and understanding of the material to solve
problems that are not completely standard. This is not something you should worry
unduly about: all mathematics topics expect this, and you will never be expected to do
anything that cannot be done using the material of the course. Second, we expect you
to be able to ‘demonstrate knowledge and understanding’ and you might well wonder
how you would demonstrate this in the examination. Well, it is precisely by being able
to grapple successfully with unseen, non-routine questions that you will indicate that
you have a proper understanding of the topic.
Generalised inverses
1.2 Reading
There are many books that would be useful for at least some parts of this course. We
recommend one in particular, and another for additional, further reading. Neither of
these books covers generalised inverses or Jordan normal form, and those topics have
therefore been discussed in this subject guide in such a way that they are self-contained.
R
You will need a copy of the following textbook.
Anthony, M. and M. Harvey. Linear Algebra: Concepts and Methods. (Cambridge
University Press, 2012) [ISBN 9780521279482].
4
1.3. Online study resources
1
1.2.2 Further reading
R
For additional reading, we suggest the following.
Anton, H. and C. Rorres. Elementary Linear Algebra with Supplemental
Applications (International Student Version). (John Wiley & Sons (Asia) Plc Ltd,
2010) tenth edition. [ISBN 9780470561577 ].1
Please note that as long as you read the Essential reading you are then free to read
around the subject area in any text, paper or online resource. You will need to support
your learning by reading as widely as possible. To help you read extensively, you have
free access to the virtual learning environment (VLE) and University of London Online
Library (see below).
Textbooks will provide more in-depth explanations than you will find in this subject
guide, and they will also provide many more examples to study and exercises to work
through. The books listed are the ones we have referred to in this subject guide.
Self-testing activities: Doing these allows you to test your own understanding of
subject material.
Electronic study materials: The printed materials that you receive from the
University of London are available to download, including updated reading lists
and references.
1
There are many editions and variants of this book. Any one is equally useful and you will not need
more than one of them. You can find the relevant sections cited in this subject guide in any edition by
using the index.
5
1. Introduction
1
Past examination papers and Examiners’ commentaries: These provide advice on
how each examination question might best be answered.
A student discussion forum: This is an open space for you to discuss interests and
experiences, seek support from your peers, work collaboratively to solve problems
and discuss subject material.
Videos: There are recorded academic introductions to the subject, interviews and
debates and, for some courses, audio-visual tutorials and conclusions.
Recorded lectures: For some courses, where appropriate, the sessions from previous
years’ Study Weekends have been recorded and made available.
Study skills: Expert advice on preparing for examinations and developing your
digital literacy skills.
Feedback forms.
Some of these resources are available for certain courses only, but we are expanding our
provision all the time and you should check the VLE regularly for updates.
6
1.5. Examination
1
provided on the VLE area for this course. Solutions to any additional exercises given in
this subject guide are provided in the subject guide.
The exercises are a very useful resource. You should try them once you think you have
mastered a particular chapter. Really try them: do not just simply read the solutions
provided. Make a serious attempt before consulting the solutions. It is vital that you
develop and enhance your problem-solving skills and the only way to do this is to try
lots of examples.
Near the end of the chapters, we also provide some feedback on the activities contained
in the chapter. Again, you should consult these only after you have attempted the
activities.
Unless it is explicitly stated otherwise, you are expected to work through and
understand the proofs in this subject guide and also those proofs in the Anthony and
Harvey book to which we refer. Understanding proofs is a good way to see how the
theory works to achieve results.
We often use the symbol to denote the end of a proof where we have finished
explaining why a particular result is true. This is just to make it clear where the proof
ends and the following text begins.
1.5 Examination
Important: the information and advice given here are based on the examination
structure used at the time this subject guide was written. Please note that subject
guides may be used for several years. Because of this we strongly advise you to always
check both the current Regulations for relevant information about the examination, and
the VLE where you should be advised of any forthcoming changes. You should also
carefully check the rubric/instructions on the paper you actually sit and follow those
instructions.
This course is assessed by a two-hour unseen written examination. A Sample
examination paper is given as the final chapter to this subject guide. In addition, a
commentary is provided on the sample paper.
Please do not think that the questions in a real examination will necessarily be very
similar to those in the Sample examination paper. An examination is designed (by
definition) to test you. You will get examination questions unlike questions in this
subject guide. The whole point of examining is to see whether you can apply knowledge
in familiar and unfamiliar settings. The Examiners (nice people though they are) have
an obligation to surprise you! For this reason, it is important that you try as many
examples as possible, from the subject guide and from the textbooks. This is not so
that you can cover any possible type of question the Examiners can think of! It is so
that you get used to confronting unfamiliar questions, grappling with them, and finally
coming up with the solution.
Do not panic if you cannot completely solve an examination question. There are many
marks to be awarded for using the correct approach or method.
Remember, it is important to check the VLE for:
7
1. Introduction
1
up-to-date information on examination and assessment arrangements for this course
where available, past examination papers and Examiners’ commentaries for the
course which give advice on how each question might best be answered.
8
Chapter 2 2
Diagonalisation, Jordan normal form
and differential equations
Reading
The Jordan normal form and its applications to differential equations are not discussed
in these textbooks, so we have tried to write this chapter of the subject guide in such a
way that the discussion of that topic is self-contained.
R
Essential reading
Anthony, M. and M. Harvey. Linear Algebra: Concepts and Methods. Chapter 9.
(For revision of diagonalisation, read Chapter 8.)
R
Further reading
Anton, H. and C. Rorres. Elementary Linear Algebra. Section 5.4.
9
2. Diagonalisation, Jordan normal form and differential equations
discussion of calculus.
For a function y = y(t), the derivative of y will be denoted by y 0 = y 0 (t) or dy/dt. We
2 will need the following result: if y(t) satisfies the differential equation
y 0 = ay,
If an initial condition, y(0), is given, then since y(0) = βe0 = β, we have a particular
(unique) solution y(t) = y(0)eat to the differential equation.
y0 = Ay,
where A = (aij ) is an n × n matrix whose entries are constants (that is, fixed numbers),
and y = (y1 , y2 , . . . , yn )T , y0 = (y10 , y20 , . . . , yn0 )T are vectors of functions.
If A is diagonal, the system y0 = Ay is easy to solve. Suppose
A = diag(λ1 , λ2 , . . . , λn ),
the diagonal matrix whose diagonal entries are (in order) λ1 , . . . , λn . Then the system is
precisely
y10 = λ1 y1 , y20 = λ2 y2 , . . . , yn0 = λn yn ,
and so
y1 = y1 (0)eλ1 t , y2 = y2 (0)eλ2 t , ..., yn = yn (0)eλn t .
Since a diagonal system is so easy to solve, it would be very helpful if we could reduce
our given system to a diagonal one, and this is precisely what the method will do.
10
2.3. Solving by diagonalisation
Example 2.1 Suppose the functions y1 (t) and y2 (t) are related as follows:
y10 = 7y1 − 15y2
y20 = 2y1 − 4y2 .
In matrix form this is y0 = Ay where A is the 2 × 2 matrix
7 −15
A= .
2 −4
This matrix is diagonalisable.
Activity 2.2 Show that the matrix is diagonalisable. Find an invertible matrix
P and a diagonal matrix D such that P −1 AP = D.
P z0 = y0 = Ay = A(P z) = AP z
and hence
z0 = P −1 AP z = Dz.
In other words,
z10
1 0 z1 z1
= = .
z20 0 2 z2 2z2
So the system for the functions z1 , z2 is diagonal and hence it is easily solved. Having
found z1 , z2 we can then find y1 and y2 through the explicit connection between the
two sets of functions: namely, y = P z.
11
2. Diagonalisation, Jordan normal form and differential equations
Let us now return to the general technique. Suppose we have the system y0 = Ay, and
that A can indeed be diagonalised. Then, there is an invertible matrix P and a diagonal
matrix D such that P −1 AP = D. Here,
2
P = (v1 . . . vn ), D = diag(λ1 , λ2 , . . . , λn ),
where λi are the eigenvalues and vi are the corresponding eigenvectors. Let z = P −1 y
(or, equivalently, let y = P z). Then
y0 = (P z)0 = P z0 ,
Therefore
P z0 = Ay = AP z,
and
z0 = P −1 AP z = Dz.
We may now easily solve for z, and hence y.
We illustrate with an example of a 3 × 3 system of differential equations, solved using
this method. Note carefully how we use the initial values y1 (0), y2 (0) and y3 (0).
Example 2.2 We find functions y1 (t), y2 (t), y3 (t) such that y1 (0) = 2, y2 (0) = 1
and y3 (0) = 1 and such that they are related by the linear system of differential
equations,
dy1
= 6y1 + 13y2 − 8y3
dt
dy2
= 2y1 + 5y2 − 2y3
dt
dy3
= 7y1 + 17y2 − 9y3 .
dt
We can express this system in matrix form as y0 = Ay where
6 13 −8
A = 2 5 −2 .
7 17 −9
We have P −1 AP = D where
1 −1 1 −2 0 0
P = 0 1 1, D = 0 1 0.
1 1 2 0 0 3
12
2.3. Solving by diagonalisation
We set y = P z, and substitute into the equation, y0 = Ay to obtain (P z)0 = A(P z).
That is, P z0 = AP z and so z0 = P −1 AP z = Dz. In other words, if
z1 2
z = z2 ,
z3
then
z10
−2 0 0 z1
z20 = 0 1 0 z2 .
z30 0 0 3 z3
So,
z10 = −2z1 , z20 = z2 , z30 = 3z3 .
Therefore,
z1 = z1 (0)e−2t , z2 = z2 (0)et , z3 = z3 (0)e3t .
Then, using y = P z, we have
z1 (0)e−2t
y1 1 −1 1
y2 = 0 1 1 z2 (0)et .
y3 1 1 2 z3 (0)e3t
It remains to find z1 (0), z2 (0), z3 (0). To do so, we use the given initial values
y1 (0) = 2, y2 (0) = 1, y3 (0) = 1. Since y = P z, we can see that y(0) = P z(0). We
could use row operations to solve this system to determine z(0). Alternatively, we
could use z(0) = P −1 y(0). Let us take the second approach. You should calculate
the inverse of P and find that
−1 −3 2
P −1 = −1 −1 1 .
1 2 −1
Therefore,
z1 (0) −1 −3 2 2 −3
z(0) = z2 (0) = P −1 y(0) = −1 −1 1 1 = −2 .
z3 (0) 1 2 −1 1 3
Therefore, finally,
z1 (0)e−2t
y1 1 −1 1
y2 = 0 1 1 z2 (0)et
y3 1 1 2 z3 (0)e3t
−3e−2t
1 −1 1
= 0 1 1 −2et
1 1 2 3e3t
−3e−2t + 2et + 3e3t
= −2et + 3e3t .
−2t t 3t
−3e − 2e + 6e
13
2. Diagonalisation, Jordan normal form and differential equations
How can we check our solution? First of all, it should satisfy the initial conditions. If we
substitute t = 0 into the equations we should obtain the given initial conditions.
The real check is to look at the derivatives at t = 0. We can take the original system,
y0 = Ay and use it to find y0 (0),
0
y1 (0) 6 13 −8 y1 (0) 6 13 −8 2 17
y20 (0) = 2 5 −2 y2 (0) = 2 5 −2 1 = 7 .
y30 (0) 7 17 −9 y3 (0) 7 17 −9 1 22
And we can differentiate our solution to find y0 , and then substitute t = 0.
0 −2t
y1 (t) 6e + 2et + 9e3t
y20 (t) = −2et + 9e3t .
0 −2t t 3t
y3 (t) 6e − 2e + 18e
Activity 2.7 Substitute t = 0 to obtain y0 (0) and check that it gives the same
answer.
Example 2.3 Let y1 (t), y2 (t), y3 (t) be functions related by the system of
differential equations
dy1
= 6y1 + 13y2 − 8y3
dt
dy2
= 2y1 + 5y2 − 2y3
dt
dy3
= 7y1 + 17y2 − 9y3 .
dt
Let the matrices A, P and D be exactly as before in Example 2.2, so that we still
have P −1 AP = D, and setting y = P z, to define new functions z1 (t), z2 (t), z3 (t), we
have
y0 = Ay ⇐⇒ P z0 = AP z ⇐⇒ z0 = P −1 AP z = Dz.
So we need to solve the equations
14
2.4. The Jordan normal form
that is,
y1 (t) = αe−2t − βet + γe3t
y2 (t) = βet + γe3t for α, β, γ ∈ R.
−2t t 3t
y3 (t) = αe + βe + 2γe
Using the general solution, you can find particular solutions for any given initial
conditions. For example, using the same initial conditions y1 (0) = 2, y2 (0) = 1 and
y3 (0) = 1 as in Example 2.2, we can substitute t = 0 into the general solution to
obtain,
y1 (0) = 2 = α − β + γ
y2 (0) = 1 = β + γ
y3 (0) = 1 = α + β + 2γ
and solve this linear system of equations for α, β, γ. Of course this is precisely the
same system y(0) = P z(0) as before, with solution P −1 y(0),
α −1 −3 2 2 −3
β = −1 −1 1 1 = −2 .
γ 1 2 −1 1 3
Activity 2.8 Find the particular solution of the system of differential equations in
Example 2.3 which satisfies the initial conditions y1 (0) = 1, y2 (0) = 1 and y3 (0) = 0.
15
2. Diagonalisation, Jordan normal form and differential equations
(x − λ1 )m1 . . . (x − λk )mk , then there will exist an invertible n × n matrix P such that
2 A1 0 ... 0
0 A2 ... 0
P −1 AP =
... .. . . . .. ,
. .
0 0 . . . Ak
λi ∗ 0 ... ... 0 0
0 λi ∗ 0 ... 0 0
0 0 λi ∗ ... 0 0
. .. .. ... . . . .. ..
Ai = .
.. ..
.
..
. . ,
.
.. . . ... λi ∗ 0
0 0 0 ... . . . λi ∗
0 0 0 ... ... 0 λi
where the ∗ are either 0 or 1 and all other entries are zeros.
Note that the special case of this in which all ∗ entries are 0 is nothing more than a
diagonal matrix. But the point of the theorem is that even if such a matrix A cannot be
diagonalised, it will be similar to a Jordan matrix, which is an ‘almost-diagonal’ matrix.
0 0 0 2
If
3 −4 0 8
2 −3 0 7
P =
1 −2 1 2 ,
0 0 0 1
then P is invertible, with
3 −4 0 4
2 −3 0 5
P −1 =
1 −2 1 4
0 0 0 1
and (check this!)
1 1 0 0
0 1 1 0
P −1 AP = J =
,
0 0 1 0
0 0 0 2
which is a Jordan matrix.
16
2.4. The Jordan normal form
There is another way to describe Jordan matrices, which is sometimes more useful. A
k × k matrix B is a Jordan block if, for some λ, if k = 1, then B = (λ) and, if k ≥ 2,
λ 1 0 0
... 0 0
2
0 λ 1 0 ... 0 0
0 0 λ 1 ... 0 0
. . . . .. . .
B= . . . .. . .. ..
.. .. .. . .. .. ..
,
.. .. .. .. . . .
0 0 0 ... ... λ 1
0 0 0 ... ... 0 λ
B1 0 ··· 0
0 B2 ··· 0
J =
... .. . . . .. ,
. .
0 0 · · · Br
Then,
B1 0
J= ,
0 B2
where B1 , B2 are the following Jordan blocks:
1 1 0
B1 = 0 1 1 , B2 = (2).
0 0 1
If J1 , J2 are two Jordan matrices similar to a given matrix A, then it turns out that the
only way in which J1 , J2 can differ is in the order of the Jordan blocks. (This is
analogous to the fact that, for a diagonalisable matrix, the only diagonal matrices
similar to it are those whose diagonal entries are the eigenvalues, in some order, and
this order can be changed.) For this reason, we can speak of the Jordan normal form of
A (or the Jordan canonical form of A), by which we mean a Jordan matrix similar to A.
(Although there are a number of such Jordan matrices, they differ only in the ordering
of the Jordan blocks, so this is why we use the word ‘the’ in ‘the Jordan normal form’:
we could instead say ‘a’ but the point is that there is essentially only one such matrix.)
17
2. Diagonalisation, Jordan normal form and differential equations
2
2 0 1 −3
0 2 10 4
A=
0
.
0 2 0
0 0 0 3
Then it turns out that A is not diagonalisable. The Jordan normal form of A is
2 0 0 0
0 2 1 0
J = 0 0 2
.
0
0 0 0 3
Note that, given what we said just a moment ago, we could equally well say that the
Jordan normal form is
3 0 0 0
0 2 1 0
J = 0 0 2 0.
0 0 0 2
It is the case that, if
0 1 0 −3
1 10 0 4
P =
0 0 1 0 ,
0 0 0 1
then P is invertible and P −1 AP = J. (Although P is a 4 × 4 matrix, it is easy to see
this because the last two rows of P are so simple.) Just to make sure you understand
the Jordan block description of a Jordan matrix, note that we can write
2 0 0 0
0 2 1 0 B 1 0 0
J =0 0 2 0 =
0 B2 0 ,
0 0 B3
0 0 0 3
We will be interested in using Jordan normal forms to solve systems of linear differential
equations, but we will not, in this course, be concerned with how to determine the
Jordan normal form. But let us make a few observations. Look at Example 2.6. What
this tells us is that if
0 1 0 −3
1 10 0 4
v1 = 0 , v 2 = 0 , v3 = 1 , v4 = 0 ,
0 0 0 1
18
2.5. Solving systems of differential equations using Jordan normal form
Example 2.7 Suppose we want to find the general solution of the following system
of differential equations:
dy1
= y1 + y3
dt
dy2
= y1 + y2 − 3y3
dt
dy3
= y2 + 4y3 .
dt
We can express this system in matrix form as y0 = Ay where
1 0 1
A = 1 1 −3 .
0 1 4
Our approach outlined earlier would attempt to diagonalise A. But A turns out not
to be diagonalisable.
19
2. Diagonalisation, Jordan normal form and differential equations
The next best thing is to work with the Jordan normal form. Suppose this is given:
explicitly, suppose we know that P −1 AP = J where
2
1 1 0
2 1 0
P = −2 −3 0 ,
J= 0 2 1.
1 2 1 0 0 2
Activity 2.11 Let v1 , v2 , v3 denote the three columns of P , in order. Verify that
v1 is an eigenvector for A corresponding to eigenvalue 2, and that
Av2 = v1 + 2v2 and Av3 = v2 + 2v3 . (Given that the three vectors are linearly
independent, J then represents A with respect to the basis {v1 , v2 , v3 }.)
20
2.5. Solving systems of differential equations using Jordan normal form
Let us look at another example and see how we might generalise from it.
and if
0 1 0 −3
1 10 0 4
P =
0 0 1 0 ,
0 0 0 1
then P −1 AP is the Jordan matrix
2 0 0 0
B 0 0
0 2 1 0 1
J =0 0 = 0 B2 0 .
2 0
0 0 B3
0 0 0 3
z10 = 2z1
z20 = 2z2 + z3
z30 = 2z3
z40 = 3z4 .
Clearly the first, third and final equations can be solved directly just as in the
diagonalisable case:
z1 = c1 e2t , z3 = c3 e2t , z4 = c4 e3t .
It turns out that the solution for z2 is z2 = c2 e2t + c3 te2t . We can then determine y
by using y = P z.
Now, looking at the explicit equations for the zi in this example, we can see that really
there are three separate ‘sub-systems’ to solve: namely z10 = 2z1 , which is directly
solvable, then
z20 = 2z2 + z3
z30 = 2z3 ,
which involves only the functions z2 and z3 , and finally z40 = 3z4 , which is directly
solvable. You can probably see in general that if we are attempting to solve
B1 0 ··· 0
0 B2 ··· 0
z0 = Jz =
... .. .. . z,
. . ..
0 0 · · · Br
21
2. Diagonalisation, Jordan normal form and differential equations
where each Bi is a Jordan block, then we can separate this into r sub-systems for the zi
which can be solved separately (and having solved these, we then determine the
solution y = P z). Let us focus on any particular sub-system. It will be of the form
2 w0 = Bw where w consists of some of the zi and where
λ 1 0 ... ... 0 0
0 λ 1 0 ... 0 0
0 0 λ 1 ... 0 0
. . . .
. . . . . . . . ... ... ,
B= .. .. ..
.. .. .. . . . λ 1 0
0 0 0 ... ... λ 1
0 0 0 ... ... 0 λ
0 0 0 ... ... 0 λ
is given by
wk = ck eλt
wk−1 = ck−1 eλt + ck teλt
t2 λt
wk−2 = ck−2 eλt + ck−1 teλt + ck e ,
2
and, in general, for j = 1, . . . , k,
λt λtt2 λt tk−j λt
wj = cj e + cj+1 te + cj+2 e + · · · + ck e .
2 (k − j)!
Proof
We have the equation wk0 = λwk , so certainly wk = ck eλt for some ck , and that concurs
with the statement of the theorem. Consider a general value of j between 1 and k − 1
and suppose that we already know that wj+1 is as stated in the theorem. Now, we have
wj0 = λwj + wj+1 . Multiplying both sides of this equation by e−λt and rearranging gives
22
2.5. Learning outcomes
Now, the left-hand side is easily seen to be (e−λt wi )0 , so we must have that e−λt wi is the
(indefinite) integral of e−λt wj+1 . Given that
e −λt
wj+1
t2
= cj+1 + cj+2 t + cj+3 + · · · + ck
tk−j−1
,
2
2 (k − j − 1)!
it follows that
Z
−λt
e wj = wj+1 dt
t2 tk−j
= cj + cj+1 t + cj+2 + · · · + ck ,
2 (k − j)!
where cj is a constant of integration. The expression for wj then follows by multiplying
both sides by eλt . This ‘inductive’ argument proves the theorem.
Activity 2.12 Now go back to the examples given earlier where we merely
presented the solutions of the differential equations, and convince yourself that the
theorem just proved shows why these are indeed the solutions.
Learning outcomes
At the end of this chapter and the relevant reading, you should be able to:
Exercises
You should now attempt the following Exercises from Anthony and Harvey: Exercises
9.7, 9.8, 9.9, 9.17, 9.18, 9.19 and 9.20.
23
2. Diagonalisation, Jordan normal form and differential equations
Now, attempt the following exercises on the Jordan normal form and its applications to
solving systems of differential equations.
2 Exercise 2.1
Let
0 0 1
A = 1 0 −3 .
0 1 3
Find the eigenvalues of A and show that A cannot be diagonalised. Let
1 1 0
v1 = −2 , v2 = −3 , v3 = 0 .
1 2 1
Av2 = v1 + v2 , Av3 = v2 + v3 .
Exercise 2.2
Find the functions y1 (t), y2 (t), y3 (t) which are such that y1 (0) = 1, y2 (0) = 1 and
y3 (0) = 1 and which satisfy
y10 = −y1 + y2
y20 = −y2 + y3
y30 = −y3 .
Exercise 2.3
Write down the general solution to the following system of differential equations:
0
y1 1 0 0 0 0 0 0 0 y1
y20 0 2 1 0 0 0 0 0 y2
0
y3 0 0 2 1 0 0 0 0 y3
0
y4 0 0 0 2 1 0 0 0 y4 .
0=
y 0 0 0 0 2 1 0 0 y5
50
y 0 0 0 0 0 3 1 0 y6
6
y70 0 0 0 0 0 0 3 1 y7
y80 0 0 0 0 0 0 0 3 y8
Exercise 2.4
Let
1 −1 0
A= 1 3 0 .
−2 1 −1
Verify that (0, 0, 1)T is an eigenvector of A. Verify also that u = (−1, 1, 1)T is an
eigenvector of A and that if v = (0, 1, 0), then Av = 2v + u. Hence find an invertible
24
2.5. Test your knowledge and understanding
matrix P and a Jordan matrix J such that P −1 AP = J. Hence find the general solution
to the following system of differential equations:
0
y1 1 −1 0
y1 2
y20 = 1 3 0 y2 .
y30 −2 1 −1 y3
Comments on exercises
See Anthony and Harvey for solutions to Exercises 9.7, 9.8, 9.9. Solutions to Exercises
9.17, 9.18, 9.19 and 9.20 may be found on the VLE pages for this course. Here are the
solutions to the remaining exercises.
Solution to exercise 2.1
There is one eigenvalue, λ = 1, but you will find that there are not three linearly
independent eigenvectors. (We omit the details here.) The required verifications are
straightforward. The three vectors are easily seen to form a linearly independent set,
and hence a basis of R3 . With respect to that basis, the transformation represented by
A with respect to the standard basis will, when represented with respect to the basis
{v1 , v2 , v3 }, be the matrix
1 1 0
J = 0 1 1
0 0 1
by the theory you will already have studied in MT1173 Algebra. That means that if
we take the matrix P with these vectors as its columns, then P −1 AP will equal J. This
P is
1 1 0
P = −2 −3 0 .
1 2 1
t2
y1 = c1 e−t + c2 te−t + c3 e−t , y2 = c2 e−t + c3 te−t , y3 = c3 e−t .
2
The initial values then easily establish that c1 = c2 = c3 = 1. So the answer is
t2 −t
y1 = e−t + te−t + e , y2 = e−t + te−t , y3 = e−t .
2
25
2. Diagonalisation, Jordan normal form and differential equations
y50 0 0 0 2 y5
y1 = c1 et
t2 t3
y2 = c2 e2t + c3 te2t + c4 e2t + c5 e2t
2 6
2
t
y3 = c3 e2t + c4 te2t + c5 e2t
2
y4 = c4 e2t + c5 te2t
y5 = c5 e2t
t2
y6 = c6 e3t + c7 te3t + c8 e3t
2
y7 = c7 e3t + c8 te3t
y8 = c8 e3t .
So, P −1 AP = J where
−1 0 0
P = 1 1 0.
1 0 1
To solve the system of differential equations, we make the substitution y = P z. Then
we will have z0 = Jz. That is,
0
z1 2 1 0 z1
z20 = 0 2 0 z2 .
z30 0 0 −1 z3
26
2.5. Feedback on selected activities
27
2. Diagonalisation, Jordan normal form and differential equations
Any non-zero scalar multiple of the vector (3, 1)T is an eigenvector of A for eigenvalue 2.
It follows, then, that if
5 3
P = ,
2 1
then P is invertible and
−1 1 0
P AP = D = .
0 2
The rows of the matrix (P z)0 are the derivatives of these linear combinations of
functions, so the i-th row is
(pi1 z1 (t) + pi2 z2 (t) + · · · + pin zn (t))0 = pi1 z10 (t) + pi2 z20 (t) + · · · + pin zn0 (t),
using the properties of differentiation, since the entries pij of P are constants. But
is just the i-th row of the n × 1 matrix P z0 , so these matrices are equal.
so the solution is
y1 (t) = −4e−2t + 2et + 3e3t
y2 (t) = −2et + 3e3t
y3 (t) = −4e−2t − 2et + 6e3t .
28
Chapter 3
Inner products and orthogonality
3
Reading
R
Essential reading
Anthony, M. and M. Harvey. Linear Algebra: Concepts and Methods. Chapter 10.
R
Further reading
Anton, H. and C. Rorres. Elementary Linear Algebra. Sections 6.1, 6.2 and 6.3.
It is important to realise that the inner product is just a number, not another vector or
a matrix.
The inner product on Rn satisfies certain basic properties as shown in the next theorem.
In R2 and in R3 the inner product is closely linked with the geometric concepts of length
and angle. This provides the background for generalising these concepts to any vector
space V , as we shall see in the next section.
29
3. Inner products and orthogonality
Activity 3.1 Prove properties (ii) and (iii). Show, also, that these two properties
are equivalent to the single property
y 6
(a1 , a2 )
>
a2
-
(0, 0) a1 x
30
3.1. The inner product of two real vectors
Its length, denoted by kak, can be calculated using Pythagoras’ theorem applied to the
right-angled triangle shown above:
q
kak = a21 + a22
so that
kak2 = ha, ai .
3
If a, b are two vectors in R2 , let θ denote the angle between them, 0 ≤ θ ≤ π. The
vectors a, b and c = b − a form a triangle, where c is the side opposite the angle θ. The
law of cosines applied to this triangle gives us the important relationship stated in the
following theorem.
Theorem 3.2 Let a, b ∈ R2 and let θ denote the angle between them. Then
ha, bi = kak kbk cos θ .
The vectors a, b and c = b − a in R3 lie in a plane and the law of cosines can still be
applied to establish the result that
ha, bi = kak kbk cos θ .
31
3. Inner products and orthogonality
where k = ha, pi is a constant. This is the equation of a plane in R3 . This plane consists
of all vectors x for which x − p is orthogonal to the given vector a. The vector a is then
called a normal to the plane. If k = 0 the plane contains the vector 0. It has equation
ha, xi = 0 and defines a (two-dimensional) subspace of R3 .
As a set of points which we can graph in R3 , taking the endpoints X = (x1 , x2 , x3 ) of all
the vectors x, we can think of ha, x − pi = 0 as the plane passing through the point
3 P = (p1 , p2 , p3 ) and normal to the vector a, meaning that the line segments from X to
P are all perpendicular to a.
In general, if a and p are given vectors in Rn , we define the set of all vectors x ∈ Rn
which satisfy the equation ha, x − pi = 0 to be the hyperplane which contains the point
P and for which the normal vector is a. It has the equation
a1 x1 + a2 x2 + · · · + an xn = k, k = ha, pi is a constant.
(iii) hx, xi ≥ 0 for all x ∈ V , and hx, xi = 0 if and only if x = 0, the zero vector of the
vector space V .
Some other basic facts follow immediately from this definition: for example,
32
3.2. Inner products more generally
It is a simple matter to check that the standard inner product on Rn defined in the
previous section is indeed an inner product according to this more general definition.
The abstract definition, though, applies to more than just the vector space Rn , and
there is some advantage in developing results in terms of the general notion of inner 3
product. If a vector space has an inner product defined on it, we refer to it as an inner
product space.
Example 3.2 (This is a deliberately strange example. It is not one you would
necessarily come up with, but its purpose is to illustrate how we can define inner
products in non-standard ways, which is why we have chosen it.) Suppose that V is
the vector space consisting of all real polynomial functions of degree at most n; that
is, V consists of all functions p : x 7→ p(x) of the form
p(x) = a0 + a1 x + a2 x2 + · · · + an xn , a0 , a1 , . . . , an ∈ R.
The addition and scalar multiplication are defined pointwise. (See Section 7.1.1 of
MT1173 Algebra.) Let x1 , x2 , . . . , xn+1 be n + 1 fixed, different, real numbers, and
define, for p, q ∈ V ,
n+1
X
hp, qi = p(xi )q(xi ).
i=1
Then this is an inner product. To see this, we check the properties in the definition
of an inner product. Property (i) is clear. For (iii), we have
n+1
X
hp, pi = p(xi )2 ≥ 0.
i=1
Clearly, if p is the zero vector of the vector space (which is the identically-zero
function), then hp, pi = 0. To finish verifying (iii) we need to check that if hp, pi = 0
then p must be the zero function. Now, hp, pi = 0 must mean that p(xi ) = 0 for
i = 1, 2, . . . , n + 1. So p(x) has n + 1 different roots. But p(x) has degree no more
than n, so p must be the identically-zero function. (A non-zero polynomial of degree
at most n has no more than n distinct roots.) Part (ii) is left to you.
33
3. Inner products and orthogonality
Definition 3.2 (Norm) Suppose that V is an inner product space and x is a vector in
V . Then the norm of x, denoted kxk, is
p
kxk = hx, xi.
34
3.2. Inner products more generally
hx, yi
cos θ = .
kxk kyk
This definition will only make sense if we can show that this number cos θ is between
−1 and 1. But this follows immediately from the Cauchy-Schwarz inequality, which can
be stated as
hx, yi
kxk kyk ≤ 1 .
Example 3.3 With the usual inner product on R4 , the vectors x = (1, −1, 2, 0)T
and y = (−1, 1, 1, 4)T are orthogonal.
35
3. Inner products and orthogonality
At first it appears that this definition has little to do with the geometric concept of
3 orthogonality. But, as we shall see, it is closely related. If P is an orthogonal matrix,
then P T P = I, the identity matrix. Suppose that the columns of P are x1 , x2 , . . . , xn .
Then the fact that P T P = I means that xTi xj = 0 if i 6= j and xTi xi = 1.
To see this, consider the case n = 3. Then, P = (x1 x2 x3 ) and since I = P T P we have
T T
1 0 0 x1 x1 x1 xT1 x2 xT1 x3
0 1 0 = xT2 (x1 x2 x3 ) = xT2 x1 xT2 x2 xT2 x3 .
0 0 1 xT3 xT3 x1 xT3 x2 xT3 x3
But, if i 6= j, xTi xj = 0 means precisely that the columns xi , xj are orthogonal. The
second statement is that kxi k2 = 1, which means (since kxi k ≥ 0) that kxi k = 1; that is,
xi is of norm 1. The converse is also true. If hvi , vj i = 0 for i 6= j and hvi , vi i = 1 then
it follows that P T P = I. This indicates the following characterisation:
A matrix P is orthogonal if and only if, as vectors, its columns are pairwise orthogonal,
and each has length 1.
Definition 3.5 (Orthonormal) A set of vectors {x1 , x2 , . . . , xk } in an inner product
space V such that any two different vectors are orthogonal and each vector has length 1:
ai = hw, vi i.
36
3.4. Gram-Schmidt orthonormalisation process
It therefore follows that the above theorem is also true if column is replaced by row: A
matrix P is orthogonal if and only if the columns (or rows) of P form an orthonormal
basis of Rn .
3
3.4 Gram-Schmidt orthonormalisation process
Given a set of linearly independent vectors {v1 , v2 , . . . , vk }, the Gram-Schmidt
orthonormalisation process is a way of producing k vectors that span the same space as
{v1 , v2 , . . . , vk }, and that form an orthonormal set. That is, the process produces a set
{u1 , u2 , . . . , uk } such that:
Lin{u1 , u2 , . . . , uk } = Lin{v1 , v2 , . . . , vk }
Activity 3.8 Try to understand why this works. Show that w2 ⊥ u1 and conclude
that u2 ⊥ u1 . Why are the linear spans of {u1 , u2 } and {v1 , v2 } the same?
Next, we define
w3 = v3 − hv3 , u1 iu1 − hv3 , u2 iu2
and set
w3
u3 = .
kw3 k
Then {u1 , u2 , u3 } is an orthonormal set and Lin{u1 , u2 , u3 } is the same as
Lin{v1 , v2 , v3 }. Generally, when we have u1 , u2 , . . . , ui , we let
i
X wi+1
wi+1 = vi+1 − hvi+1 , uj iuj , ui+1 = .
j=1
kwi+1 k
37
3. Inner products and orthogonality
Example 3.4 In R4 , let us find an orthonormal basis for the linear span of the
three vectors
First, we have
3 v1 v1 1
u1 = =√ = v1 = (1/2, 1/2, 1/2, 1/2)T .
kv1 k 12 + 12 + 12 + 12 2
Next, we have
−1 1/2 −5/2
4 1/2 5/2
w2 = v2 − hv2 , u1 iu1 =
4 − 3 1/2 = 5/2 ,
−1 1/2 −5/2
and we set
−1/2
w2 1/2
u2 = = .
kw2 k 1/2
−1/2
(Note: to do this last step, we merely noted that a normalised vector in the same
direction as w2 is also a normalised vector in the same direction as (−1, 1, 1, −1)T ,
and this second vector is easier to work with.) Continuing, we have
0 1/2 −1/2 −2
Then,
w3
u3 = = (1/2, −1/2, 1/2, −1/2)T .
kw3 k
So
1/2 −1/2 1/2
1/2 1/2 −1/2
{u1 , u2 , u3 } = , , .
1/2 1/2 1/2
1/2 −1/2 −1/2
Activity 3.9 Verify that the set {u1 , u2 , u3 } of this example is an orthonormal set.
Learning outcomes
At the end of this chapter and the relevant reading, you should be able to:
38
3.4. Test your knowledge and understanding
αhx, yi = α(x1 y1 + x2 y2 + · · · + xn yn )
= αx1 y1 + αx2 y2 + · · · + αxn yn
= (αx1 )y1 + (αx2 )y2 + · · · + (αxn )yn = hαx, yi.
The single property hαx + βy, zi = αhx, zi + βhy, zi implies property (ii) by letting
β = 0 and then letting α = 0, and property (iii) by letting α = β = 1. On the other
hand, if properties (ii) and (iii) hold, then
39
3. Inner products and orthogonality
40
3.4. Feedback on selected activities
41