189-251B: Honors Linear Algebra
Blog
Week 1. January 6-10.
This week was devoted to the basic definitions and properties
of abstract vector spaces over a field F, and of the
linear transformations between them.
We discussed what it means for a set of vectors to span a vector space,
to be linearly independent, and to be a basis for the vector space.
This led us to the notion of dimension of a finite-dimensional vector space.
Next week, we will discuss the final exam of last semester, and I will give
some solutions to the problems there. I advise that you have a look at the
final exam before then, in order to refresh your
memory about the questions whose solutions I will be presenting.
Week 2. January 13-17.
Monday and Wednesday saw the proof of
basic facts concerning linear independence
of vectors and spanning sets. A basis of a vector space V
was defined to be a minimal spanning set,
or, equivalently, a maximal linearly independent subset,
or, equivalently, a subset B of V such
that every vector in V can be
expressed uniquely as a (finite, by definition)
linear combination of the vectors in B.
A vector space which has a finite basis is called finite-dimensional.
We proved that the cardinality of a basis for a
given finite-dimensional vector space V depends only on V
and not on the choice of basis. This cardinality is therefore an invariant of
V caled its dimension.
We proved various useful facts about bases and linearly independent sets,
based on the slightly technical but useful Steinitz substittion lemma.
One of the particularly useful corollaries is that every linearly independent subset of a finite-dimensional vector space can be completed to a basis for V (and hence, its cardinality is less than the dimension of V.
This flexibility in the choice of basis was a key ingredient in the
rank-nullity theorem proved on Wednesday, which, given a linear
transformation defined on a finite-dimensional vector space V,
relates the dimensions of its image and kernel to the dimension of V.
Friday's lecture was devoted to a quick review of the questions in the final exam from last semester.
Week 3. January 20-24.
This week was devoted to some of the basic material covered in
chapter 4 of the on-line notes, most notably the idea of representing
vectors in finite dimensional vector spaces by coordinates,
and linear transformations between such spaces as matrices,
after a suitable choice of bases of the abstract vector spaces involved.
We dicussed quotients, the isomorphism theorem for vector spaces, and the
theorem about the dimension of the kernel and image. Hopefully this discussion
left you with a feeling of deja vu from our very similar treatment
of groups last semester -- and if not, go ccarefully through sections 4.2. and 4.3. of the notes.
We illustrated some of the concrete applications of the material seen in class
by discussing linear codes, which are simply subspaces of
F2n,
viewed as vector spaces over the field {0,1} with two elements.
Friday's lecture concluded with the construction of the remarkable
Hamming code, the (essentially unique)
error correcting code of length 7 and dimension 4.
Week 4. January 27-31.
This week was devoted to the study of a single linear transformation
T from a vector space V to itself.
We saw that one can associate to such a T its minimal polynomial, and we proved that, when V has dimension n, the
degree of this minimal polynomial is at most n.
We also defined eigenvalues and eigenvectors, and proved that the set of eigenvalues of a linear transformation is equal to the set of roots of its minimal
polynomial.
Week 5. February 3-7.
We continued our discussion of eigenvalues, eigenvectors, and diagonalisability and its connection with the characteristic polynomial. The most important theorem we proved this week is the primary decomposition theorem which asserts
that a vector space V on which a linear transformation T is defined
can be broken up into a direct sum of T-stable subspaces indexed by
the irreducible factors of the minimal polynomial. If p(x) is such an
irreducible factor, and
p(x)e divides
the minimal polynomial exactly, then the restriction of T to the associated stable subspace has
p(x)e as its minimal polynomial.
From this we deduced that a linear transformation is diagonalisable if and only if
its minimal polynomial factors into a product of distinct linear factors.
Week 6. February 10-14.
This week was devoted to a discussion of determinants, which were treated
somewhat differently from the on-line notes. By way of motivation, we started
with a discussion of the notion of volume on the real vector space
Rn. (This volume is usually
called the length when n=1,
the area when n=2, the volume when n=3, and continues to be
designated as a volume for all larger values of n; this is just for
lack of further words in the english vocabulary, and the
n-dimensional
volume differs just as much from its three-dimensional avatar as the surface
area does from the length....
We saw that the function which to (v1, ..., vn)
associates the signed volume
of the n-dimensional parallelopiped spanned by
these n vectors satisfies two properties: it is multilinear in
the vi, and it is alternating. Remarkably, these
two properties are purely algebraic and
make sense over an arbitrary field F; indulging in the
mathematician's
taste for abstraction, a volume function on an n-dimensional
vector space over F
can just be defined to be a multilinear, alternating function from
Vn to F.
Even more remarkably, such volume functions are essentially unique:
any two multlinear
alternating functions in n variables on an n-dimensional
vector space just differ by a constant of proportionality.
(This important uniqueness property was proved in class by induction.)
The determinant of a transformation T: V --> V can then be defined as
the ratio between the volume
function D(T v1,...,T vn)
and
D( v1,...,vn),
for any non-zero volume function D. In effect, the determinant
of T represents the amount by which T distorts (any)
volume on
V.
We used this somewhat abstract definition of the determinant to desribe
the well-known algorithm for computing it via row-reduction, and gave a closed
formula for the determinant of a matrix in terms of its entries, as a sum
of n! terms indexed by the elements of the symmetric group
on n letters.
With the determinant behind our belt, we were finally
able to define the
characteristic polynomial of a linear transformation T.
This polynomial, denoted fT(x), is simply the
determinant of the transformation x I - T,
where I denotes the identity tansformation on V.
The explicit formula for the determinant of a matrix shows that this function of
x is in fact a polynomial in x whose degree is always equal
to n=dim(V).