lecture notes on determinants
1 Introduction.
The determinant operation is an algebraic formula involving addition and multiplication that combines the entries of an matrix of numbers into a single number. The determinant has many useful and surprising properties. In particular the determinant “determines” whether or not a matrix is singular. If the determinant is zero, the matrix is singular; if not, the matrix is invertible.
2 Notation.
We can regard an matrix as an entity in and of itself, as a collection of numbers arranged in a table, or as a list of column vectors:
where
Correspondingly, we employ the following notation for the determinant:
3 Defining properties.
The determinant operation obeys certain key properties. The correct application of these rules allows us to evaluate the determinant of a given square matrix. For the sake of simplicity, we describe these rules for the case of the and determinants. Determinants of larger sizes obey analogous rules.
-
1.
Multi-linearity. The determinant operation is linear in each of its vector arguments.
-
(a)
The determinant distributes over addition. Thus, for a matrix and a column vector , we have
Thus, if are two matrices the determinant of the sum will have a total of terms (think FOIL):
Warning: the formula is most certainly wrong; it has the F and the L terms from FOIL, but is missing the O and the I terms. Remember, the determinant is not linear; it’s multi-linear!
-
(b)
Scaling one column of a matrix, scales the determinant by the same amount. Thus, for a scalar , we have
Similarly,
Let’s see what happens to the determinant if we scale the entire matrix:
-
(c)
A matrix with a zero column has a zero determinant. For example,
-
(a)
-
2.
Skew-symmetry. A multi-variable function or formula is called symmetric if it does not depend on the order of the arguments/variables. For example, ordinary addition and multiplication are symmetric operations. A multi-variable function is said to be skew-symmetric if changing the order of any two arguments changes the sign of the operation. The 3-dimensional cross-product is an example of a skew-symmetric operation:
Likewise, the determinant is a skew-symmetric operation, albeit one with arguments.
Thus, for a matrix we have
There are six possible ways to rearrange the columns of a matrix. Correspondingly, for a matrix we have
The determinants in the 3rd line are equal to because, in each case, the matrices in question differ from by 2 column exchanges.
Skew-symmetry of the determinant operation has an important consequence: a matrix with two identical columns has zero determinant. Consider, for example a matrix with identical first and second columns. By skew-symmetry, if we exchange the first two columns we negate the value of the determinant:
Therefore, is equal to its negation. This can only mean that .
-
3.
The identity rule. The determinant of the identity matrix is equal to . Written out in symbols for the case, this means that
4 Evaluation.
The above properties dictate the rules by which we evaluate a determinant. Overall, the process is very reminiscent of binomial expansion — the algebra that goes into expanding an expression like . The essential difference in evaluating determinants is that for scalar algebra the order of the variables is unimportant: . However, when we evaluate determinants, the choice of order can introduce a minus sign, or result in a zero answer, if some of the arguments are repeated.
Let’s see how evaluation works for determinants.
In the above evaluation we used skew-symmetry 3 times:
At the very end, we also used the identity rule. It is useful to contrast the above manipulations with the expansion of
The distributive step is the same. The different answers arise because ordinary multiplication is a symmetric operation.
Next, let’s see how to evaluate a determinant.
The symmetric analogue would be an expansion of the form
In both cases, if we were to expand fully, we would get an expression involving terms. However, skew-symmetry tells us that a determinant that involves repeated standard vectors will equal to zero. Thus, from the 27 terms we only keep the 6 terms that involve all three standard standard vectors:
Using skew symmetry once again (one flip gives a minus sign; two flips give a plus sign), and the identity rule we obtain
Notice that each of the 6 terms in the above expression involves an entry from all 3 rows and from all 3 columns; it’s a sudoku kind of thing. The choice of sign depends on the number of column flips that takes the corresponding list of standard vectors to the identity matrix; an even number of flips gives a , an odd number a .
5 Row operations.
Above, we observed that the terms in the final expression treat rows and columns on an equal footing.
Theorem 1.
Let be a square matrix. Then, .
In other words, transposing a matrix does not alter the value of its determinant. One consequence of the row-column symmetry is that everything one can say about determinants in terms of columns, one can say using rows. In particular, we can state some useful facts about the effect of row operations on the value of a determinant.
Again, for the sake of simplicity we focus on the case. Let be a matrix, expressed as a column of row vectors. The hats above the symbols are there to remind us that we are dealing with row rather than column vectors.
Below we list the effect of each of the 3 types of elementary row operations on the determinant of a matrix, and a provide an explanation.
-
1.
Row replacements do not affect the value of the determinant. Consider, for example the effect of a row replacement operation . Here is a scalar and is an elementary matrix that encodes the row operation in question:
-
2.
Row scaling operations scale the determinant by the same factor. Consider, for example, the operation :
-
3.
A row exchange operation negates the determinant. Consider, for example,the exchange of rows and :
Above, if we take to be the identity matrix, we obtain the value
for the determinant of an elementary matrix. We summarize as follows.
Theorem 2.
Let be an elementary matrix. Then,
where
Of course, a sequence of elementary row operations can be used to transform every matrix into reduced echelon form. This is a useful observation, because it gives us an alternative method for computing determinants; namely, we can compute the determinant of an square matrix by row reducing to row-echelon form. Let’s summarize this process by writing
where is a sequence of elementary row operations, and where is a matrix in reduced echelon form. If is singular, then the bottom row of will consist of zeros, and hence . Since the determinant of an elementary matrix is never zero, this implies that , also.
If is invertible, then is the identity matrix, and hence
Since the determinant of an elementary matrix is explicitly known (see Theorem 2), this gives us a way of calculating . We are also in a position to prove some important theorems.
Theorem 3.
Let be a square matrix. Then, if and only if is singular.
Theorem 4.
Let be matrices. Then,
Proof. As above, Let be the elementary matrices that row reduce to reduced echelon form;
Above, we showed that
If is singular, then the bottom row of will be zero, and so will the bottom row of . Hence, in this case, , because , also. Suppose then that is invertible. This means that is the identity matrix, and hence
However,
and the desired conclusion follows.
Theorem 5.
Let be an invertible matrix. Then
Proof. By the above theorem,
6 Cofactor expansion.
Cofactor expansion is another method for evaluating determinants. It organizes the computation of larger determinants, and can be useful in calculating the determinants of matrices containing zero entries. At this point, we introduce some useful jargon. Given an matrix , we call the matrix obtained by deleting row and column the minor of , and denote it by . We also set
and call the signed cofactor of . We are going to prove the following.
Theorem 6.
Consider a matrix
The determinant of can be obtained by means of cofactor expansion along the first column:
or, along the first row:
More generally, the determinant of an matrix can be obtained by cofactor expansion along any column :
or, along any row :
The proof works by writing
and then using multi-linearity:
The intermediate steps, namely
need to be explained.
Theorem 7.
Let be an matrix. Then,
Proof. Expanding we obtain 9 terms:
However, there can’t be any terms with a double occurrence of , and so we end up evaluating a determinant:
Similarly,
This argument generalizes to matrices of arbitrary size.
Next, by way of example, let’s consider the expansion of a matrix along the 2nd column:
The above argument generalizes to expansions along any column, and indeed to expansions of a matrix of arbitrary size. For example, for a matrix we write
and use multi-linearity to obtain
Working with row vectors, the same argument also establishes the validity of cofactor expansion along rows. For example, here is the derivation of cofactor expansion along the 2nd row of a matrix:
Here denote the elementary row vectors.
Here is a useful theorem about determinants that can be proved using cofactor expansions.
Theorem 8.
The determinant of an upper triangular matrix is the product of the diagonal entries.
Consider, for example the determinant of the following upper triangular matrix; the stars indicate an arbitrary number. We repeatedly use cofactor expansion along the first column. The multiple zeros mean that, each time, the cofactor expansion has only one term.
7 The adjugate matrix.
The cofactors of an matrix can be arranged into an matrix, called , the adjugate of . In the case we define
Note that the entries of are indexed differently than the entries of . For , the entry in the -th row and -th column is denoted by . However the cofactor is placed in row and column . Remarkably, the matrix of cofactors is closely related to the inverse of .
Theorem 9.
Let be an matrix. Then, . Furthermore, if is invertible, then .
8 Cramer’s rule
We can also use determinants and cofactors to solve a linear system , where is an invertible, square matrix. Of course, if is invertible, then a solution exists and is unique; indeed, . However, Cramer’s rule allows us to calculate directly, without first calculating and then performing a matrix-vector multiplication.
Let’s see how this works for the case of a matrix . Given a , we are searching for the numbers such that
Substituting this into the following determinant and expanding produces a useful equation:
Similarly
Therefore, the desired solution can be obtained as follows:
We generalize and summarize as follows.
Theorem 10.
Let be an invertible matrix, and a vector. Then the unique solution to the linear equation is given by
where denotes the matrix obtained by replacing column of with .
Title | lecture notes on determinants |
---|---|
Canonical name | LectureNotesOnDeterminants |
Date of creation | 2013-03-22 17:33:37 |
Last modified on | 2013-03-22 17:33:37 |
Owner | rmilson (146) |
Last modified by | rmilson (146) |
Numerical id | 6 |
Author | rmilson (146) |
Entry type | Topic |
Classification | msc 15A15 |