The determinant is a map that assigns to a square matrix \(A\) a scalar \(\det A\in\R \text{.}\) The definition given below of the determinant is far from intuitive, and we will do little to motivate it up front. Instead, we allow its various properties to speak for themselves by way of retroactive motivation. In particular, we will see that
\begin{equation*}
A \text{ is invertible } \iff \det A \ne 0\text{,}
\end{equation*}
making the determinant an important tool for investigating invertibility.
Subsection2.5.1Definition of the determinant
Our definition of the determinant is a recursive one; given an \(n\times n\) matrix \(A\) its determinant is defined in terms of the determinant of certain submatrices of dimension \((n-1)\times (n-1)\text{.}\) This necessitates some notation to help our discussion along.
Definition2.5.1.Submatrix notation.
Let \(A\) be an \(n\times n\) matrix with \(n\geq 2\text{.}\) Given \(1\leq i, j\leq n\text{,}\) the submatrix of \(A\) obtained by removing the \(i\)-th row and \(j\)-th column of \(A\) is denoted \(A_{i j}\text{.}\)
Warning2.5.2.
Do not conflate the submatrix notation \(A_{ij}\) with matrix entry notation \([A]_{ij}\text{:}\) the former returns the submatrix of \(A\) obtained by deleting the \(i\)-th row and \(j\)-th column; the latter returns the \(ij\)-th entry of \(A\text{.}\)
Definition2.5.3.The determinant.
Let \(A=[a_{ij}]_{n\times n}\text{.}\) The determinant is defined as follows:
Base case: \(n=1\)
When \(n=1\) we have \(A=[a_{11}]\) and we define \(\det A=a_{11}\text{.}\)
Let’s look at determinant formulas for the \(n=2,3\) cases. You may remember the formula for \(2\times 2\) matrices from Theorem 2.3.5; we will make the connection more explicit in Theorem 2.5.17.
Given \(A=\abcdmatrix{a}{b}{c}{d}\text{,}\) we have
The formula for the \(n=2\) case is simple enough to serve as a “second base case”, allowing us to end the recursive process of computing a general \(n\times n\) matrix once we get to expressions involving \(2\times 2\) matrices.
The recursive nature of the determinant definition makes induction arguments particularly useful when proving properties of the determinant, as illustrated by the next theorem.
Theorem2.5.5.Determinant of triangular matrices.
Let \(A=[a_{ij}]_{n\times n}\) be triangular (upper, lower, or diagonal). Then
We only give the proof for lower triangular matrices; the proof in the upper triangular case is nearly identical.
For any \(n\geq 1\) let \(P(n)\) denote the proposition: “The determinant of any \(n\times n\) lower triangular matrix is the product of its diagonal entries”. We prove by induction that \(P(n)\) is true for all \(n\geq 1\text{.}\)
Base step: show \(P(1)\) is true.
In this case \(A=[a_{11}]\text{,}\) and \(\det A=a_{11} \) is indeed the product of the diagonal entries of \(A\text{.}\)
Induction step: show \(P(n)\implies P(n+1)\) for all \(n\geq 1\).
Let \(A=[a_{ij}]_{(n+1)\times (n+1)}\) be a lower triangular matrix. Then \(a_{1j}=0\) for all \(j\geq 2\text{,}\) and hence the determinant of \(A\) is given by
for all \(1\leq i,j\leq n\text{;}\) by deleting the first row and first column we effectively bump each index up by one. Since \(A\) is lower triangular we have \(a_{ij}=0\) for all \(1\leq i \lt; j\leq n+1\text{,}\) and hence also
This follows directly from Theorem 2.5.5 since the diagonal entries of \(I\) are all ones.
Subsection2.5.2Expansion along rows and columns
Morally speaking, we should give some examples of higher-dimensional determinants, but we first introduce some theory that affords us more leeway in our computations.
Definition2.5.7.Minors and expansions along rows/columns.
Given an \(n\times n\) matrix \(A\text{,}\) for any pair \(1\leq i,j\leq n\) the \(ij\)-th minor of \(A\) is defined as
The matrix \((A_{1j})_{(i-1) k}\) is the result of first deleting row 1 and column \(j\) from \(A\text{,}\) and then deleting row \((i-1)\) and column \(k\) of the resulting matrix. To deal with such iterated submatrices, we make some simple observations relating the rows and columns of \(A_{1 j}\) and \(A_{i k}\) with those of \(A\text{.}\)
The \((i-1)\)-th row of \(A_{1 j}\) corresponds to the \(i\)-th row of \(A\text{,}\) and the first row of \(A_{i k}\) corresponds to the first row of \(A\text{.}\)
If \(k\lt j\text{,}\) then the \(k\)-th column of \(A_{1 j}\) corresponds to the \(k\)-th column of \(A\text{;}\) if \(k\geq j\text{,}\) then the \(k\)-th column of \(A_{1 j}\) corresponds to the \((k+1)\)-th column of \(A\text{.}\)
If \(j\lt k\text{,}\) then the \(j\)-th column of \(A_{ik}\) corresponds to the \(j\)-th column of \(A\text{;}\) if \(j\geq k\text{,}\) then the \(j\)-th column of \(A_{ik}\) corresponds to the \((j+1)\)-th column of \(A\text{.}\)
From these observations we derive the following table of formulas:
This completes the induction step, and thus the proof is finished.
Surprisingly, it turns out that we can compute the determinant of a matrix by expanding along any column (Corollary 2.5.10). This is a consequence of the following theorem, which is useful in its own right. The proof below is taken from Robert Beezer’s A First Course in Linear Algebra 1 . (See Theorem DT 2 .) It uses induction and a wonderful trick starting from the observation that \(c=\frac{1}{n}\sum_{i=1}^n c\) for any \(c\in\R\text{.}\)
The proof is by induction on \(n\text{.}\) The base case (\(n=1\)) is trivial since \([a]^T=[a]\) for any \(1\times 1\) matrix \([a]\text{.}\)
For induction we assume that for all \(n\geq 2\) we have \(\det B^T=\det B\) for any \((n-1)\times (n-1)\) matrix. Suppose \(A\) is an \(n\times n\) matrix. We have
This completes the proof by induction. (Note how in the second equality in the chain above we compute \(\det A^T\) in the \(i\)-th term of \(\sum_{i=1}^n\det A^T\) by expanding along the \(i\)-th row of \(\det A^T\text{.}\) A similar observation applies to the penultimate equality.)
Corollary2.5.10.Expansion along columns.
Let \(A=[a_{ij}]_{n\times n}\text{.}\) For any \(1\leq j\leq n\) we have
When expanding along a row or column, it is easy to get tripped up by the sign \((-1)^{i+j}\) in front of the \(ij\)-th coefficient. A “matrix of signs” is a sort of mnemonic device to help you in this regard. It is easily generated by observing that the sign in front of the \((1,1)\)-th entry is always a \(+\) (since \(1=(-1)^{1+1}\)), and that any horizontal or vertical step within the matrix is accompanied by a change of sign. As an example, for \(n=4\) we have the following matrix of signs:
The first statement is obvious since according to Theorem 2.5.8 and Corollary 2.5.10 we may compute the determinant by expanding along the zero row or zero column in question.
The third statement follows from the second. Indeed, if \(A\) has two identical rows or columns, then the matrix \(A'\) obtained from \(A\) by swapping the rows (or columns) in question is \(A\) itself. Thus \(\det A=-\det A\) by the second statement, and we conclude that \(\det A=0\text{.}\)
It remains only to show the second statement. We prove only the statement regarding swapping rows; the corresponding statement about columns follows from Theorem 2.5.9. The proof is by induction.
Base step: \(n=2\).
Let \(A=\begin{bmatrix}a\amp b\\c\amp d \end{bmatrix}\text{.}\) Then \(A'=\abcdmatrix{c}{d}A{b}\text{,}\) and \(\det(A')=cb-ad=-(ad-bc)=-\det A\text{.}\)
Induction step.
We assume by induction that the result holds for any \(n\times n\) matrices, \(n\geq 2\text{,}\) and show the same is true for any \((n+1)\times (n+1)\) matrix.
Let \(A\) be an \((n+1)\times (n+1)\) matrix, and suppose \(A'\) is the result of swapping the \(i\)-th and \(k\)-th rows of \(A\text{.}\) We compute the determinants of \(A\) and \(A'\) by expanding along the \(\ell\)-th row, where \(\ell\ne i\) and \(\ell\ne k\text{.}\) This is possible since \((n+1)\geq 3\text{.}\)
Moving along the \(\ell\)-th row, notice that each submatrix \({A'}_{\ell j}\) is the result of swapping the two rows of \(A_{\ell j}\) that originally corresponded to the \(i\)-th and \(k\)-th rows of \(A\text{.}\) Since these submatrices are of dimension \(n\times n\text{,}\) we have \(\det {A'}_{\ell j}=-\det A_{\ell j}\) by induction. Lastly, since the \(\ell\)-th rows of \(A\) and \(A'\) are the same we have
Let \(A\) be an \(n\times n\) matrix. The adjoint matrix of \(A\text{,}\) denoted \(\adj A\text{,}\) is the \(n\times n\) matrix whose \(ij\)-th entry is defined as follows:
Be careful of the order reversal in this definition. The \(ij\)-th entry of \(\adj A\) is equal to plus or minus the \(ji\)-th minor of \(A\text{.}\) Let’s see this in action for some small matrices.
First observe that the second statement regarding invertibility follows directly from (2.5.7), since in this case setting \(B=\frac{1}{\det A}\, \adj A\) we have
where \(C\) is the matrix obtained by replacing the \(j\)-th row of \(A\) with a copy of its \(i\)-th row. Since \(C\) has two identical rows Theorem 2.5.14 implies
Before you get too excited about the adjoint matrix formula, you should know that as \(n\) grows, this procedure becomes much more costly in terms of number of arithmetic operations involved than our inverse algorithm based on Gauss-Jordan elimination. You get a sense of this already from the previous \(3\times 3\) example. In general, the Gauss-Jordan inverse algorithm is the way to go.
Subsection2.5.3Row operations and determinant
Suppose the square matrix \(A\) can be row reduced to \(B\) via sequence of row operations. In general we do not have \(\det A=\det B\text{,}\) but we can compute \(\det A\) from \(\det B\) by keeping track of which operations are used.
Theorem2.5.20.Row operations and determinant.
Let \(A\) be an \(n\times n\) matrix. Using the notation from Definition 2.4.1 we have:
The first statement follows easily by computing \(\det (\underset{cr_i}{E}\, A)\) by expanding along the \(i\)-th row. The second statement is in fact a rephrasing of the second statement of Theorem 2.5.14. It remains to prove the third statement.
Let \(A=[a_{ij}]_{n\times n}\text{,}\) and set \(B=\underset{r_i+cr_j}{E}\cdot A\text{.}\) Then \(B\) is identical to \(A\) with the exception of the \(i\)-th row, whose \(k\)-th entry is
\begin{align*}
\det B \amp=\sum_{k=1}^n(-1)^{i+k}(a_{ik}+ca_{jk})M_{ik} \\
\amp = \sum_{k=1}^n(-1)^{i+k}a_{ik}M_{ik}+c\sum_{k=1}^n(-1)^{i+k}a_{jk}M_{ik}\\
\amp =\det A+\det C \text{,}
\end{align*}
where \(C\) is the matrix obtained by replacing the \(i\)-th row of \(A\) with a row identical with its \(j\)-th row. By Theorem 2.5.14 we conclude \(\det C=0\text{,}\) and thus
In the language of row operations, Theorem 2.5.20 translates as follows:
Scaling a row of a matrix by \(c\) has the effect of scaling the determinant by \(c\text{.}\)
Swapping two rows of a matrix changes the sign of the determinant.
Performing a row addition operation on a matrix has no effect on the determinant.
Remark2.5.22.Column operations and the determinant.
As shown in Exercise 2.5.4.26 the determinant behaves in a similar manner with respect to elementary column operations: i.e., scaling a column by a nonzero constant scales the determinant by \(c\text{,}\) swapping columns multiplies the determinant by \(-1\text{,}\) adding a multiple of one column to another leaves the determinant unchanged.
Corollary2.5.23.Determinant and products of elementary matrices.
Let \(A\) be an \(n\times n\) matrix, and suppose we have
This is an easy proof by induction on the number \(r\) of elementary matrices involved, the base case (\(r=1\)) of which is covered by Theorem 2.5.20.
Corollary 2.5.23 has both computational and theoretical applications.
On the computational side, it suggests an alternative method of computing \(\det A\text{:}\) first row reduce \(A\) to a simpler matrix \(B\text{,}\) making sure to keep track of the operations you use; set up an equation as in (2.5.9) representing the row reduction; then solve the corresponding equation (2.5.10) for \(\det A\) in terms of \(\det B\) and the \(\det E_i\text{.}\)
We consider two cases based on the invertibility of \(A\) and/or \(B\text{.}\)
\(A\) or \(B\) is not invertible.
In this case \(AB\) is not invertible (Corollary 2.4.14), and hence \(\det AB=0\) by Theorem 2.5.25. By the same reasoning we must have \(\det A=0\) or \(\det B=0\text{.}\) It follows that
In our proof of statement (2) of Theorem 2.5.14 we only showed that if \(A\) is a square matrix with two identical rows, then \(\det A=0\text{.}\) Assuming this, show that the same is true if \(A\) has two identical columns.
26.
State and prove an analogue to Theorem 2.5.20 that describes how the corresponding column operations (i.e., scale a column by \(c\text{,}\) swap two columns, column addition) affect the determinant of a matrix. (See Remark 2.5.22).