Recall from Definition Definition 2.4.1 that the dot product is a function which takes input from \(\R^n\times\R^n\) (ordered pairs of \(n\)-vectors) and outputs the sum of the products of corresponding components of \(\u\) and \(\v.\) The value of the dot product provides information about the relationship between the two input vectors: for example, if the dot product of nonzero \(\u,\,\v\) equals zero the vectors are orthogonal. This is evident from Equation (2.4.5): \(\dpr{\u}{\v}=\|\u\|\,\|\v\|\cos\th.\)
The determinant is a real-valued function of square matrices of any size: \(\det:\R^{n\times n}\rightarrow\R;\) the determinant of \(\A\in\R^{n\times n}\) is denoted \(\det(\A).\) That is, the determinant takes input from ordered \(n\)-tuples of \(n\)-vectors and outputs a real number computed as the sum of \(n\) (signed) products of the entries of a matrix \(\A,\) one entry from each column. The formula for \(\det(\A)\) is simple for \(n=2,\) a bit more complex for \(n=3,\) still more elaborate for \(n=4,\) and so on.
The determinant of \(\A\) is so named because its value (more specifically whether it is nonzero or not) determines whether or not a linear system associated with \(\A\) has a unique solution or not.
Ultimately, we will give a recursive formula for \(\det(\A),\) which will allow us to compute the determinant of a \(4\times 4\) matrix by computing four \(3\times 3\) determinants (and of course for \(n=5\) we’d need to compute five \(4\times 4\) determinants, each of which would require the computation of four \(3\times 3\) determinants, and so on). The determinant function will tell us a lot about the matrix, including whether or not it is invertible, and we will find that it satisfies many interesting properties.
Definition4.1.1.The determinant of a matrix.
The determinant of a square matrix \(\A,\) denoted \(\det(\A)\) or \(|\A|,\) is the unique function \(\det:\R^{n\times n}\longrightarrow\R\) satisfying the following properties:
For any \(n\in\N,\det(\boldsymbol{I_n})=1.\)
If \(\B\) is produced from \(\A\) by a single row-swap, \(\det(\B)=-\det(\A).\)
The determinant is linear in each row.
The notation for the determinant of \(\A\) is any of:
There is a subtlety to Definition 4.1.1, namely that a function satisfying the properties therein is unique. We omit the general proof of this uniqueness. From here on we will take the uniqueness of such a function as a given.
Definitions are crucial but do not always lead to convenient computation. We provide formulas for calculating determinants of matrices of various sizes, and show that our formulas do indeed satisfy the properties in Definition 4.1.1.
Theorem4.1.2.Determinant of a \(\boldsymbol{2\times2}\) matrix.
Let \(\A=\begin{pmatrix}a\amp\!b\\c\amp\!d\end{pmatrix}\in\R^{2\times2}.\) Then
We show that the proposed formula \(\det(\A)=ad\!-\!bc\) satisfies the properties in Definition 4.1.1.
By the proposed formula applied to the \(2\times2\) identity we have \(\left|\begin{array}{cc}1\amp 0\\0\amp 1\end{array}\right|=1\cdot1-0\cdot0=1\) so the first property in Definition 4.1.1 holds.
Swapping the rows of \(\A\) and using the proposed formula yields
Finally, we show that \(\det\) is linear in the first row by showing the effects of scalar multiplication and addition separately. Multiplying the first row through by the constant \(r,\) the proposed formula gives
Since scalar multiples of the first row factor out of the proposed formula, and the proposed formula is additive in the first row, the \(2\times2\) determinant formula satisfies linearity in the first row. Similar calculations demonstrate linearity in the second row.
Thus the proposed \(2\times2\) determinant formula \(\left|\begin{array}{cc}a\amp b\\c\amp d\end{array}\right|=ad\!-\!bc\) is valid by Definition 4.1.1.
Example4.1.3.Determinant of a \(\boldsymbol{2\times2}\) matrix.
For \(\A=\lmatrix{rr}7\amp 2\\3\amp 4\rmatrix\) we have
To show Property 1 for \(n=3,\) we have \(I_3 = \lmatrix{ccc} 1 \amp 0 \amp 0 \\ 0 \amp 1 \amp 0 \\ 0 \amp 0 \amp 1\rmatrix.\) Using Theorem 4.1.4 we see
The result for swapping any other two rows follows similarly.
Finally, to show Property 3 for \(n=3\) let \((B_{11}\,\,B_{12} \,\, B_{13})\) be a row vector and let \(r\in\R.\) We form two matrices, for the purpose of proving separately each of the effects of linearity:
\begin{align*}
\amp \lmatrix{ccc} A_{11}+B_{11} \amp A_{12}+B_{12} \amp A_{13}+B_{13} \\ A_{21} \amp A_{22} \amp A_{23}\\ A_{31} \amp A_{32} \amp A_{33} \rmatrix\\
\amp \text{where the first row is the sum } (A_{11}\,\,A_{12}\,\, A_{13})+ (B_{11}\,\,B_{12} \,\, B_{13}),\text{ and}\\
{ }^{ }\\
\amp \left(\begin{array}{rrr} rA_{11} \amp rA_{12} \amp rA_{13}\\ A_{21} \amp A_{22} \amp A_{23} \\ A_{31} \amp A_{32} \amp A_{33} \end{array} \right)\\
\amp \text{where the first row is } r(A_{11}\,\,A_{12} \,\, A_{13}).
\end{align*}
If we switch rows 1 and 3 of \(\A\) to get matrix \(\B=\lmatrix{rrr} 0 \amp 1 \amp 4\\ -2 \amp 0 \amp 1\\1 \amp -1 \amp 1 \rmatrix\) then by Definition 4.1.1 part 2 we know \(\det(\B)=-\det(\A)=11.\) We check this directly:
If we add row vector \((2\,\,2\,\,1)\) to row \(3\) of matrix \(\A\) to get \(\B=\left( \begin{array}{rrr}1 \amp -1 \amp 1 \\ -2 \amp 0 \amp 1\\ 2 \amp 3 \amp 5\end{array}\right)\) then by Definition 4.1.1
If we multiply row one of \(\A\) by 2 and row two of \(\A\) by 3 we get \(\B=\lmatrix{rrr} 2 \amp -2 \amp 2 \\ -6 \amp 0 \amp 3\\ 0 \amp 1 \amp 4\rmatrix\) and by Definition 4.1.1 part 3 (applied twice, once for each row) we get
Before we give the general formula for the determinant of an \(n \, \times \, n\) matrix let us observe some patterns in the formulas for the \(n=2\) and \(n=3\) cases. Recall
\begin{equation*}
\left| \begin{array}{rr} a \amp b \\ c \amp d \end{array} \right| = ad-bc.
\end{equation*}
Note the determinant of the \(3 \times 3\) matrix is given as a linear combination of the determinants of \(2 \times 2\) matrices that can be found within the larger matrix. Indeed if we let the dash \(-\) stand for a deleted entry the determinant of a \(3\times3\) matrix \(\A\) may be expressed a linear combination of \(2\times2\) determinants, as in for example
Definition 4.1.1 tells us that if we swap two rows of a matrix then the determinant changes by a sign. This means that if we swap any other row of our \(3 \times 3\) matrix with the top row, then the determinant changes by a sign. So for example, if we swap the first and second rows we find
In each of the terms in (4.1.2) the \(2\times 2\) matrix formed by deleting the \(i^{\text{th}}\) row and \(j^{\text{th}}\) column of the \(3 \times 3\) plays a crucial role; these matrices are defined below.
Definition4.1.7.
Let \(\A\in\R^{n \times n}\) and let \(1\le i,j\le n.\) The \((n-1)\times (n-1)\) matrix formed by deleting the \(i^{\text{th}}\) row and \(j^{\text{th}}\) column of \(\A\) is called the \(ij^{\text{th}}\,\) minor of \(\A\) and denoted by \(\boldsymbol{M_{ij}}.\)
Figure4.1.8.\(\boldsymbol{M_{ij}}:\) the minor matrix formed by eliminaing row i and column j of the matrix.
This process is called Laplacian expansion by minors (or just expansion) and is valid for square matrices of any size. In the above example line (4.1.1) we have expanded the determinant along row \(1\) and in example line (4.1.2) we have expanded the determinant along row \(2.\) We similarly could have put row \(3\) in position \(1\) and adjusting for signs we’d get a similar formula.
We are now ready to give our general formula for a determinant of an \(n\, \times \, n\) matrix \(\A \in \R^{n \times n}.\) The formula is found recursively by taking linear combinations of the determinants of the \((n-1) \, \times \, (n-1)\) minors of the matrix. Similar to the \(3 \, \times \, 3\) case we can expand by any row that we like. This is reflected in the following formula.
Theorem4.1.9.Determinant expansion by rows.
Let \(\A \in \R^{n \times n}.\) Then \(\det(\A)\) is given by
By Theorem 4.4.11\(\,\det(\A)=\det(\A^T)\) so instead of expanding by rows we could expand by a column of \(\A\) if that is more convenient. The formula is modified in the obvious way; the expansion about column \(j\) is given in the next theorem.
Theorem4.1.10.Determinant expansion by columns.
Let \(\A\in\R^{n\times n}.\) Then \(\det(\A)\) is given by
The proof is almost immediate by Theorem 4.1.9, once we have proven that for any \(\A\in\R^{n\times n},\,\det(\A^T)=\det(\A).\) We do this in the next section.
The signs of the coefficients in the linear combinations alternate. It can be useful to remember how to place the signs by visualizing the matrix of signs in an \(n\, \times\, n\) matrix starting with a \(+\) in the upper left and alternating. For example in the \(3\times3\) case:
The benefit of being able to expand by any row or column we choose is that we can choose the row or column that most simplifies our calculation. For example for \(\A=\left(\begin{array}{rrr}-5 \amp {\bf 0} \amp 2 \\ 2 \amp {\bf 2} \amp 9\\ 5 \amp {\bf 0} \amp 11 \end{array}\right)\) we would choose to expand by column \(2\) since there are two zeros:
Let \(\displaystyle{\A=\left(\begin{array}{rrrr} 1 \amp 2\amp 3 \amp -1 \\0 \amp 4 \amp -1 \amp 0 \\ 3 \amp 2 \amp 1 \amp 4\\ -1 \amp 1 \amp -1 \amp -3 \end{array} \right)}.\) To find the determinant we will expand about row \(2\) since there are two zeros: