Skip to main content

Section 2.10 Matrix Algebra

In the last section we showed how to add and multiply matrices of appropriate sizes. In this section we discuss matrix subtraction and what amounts to matrix division, and various rules of algebra as they apply to the operations of addition and multiplication on matrices.

Definition 2.10.1. Additive identity.

An identity for matrix addition (or additive identity) in \(\R^{m \times n}\) is a matrix \(\mathcal{O}\in\R^{m \times n}\) such that \(\A+\mathcal{O} = \mathcal{O} + \A = \A\) for all \(\A \in \R^{m \times n}.\)

Proof.

The proof is left an as exercise.

Proof.

The proof is left an as exercise.

Definition 2.10.4. Matrix subtraction.

Suppose \(\A,\B\in\R^{m\times n}.\) Then \(\A-\B\) is defined as \(\A-\B=\A+(-1)\B.\) That is,
\begin{equation*} (\A-\B)_{ij}=A_{ij}-B_{ij} \end{equation*}
Matrix subtraction inverts matrix addition in the following sense: for \(\A,\B\in\R^{m\times n},\,\A+\B-\B=\A.\)

Proof.

The proof is left an as exercise.

Proof.

The proof is left an as exercise.

Proof.

The proof is left an as exercise.

Proof.

The proof is left an as exercise.

Proof.

The proof is left an as exercise.

Remark 2.10.10.

For real number multiplication, there is one identity \((1)\) and for any nonzero real number \(x,\) there is a unique inverse \(\displaystyle{\frac{1}{x}}\) for which \(\displaystyle{x\cdot\frac{1}{x}=\frac{1}{x}\cdot x=1.}\)
The non-commutativity of matrix multiplication prevents such simplicities in linear algebra.
Indeed for a given nonsquare matrix, identities are different from each side and inverses are defined from only one side and are highly non-unique. (Square matrices always have a unique two-sided identity and have a unique two-sided inverse if they are invertible from either side.)

Definition 2.10.11. Left- and right identities for matrix multiplication.

Let \(\A\in\R^{m\times n}.\) A left-identity \(\I_{L\A}\) for \(\A\) satisfies \(\I_{L\A}\A=\A\) and a right-identity \(\I_{R\A}\) for \(\A\) satisfies \(\A\I_{R\A}=\A.\)

Proof.

The proof is left an as exercise.

Example 2.10.13. Left-identities of \(\A\in\R^{3\times2}\).

To find all left-identities (it turns out that there will be an infinite number of them) for \(\A=\left(\begin{array}{rr}1\amp 2\\3\amp 5\\2\amp 7\end{array}\right)\) we set up the equations for the unknown entries of \(\I_{L\A}\text{:}\)
\begin{equation*} \left(\begin{array}{rrr}a\amp b\amp c\\d\amp e\amp f\\g\amp h\amp i\end{array}\right)\left(\begin{array}{rr}1\amp 2\\3\amp 5\\2\amp 7\end{array}\right)=\left(\begin{array}{rr}1\amp 2\\3\amp 5\\2\amp 7\end{array}\right). \end{equation*}
Expressing this more explicitly we have
\begin{align*} a\,\,+\,\,3b\,\,+\,\,2c\amp\,\,\,=\,\,\,1\\ 2a\,\,+\,\,5b\,\,+\,\,7c\amp\,\,\,=\,\,\,2\\ d\,\,+\,\,3e\,\,+\,\,2f\amp\,\,\,=\,\,\,3\\ 2d\,\,+\,\,5e\,\,+\,\,7f\amp\,\,\,=\,\,\,5\\ g\,\,+\,\,3h\,\,+\,\,2i\amp\,\,\,=\,\,\,2\\ 2g\,\,+\,\,5h\,\,+\,\,7i\amp\,\,\,=\,\,\,7 \end{align*}
which is a system of six \((m\cdot n)\) equations in nine \((m\cdot m)\) unknowns so three of the entries of \(\I_{L\A}\) will be arbitrary. We choose \(c,f,i\) to be arbitrary and express the other variables \(a,b,d,e,g,h\) in terms of \(c,f,i.\) After some algebra we find
\begin{equation*} \I_{L\A}=\left(\begin{array}{ccc}1-11c\amp 3c\amp c\\-11f\amp 1+3f\amp f\\11-11i\amp -3+3i\amp i\end{array}\right). \end{equation*}
We may express the set of left identities for \(\A\) as
\begin{equation*} \left\{\left.\left(\begin{array}{ccc}1-11c\amp 3c\amp c\\-11f\amp 1+3f\amp f\\11-11i\amp -3+3i\amp i\end{array}\right)\,\right|\,c,f,i\in\R\right\}. \end{equation*}
By Theorem 2.10.12 the unique right-identity for \(\A\) is \(\I_2.\)
Given \(\A\in\R^{m\times n}\) with \(m\gt n,\) the number of equations defining \(\I_{L\A}\) will always be fewer than the number of entries in \(\I_{L\A}\) so there will always be an infinite number of left-identities when defined. The analogous statement holds for right-identities.
This fact makes general left- and right-identities less useful than we might hope, so we will not dwell on left- nor right-identities though we do offer one homework problem which asks us to prove Theorem 2.10.12 and another which asks us to compute left-and right-identities for a few matrices.
We will focus generally on the case when \(\A\) has both a left-identity and a right-identity. This requires 1) that \(\A\) is square and 2) that \(\I_{L\A}\) and \(\I_{R\A}\) are equal, and also determines completely the entries of the (now unique) inverse.

Definition 2.10.14. Matrix multiplicative identity.

Let \(n\in\N.\) An identity for matrix multiplication in \(\R^{n\times n}\) is a matrix \(\I\in\R^{n \times n}\) such that \(\A\I=\I\A=\A\) for all \(\A \in \R^{n \times n}.\) Note also that \(\fa\b\in\R^n,\,\I\b=\b.\)

Definition 2.10.15.

Let \(n\in\N.\) Define \(\I_n\) as follows:
\begin{align*} \I_n=\left(\begin{array}{cccc}1\amp 0\amp \cdots\amp 0\\ 0\amp 1\amp \cdots\amp 0\\ \vdots\amp \amp \ddots\amp \vdots\\ 0\amp 0\amp \cdots\amp 1\end{array}\right) \end{align*}

Proof.

The proof is left an as exercise.
When the size is unambiguous we will just write \(\I\) instead of \(\I_n\) and when the context is clear we will use the term identity instead of additive identity or multiplicative identity.
We solve \(3x=6\) by multiplying each side by \(\displaystyle{3^{-1}=\frac{1}{3}}\) which yields \(\displaystyle{x=3^{-1}\cdot6=\frac{6}{3}=2}.\) It is natural to ask if there is such a thing as an inverse under matrix multiplication for a given \(\A\in\R^{n\times n}.\) Though we provide a definition which works for any size square matrices, we will focus for now on \(\R^{2\times 2},\) determining the condition under which \(\A\in\R^{2\times2}\) is invertible and finding a formula for \(\A^{-1}\) when the condition is met. We will later extend our treatment to \(\R^{n\times n}.\)

Definition 2.10.17. Left- and right-inverses.

Let \(\A\in\R^{m\times n}.\) A matrix \(\B\) is a left-inverse of \(\A\) if \(\B\A=\I,\) and if such a \(\B\) exists we denote it \(\A^{-1}_L.\) A matrix \(\C\) is a right-inverse of \(\A\) if \(\A\C=\I,\) and if such a \(\C\) exists we denote it \(\A^{-1}_R.\)

Proof.

The proof is left an as exercise.

Example 2.10.19. Left-inverses of \(\A\in\R^{3\times2}\).

To find all left-inverses (similar to Example 2.10.19 there will be an infinite number of them) of \(\A=\left(\begin{array}{rr}1\amp 2\\3\amp 5\\2\amp 7\end{array}\right)\) we set up the equations for the unknown entries of \(\I_{L\A}\text{:}\)
\begin{equation*} \left(\begin{array}{rrr}a\amp b\amp c\\d\amp e\amp f\end{array}\right)\left(\begin{array}{rr}1\amp 2\\3\amp 5\\2\amp 7\end{array}\right)=\left(\begin{array}{rr}1\amp 0\\0\amp 1\end{array}\right). \end{equation*}
Expressing this more explicitly we have
\begin{align*} a\,\,+\,\,3b\,\,+\,\,2c\amp\,\,\,=\,\,\,1\\ 2a\,\,+\,\,5b\,\,+\,\,7c\amp\,\,\,=\,\,\,0\\ d\,\,+\,\,3e\,\,+\,\,2f\amp\,\,\,=\,\,\,0\\ 2d\,\,+\,\,5e\,\,+\,\,7f\amp\,\,\,=\,\,\,1 \end{align*}
which is a system of four equations in six unknowns so two of the entries of \(\A^{-1}_L\) will be arbitrary. We choose \(c,f\) to be arbitrary and express the other variables \(a,b,d,e\) in terms of \(c,f.\) After some algebra we find
\begin{equation*} \A^{-1}_L=\left(\begin{array}{ccc}-5-11c\amp 2+3c\amp c\\3-11f\amp -1+3f\amp f\end{array}\right). \end{equation*}
We may express the set of left-inverses of \(\A\) as
\begin{equation*} \left\{\left.\left(\begin{array}{rrr}-5-11c\amp 2+3c\amp c\\3-11f\amp -1+3f\amp f\end{array}\right)\,\right|\,c,f\in\R\right\}. \end{equation*}
The matrix \(\A\) has no right-inverse but a matrix such as \(\B=\left(\begin{array}{rrr}1\amp 5\amp -4\\2\amp 7\amp -6\end{array}\right)\) has a right-inverse \(\B^{-1}_R\) which may be found in essentially the same manner as \(\A^{-1}_L\) above.
Given \(\A\in\R^{m\times n}\) with \(m\gt n,\) the number of equations defining \(\A^{-1}_L\) will always be fewer than the number of entries in \(\A^{-1}_L\) so there will always be an infinite number of left-identities when defined. The analogous statement holds for right-identities.
Though they are more useful than left- and right-identities, we will not focus much on left- nor right-inverses though we do offer one homework problem which asks us to compute a few left-and right-inverses and another which introduces a specific left- and right-inverse.
We will focus generally on the case when \(\A\) has both a left-inverse and a right-inverse. Then 1) \(\A\) must be square and 2) \(\A^{-1}_L\) and \(\A^{-1}_R\) must be equal, and also determines completely the entries of the (now unique) inverse.

Proof.

The proof is a homework exercise.

Definition 2.10.21. Matrix inverse.

Let \(\A\in\R^{n\times n}.\) If there exists a matrix \(\B\in\R^{n\times n}\) satisfying \(\A\B=\I\) then we say that \(\A\) is invertible (or nonsingular) with \(\B=\A^{-1},\) and we call \(\A^{-1}\) the multiplicative inverse (or just inverse) of \(\A\) and say “\(\A\)-inverse.” 
Square matrices which are not invertible are called singular.

Proof.

The proof is left an as exercise.

Proof.

The proof is left an as exercise.
So, when it exists we can find the matrix inverse for \(2\times2\) matrices and for diagonal matrices by easy computations.
For \(3\times3\) and larger matrices the process of finding the inverse, when it exists, is more complicated and addressed in Section 3.10.

Proof.

The proof is left an as exercise.
Using the matrix inverse. Although Theorem 2.10.25 yields an elegant solution to \(\A\x=\b,\) it turns out to be more efficient to solve \(\A\x=\b\) by a different method called Gaussian elimination, discussed in Section 3.1. Matrix inverses do have important real-world applications in 3D animation and beam deflection and in many theoretical applications of which we will enounter several.

Proof.

If \(\A\) is invertible, then by Definition 2.10.21, \(\A\A^{-1}=\A^{-1}\A=\I.\) Thus \(\A\) satisfies the condition to be the inverse of \(\A^{-1}\text{;}\) i.e., \((\A^{-1})^{-1}=\A.\)
Multiplication by the matrix inverse, when it exists, inverts matrix multiplication in the following senses:
\begin{equation*} \B^{-1}\B\A=\I\A=\A=\A\I=\A\B\B^{-1}. \end{equation*}
However, since matrix multiplication is not commutative we should expect neither \(\B\A\B^{-1}=\A\) nor \(\B^{-1}\A\B=\A.\) Indeed, we can construct a counterexample pair \(\A,\B\) as follows.
For \(\A=\left(\begin{array}{rrr}1\amp 1\amp 0\\0\amp 1\amp 0\\0\amp 0\amp 1\end{array}\right)\) and \(\B=\left(\begin{array}{rrr}1\amp 0\amp 0\\1\amp 1\amp 0\\0\amp 0\amp 1\end{array}\right)\) with (as the reader can check) \(\B^{-1}=\left(\begin{array}{rrr}1\amp 0\amp 0\\-1\amp 1\amp 0\\0\amp 0\amp 1\end{array}\right),\) we have (as the reader can check)
\begin{equation*} \B\A\B^{-1}=\left(\begin{array}{rrr}0\amp 1\amp 0\\-1\amp 2\amp 0\\0\amp 0\amp 1\end{array}\right) \end{equation*}
while
\begin{equation*} \B^{-1}\A\B=\left(\begin{array}{rrr}2\amp 1\amp 0\\-1\amp 0\amp 0\\0\amp 0\amp 1\end{array}\right). \end{equation*}

Proof.

The proof is left an as exercise.

Proof.

The proof is left an as exercise.

Definition 2.10.29. Matrix transpose.

Let \(\A=(A_{ij})\in\R^{m\times n}.\) Then the transpose of \(\A\) is \(\A^T:=(A_{ji})\in\R^{n\times m}.\) The operation of taking the transpose of \(\A\) is called the transposition of of \(\A.\)
We may think of the transpose of a matrix as the reflection across the diagonal \(i=j,\) even if part of that diagonal lies outside the matrix.
Transposing a column-vector \(\x\in\R^{n\times1}\) results in a row vector \(\x^T\in\R^{1\times n}\text{,}\) and vice-versa.

Example 2.10.30.

We have \(\left(\begin{array}{rrr}3\amp -2\amp -4\\6\amp -6\amp -9 \end{array} \right)^T=\left(\begin{array}{rr}3\amp 6\\-2\amp -6\\-4\amp -9 \end{array} \right).\)

Proof.

The proof is left an as exercise.

Proof.

The proof is left an as exercise.

Proof.

The proof is left an as exercise.

Proof.

\(\left(\Rightarrow\right):\) Suppose \(\A\in\R^{n\times n}\) is invertible and by Definition 2.10.21 fix \(\B=\A^{-1}\in\R^{n\times n}.\) Then \(\A\B=\I\) and by Theorem 2.10.33 \(\I=\I^T=\B^T\A^T.\) Now by Theorem 2.10.20 because both equal \(\I,\) we have \(\B^T\A^T=\A^T\B^T=\I\) so by Definition 2.10.21, \(\left(\A^T\right)^{-1}=\B^T=\left(\A^{-1}\right)^T.\)
\(\left(\Leftarrow\right):\) Letting \(\C=A^T\) and applying Theorem 2.10.31 as appropriate in the above argument yields the reverse implication.

Definition 2.10.35. Symmetric matrices.

Let \(\A\in\R^{n\times n}.\) We say that \(\A\) is symmetric if \(\A^T=\A.\)

Example 2.10.36. Symmetric and non-symmetric matrices.

The matrices
\begin{equation*} \left(\begin{array}{rr}8\amp-3\\-3\amp5\end{array}\right) \end{equation*}
and
\begin{equation*} \left(\begin{array}{rr}2\amp-1\amp7\\-1\amp4\amp-6\\7\amp-6\amp0\end{array}\right) \end{equation*}
are symmetric, while
\begin{equation*} \left(\begin{array}{rr}8\amp-4\\-3\amp5\end{array}\right) \end{equation*}
and
\begin{equation*} \left(\begin{array}{rr}2\amp-1\amp8\\-1\amp4\amp-6\\7\amp-6\amp0\end{array}\right) \end{equation*}
are not symmetric.

Example 2.10.37. The Gram matrix.

For \(\A\in\R^{m\times n}\) the matrix
\begin{equation*} \A^T\A=\left(\dpr{\A_{i\ast}}{\A_{\ast j}}\right) \end{equation*}
is called the Gram matrix. It is symmetric since
\begin{align*} \qquad\qquad\left(\A^T\A\right)^T=\amp\,\A^T\A^{TT}\qquad\qquad\amp\amp\text{ by } \knowl{./knowl/xref/producttranspose.html}{\text{Theorem 2.10.33}}\\ \qquad\qquad=\amp\A^T\A\qquad\qquad\amp\amp\text{ by }\knowl{./knowl/xref/transposetranspose.html}{\text{Theorem 2.10.31}}. \end{align*}

Proof.

The proof is left an as exercise.

Definition 2.10.39. Skew-symmetric matrices.

Let \(\A\in\R^{n\times n}.\) We say that \(\A\) is skew-symmetric if \(\A^T=-\A.\)

Proof.

The proof is left an as exercise.

Proof.

Let \(\A\in\R^{n\times n}\text{;}\) we have
\begin{align*} \qquad\qquad\B^T=\amp\,\A^T+\A^{TT}\qquad\qquad\amp\amp\text{ by } \knowl{./knowl/xref/sumtranspose.html}{\text{Theorem 2.10.32}}\\ \qquad\qquad=\amp\,\A^T+\A\qquad\qquad\amp\amp\text{ by }\knowl{./knowl/xref/transposetranspose.html}{\text{Theorem 2.10.31}}\\ \qquad\qquad=\amp\,\A+\A^T=\B\qquad\qquad\amp\amp \end{align*}
and
\begin{align*} \qquad\qquad\C^T=\amp\,\A^T-\A^{TT}\qquad\qquad\amp\amp\text{ by } \knowl{./knowl/xref/sumtranspose.html}{\text{Theorem 2.10.32}}\\ \qquad\qquad=\amp\,\A^T-\A\qquad\qquad\amp\amp\text{ by }\knowl{./knowl/xref/transposetranspose.html}{\text{Theorem 2.10.31}}\\ \qquad\qquad=\amp\,-(\A-\A^T)=-\C.\qquad\qquad\amp\amp \end{align*}
Since \(\B=\A+\A^T\) satisfies Definition 2.10.35 and \(\C=\A-\A^T\) satisfies Definition 2.10.39, the theorem is proved.

Proof.

The proof is left an as exercise.