In Section 3.4 we introduced the notion of a subspace of \(\mathbb{R}^n\text{.}\) In this section we will describe three important subspaces associated to any matrix, and also see how understanding these subspaces can shed light on some earlier topics from the course.
Subsection4.6.1The row space
Definition4.6.1.
Let \(A\) be an \(m \times n\) matrix. The row space of \(A\text{,}\) denoted \(\row(A)\text{,}\) is defined to be the span of the rows of \(A\text{.}\)
Note that \(\row(A)\) is a subspace of \(\mathbb{R}^n\text{,}\) by Theorem 3.4.6.
Theorem4.6.2.
For any matrix \(A\text{,}\) the non-zero rows of \(\RREF(A)\) form a basis for \(\row(A)\text{.}\)
It follows from the definition of the reduced row echelon form that the non-zero rows of \(\RREF(A)\) form a basis for \(\row(\RREF(A))\text{,}\) so all we need to show is that \(\row(\RREF(A)) = \row(A)\text{.}\) To do that, it suffices to show that if we perform one row operation on \(A\) then we do not change the span of the rows. The verification that this is true is left as an exercise.
By Theorem 4.6.2 a basis for \(\row(A)\) is \(\left\{\begin{bmatrix}1 \\ 0 \\ 0 \\ 3/2\end{bmatrix}, \begin{bmatrix}0 \\ 1 \\ 0 \\ -1/2\end{bmatrix}, \begin{bmatrix}0 \\ 0\\ 1 \\-2\end{bmatrix}\right\}\text{.}\)
Note4.6.4.
Warning: The rows of \(A\) corresponding to the non-zero rows of \(\RREF(A)\) are not necessarily a basis for \(\row(A)\text{.}\) Indeed, in the example above, the first three rows of \(A\) itself are \(not\) a basis for \(\row(A)\text{.}\)
Theorem4.6.5.
For any matrix \(A\text{,}\)\(\dim(\row(A)) = \rank(A)\text{.}\)
By Theorem 4.6.2, \(\dim(\row(A))\) is equal to the number of non-zero rows of \(\RREF(A)\text{.}\) On the other hand, \(\rank(A)\) is also equal to the number of non-zero rows of \(\RREF(A)\text{,}\) by Lemma 3.3.3.
Subsection4.6.2The column space
Definition4.6.6.
Let \(A\) be an \(m \times n\) matrix. The column space of \(A\text{,}\) denoted \(\col(A)\text{,}\) is defined to be the span of the columns of \(A\text{.}\)
As with the row space, Theorem 3.4.6 tells us that \(\col(A)\) is a subspace (this time it is a subspace of \(\mathbb{R}^m\)).
One way that we could look for a basis for \(\col(A)\) is to notice that \(\col(A) = \row(A^t)\text{,}\) so we could apply Theorem 4.6.2 to \(A^t\) to give us a basis for \(\col(A)\text{.}\) However, doing this requires row reducing \(A^t\text{.}\) It is useful to have a method for finding a basis for \(\col(A)\) that can be calculated from \(\RREF(A)\text{,}\) rather than \(\RREF(A^t)\text{.}\)
Theorem4.6.7.
For any matrix \(A\text{,}\) the pivot columns of \(A\) form a basis for \(\col(A)\text{.}\)
We won't give a complete proof of this fact, but the key idea is that when we do row operations we do not change the linear dependency relations between the columns. Thus, for instance, if in \(\RREF(A)\) the third column is equal to the sum of the first two then also in \(A\) the third column is the sum of the first two, and vice versa. From the definition of the reduced row echelon form we see that the pivot columns of \(\RREF(A)\) form a basis for \(\col(\RREF(A))\text{,}\) and therefore the corresponding columns of \(A\) form a basis for \(\col(A)\text{.}\)
We see that columns \(1\text{,}\)\(2\text{,}\) and \(4\) are the pivot columns. Therefore, by Theorem 4.6.7, a basis for \(\col(A)\) is \(\left\{\begin{bmatrix}2\\1\\4\\-1\end{bmatrix}, \begin{bmatrix}-1\\1\\0\\1\end{bmatrix}, \begin{bmatrix}3\\1\\-1\\0\end{bmatrix}\right\}\text{.}\)
Note4.6.9.
Warning: It usually happens that \(\col(A) \neq \col(\RREF(A))\text{,}\) so the pivot columns of \(\RREF(A)\) are not necessarily a basis for \(\col(A)\text{.}\)
Notice that the process for the row and column spaces are different, even though both use the reduced row echelon form. When finding a basis for the row space we use the rows of \(\RREF(A)\text{.}\) When finding a basis for the column space we use the columns of \(A\text{,}\) and use \(\RREF(A)\) to determine which columns to take.
Theorem4.6.10.
For any matrix \(A\text{,}\)\(\dim(\col(A)) = \rank(A)\text{,}\) and therefore also \(\dim(\col(A)) = \dim(\row(A))\text{.}\)
From Theorem 4.6.7 we have that \(\dim(\col(A))\) is the number of pivot columns of \(A\text{,}\) which is by definition \(\rank(A)\text{.}\) Using Theorem 4.6.5 we therefore have
The conclusion of the above theorem is surprising! For instance, if \(A\) is a \(17 \times 3\) matrix then \(\col(A)\) is a subspace of \(\mathbb{R}^{17}\) and \(\row(A)\) is a subspace of \(\mathbb{R}^{3}\text{.}\) A priori it is surprising that these two subspaces should have anything to do with each other - they certainly do not contain the same vectors! Yet Theorem 4.6.10 tells us that they do have the same dimension. In particular, even though \(\col(A)\) is a subspace of \(\mathbb{R}^{17}\text{,}\) we know that \(\dim(\col(A)) = \dim(\row(A)) \leq 3\text{.}\)
Corollary4.6.12.
For any matrix \(A\text{,}\)\(\rank(A) = \rank(A^t)\text{.}\)
The column space of a matrix has a nice geometric interpretation, though we won't spend much time on this point.
Suppose that \(T : \mathbb{R}^n \to \mathbb{R}^m\) is a linear transformation. The image of \(T\) is the collection of all of the vectors in \(\mathbb{R}^m\) that actually appear as outputs of \(T\text{;}\) that is, it is the set of all vectors of the form \(T(\vec{v})\) where \(\vec{v}\) is a vector in \(\mathbb{R}^n\text{.}\) Using the fact that \(T(\vec{v}) = [T]\vec{v}\) (Theorem 4.1.13) and the fact that multiplying a matrix by a vector gives a linear combination of the columns (Definition 4.1.12), one can show that the image of \(T\) is exactly the same as the column space of \([T]\text{.}\)
Subsection4.6.3The null space
Definition4.6.13.
Let \(A\) be an \(m \times n\) matrix. The null space of \(A\text{,}\) denoted \(\NullSp(A)\text{,}\) is the collection of all vectors \(\vec{v}\) in \(\mathbb{R}^n\) such that \(A\vec{v} = \vec{0}\text{.}\)
Theorem4.6.14.
If \(A\) is any \(m \times n\) matrix then \(\NullSp(A)\) is a subspace of \(\mathbb{R}^n\text{.}\)
This shows that \(c\vec{v}\) is in \(\NullSp(A)\text{.}\)
Definition4.6.15.
Let \(A\) be an \(m \times n\) matrix. The nullity of \(A\) is \(\nullity(A) = \dim(\NullSp(A))\text{.}\)
We've seen both the null space and the nullity before, but they were in disguise. Notice that, by Theorem 4.1.19, we have that \(\vec{v}\) is in \(\NullSp(A)\) if and only if \(\vec{v}\) is a solution to the homogeneous system \([A|\vec{0}]\text{.}\) Thus to find a basis for \(\NullSp(A)\) we just take \([A|\vec{0}]\) to reduced row echelon form and extract a collection of linearly independent solutions, as we have done before. Moreover, as we have seen before, the number of free variables of the system will be the dimension of the space of solutions (see Section 3.3). These observations give us a reformulation of Theorem 3.3.4:
Theorem4.6.16.Rank-Nullity Theorem (revisited).
Let \(A\) be an \(m \times n\) matrix. Then \(\rank(A) + \nullity(A) = n\text{.}\)
Let \(A = \begin{bmatrix}3 \amp 2 \amp -1 \amp 0 \\ 1 \amp 2 \amp -1 \amp 3\end{bmatrix}\text{.}\) Then \(\RREF(A) = \begin{bmatrix}1 \amp 0 \amp 0 \amp -3/2 \\ 0 \amp 1 \amp -1/2 \amp 9/4\end{bmatrix}\text{,}\) so we have that \(\begin{bmatrix}x\\y\\z\\w\end{bmatrix}\) is in \(\NullSp(A)\) if and only if \(x=\frac{3}{2}w\) and \(y = \frac{1}{2}z - \frac{9}{4}w\text{.}\) That is,
Therefore \(\left\{\begin{bmatrix}0 \\ 1/2 \\ 1 \\ 0\end{bmatrix}, \begin{bmatrix}3/2 \\ -9/4 \\ 0 \\ 1\end{bmatrix}\right\}\) is a basis for \(\NullSp(A)\text{,}\) and \(\nullity(A) = 2\text{.}\)
In Corollary 3.3.6 we proved that any system of linear equations has either no solution, exactly one solution, or infinitely many solutions. The material we have developed in this chapter allows us to give a different proof, which also gives some information about what the solutions to a system look like.
Theorem4.6.18.
Consider a system of linear equations written in augmented matrix form as \([A|\vec{b}]\text{,}\) and suppose that \(\vec{v}\) is a solution to the system. A vector \(\vec{w}\) is a solution to the system \([A|\vec{b}]\) if and only if it can be written in the form \(\vec{w} = \vec{v} + \vec{z}\text{,}\) where \(\vec{z}\) is a vector in \(\NullSp(A)\text{.}\)
First, suppose that \(\vec{w}\) is a solution to \([A|\vec{b}]\text{.}\) Then by Theorem 4.1.19 we have \(A\vec{w} = \vec{b}\text{,}\) and since \(\vec{v}\) is also a solution we have \(A\vec{v} = \vec{b}\text{.}\) If we let \(\vec{z} = \vec{w}-\vec{v}\) and subtract the above equations, we get:
Suppose that a system \([A|\vec{b}]\) has more than one solution, say \(\vec{v_1}\) and \(\vec{v_2}\) are two different solutions. By Theorem 4.6.18 there is a vector \(\vec{z}\) in \(\NullSp(A)\) such that \(\vec{v_1} = \vec{v_2}+\vec{z}\text{.}\) For any scalar \(a\) the vector \(a\vec{z}\) is also in \(\NullSp(A)\text{,}\) because \(\NullSp(A)\) is a subspace. Therefore, by the other direction of Theorem 4.6.18, each of the vectors \(\vec{v_1} + a\vec{z}\) is a solution to \([A|\vec{b}]\text{.}\) Finally, since \(\vec{v_1} \neq \vec{v_2}\) we have \(\vec{z} \neq \vec{0}\text{,}\) so if \(a \neq b\) then \(\vec{v_1} + a\vec{z} \neq \vec{v_1} + b\vec{z}\text{.}\) Thus the vectors of the form \(\vec{v_1}+a\vec{z}\) are infinitely many different solutions to the system \([A|\vec{b}]\text{.}\)
Subsection4.6.4The Fundamental Theorem, again
In this section we introduced a lot of terminology related to ideas that already appeared in the Fundamental Theorem, so we take this opportunity to re-state that theorem. Each of the new items on this version of the theorem is just a rephrasing of something that was already on the list in Theorem 4.5.14.
Theorem4.6.20.Fundamental Theorem - Version 4.
Let \(A\) be an \(n \times n\) matrix. The following are equivalent:
\(\RREF(A) = I_n\text{.}\)
\(A\) is invertible.
The system \([A|\vec{0}]\) has a unique solution.
The equation \(A\vec{x} = \vec{0}\) has a unique solution.
The only vector in \(\NullSp(A)\) is \(\vec{0}\text{.}\)
For every vector \(\vec{b}\) in \(\mathbb{R}^n\text{,}\) the system \([A|\vec{b}]\) has a unique solution.
For every vector \(\vec{b}\) in \(\mathbb{R}^n\text{,}\) the equation \(A\vec{x} = \vec{b}\) has a unique solution.
The columns of \(A\) are linearly independent.
The span of the columns of \(A\) is \(\mathbb{R}^n\text{.}\)
\(\col(A) = \mathbb{R}^n\text{.}\)
\(\row(A) = \mathbb{R}^n\text{.}\)
\(\rank(A) = n\text{.}\)
\(\nullity(A) = 0\text{.}\)
\(A\) can be written as a product of a finite collection of elementary matrices.
The new entries in this version of the theorem are (5), (10), (11), and (13). (5) is just a rephrasing of (4), and (10) is a rephrasing of (9). (11) is equivalent to (2) because \(\row(A) = \mathbb{R}^n\) if and only if \(\dim(\row(A)) = n\) if and only if \(\rank(A) = n\) (by Theorem 4.6.5). Finally, (13) is equivalent to (12) by the Rank-Nullity Theorem.
Exercises4.6.5Exercises
1.
Determine the rank and nullity and find a basis of the column space, row space, and null space of each of the following matrices.
The null space of \(A\) is the solution space of the homogeneous system \(A\mathbf{x}=\mathbf{0}\text{.}\) Thus, finding a basis of the null space is the same as finding a set of basic solutions. From the reduced echelon form, we can easily find the general solution of \(A\mathbf{x}=\mathbf{0}\text{,}\) using parameters corresponding to the non-pivot columns of \(\RREF(A)\text{.}\)
Let \(A\) be a matrix and consider any echelon form of \(A\text{.}\) Then, the number of pivot entries of \(A\) does not depend on the echelon form we choose, and is called the rank of \(A\). We denote it by \(\rank(A)\text{.}\) The dimension of the null space is called the nullity of \(A\text{,}\) and is written \(\nullity(A)\text{.}\)
Let \(A\) and \(B\) be row equivalent matrices. Then \(\row(A) =\row(B)\) and \(\NullSp(A) =\NullSp(B)\text{.}\) In particular, the non-zero rows of a row echelon form of \(A\) form a basis of the row space, \(\row(A)\text{.}\)
Input: a list of \(k\) vectors \(u_1,...,u_k\in \mathbb{R}^n\text{.}\)
Output: the set of indices \(j\) such that \(u_j\) is redundant.
Algorithm: Write the vectors \(u_1,\ldots,u_k\) as the columns of an \(n\times k\) matrix, and reduce to echelon form. Every non-pivot column, if any, corresponds to a redundant vector.
2.
Find a basis and calculate the dimension of the following subspaces of \(\mathbb{R}^4 \text{.}\)
You can write the vectors as rows of a matrix. Then use Hint 4.6.5.1.3 to determine a basis of its row space.
3.
Can a \(3 \times 4 \) matrix have independent columns? Independent rows? Explain. Solution.
Recall that the dimension of the row space always coincides with the dimension of the column space. Since the matrix only has three rows, its column space can have at most dimension three. Since four vectors in a three dimensional space are always linearly dependent, the answer is: No, it cannot have independent columns. For the other statement, consider
This matrix has three rows and four columns (i.e. it is \(3 \times 4 \)) and its rows are linearly independent. So the answer is yes, it can have independent rows.
By Hint 4.6.5.1.3, we know that the number of non-zero rows in a row-echelon form of \(A\) is the number of basis vectors for the row space of \(A\text{,}\) i.e. the dimension of \(\row (A)\text{.}\) Recall the definition of \(\rank (A)\) from Hint 4.6.5.1.2: since \(\rank(A)\) is the number of pivot entries in a row-echelon form of \(A\text{,}\) we see that \(\dim \row (A)= \rank(A)\text{,}\) which was assumed to be \(2\text{.}\) Since the matrix has \(4\) rows but its row space is only of dimension \(2\text{,}\) it cannot have linearly independent rows. Similarly, we know that \(\dim\row(A)=\dim\col(A)\text{,}\) so the column space of \(A\) has dimension \(2\text{,}\) which is strictly smaller than the number of columns, \(3\text{.}\) Thus, it cannot have linearly independent columns.
If \(A \) is an \(m \times n \) matrix and \(\rank(A) = m \text{,}\) show that \(m \le n \text{.}\)Solution.
As we have argued in Solution 4.6.5.3.b.1, \(\rank (A) = \dim\row(A)\text{,}\) and this number also coincides with \(\dim\col(A)\text{.}\) Thus, \(\dim\col(A) = m\) by assumption. This means, in particular, that among the \(n\) many columns of \(A\text{,}\) there must exist \(m\) many linearly independent ones. Thus, \(m \le n \text{.}\)
Can a nonsquare matrix have its rows independent and its columns independent? Explain. Solution.
Suppose \(A\) is of size \(m\times n\text{.}\) If all rows of \(A\) are linearly independent, then the row space dimension the number of rows in \(A\text{,}\) i.e. \(\dim\row(A)=m\text{.}\) Similarly, if all columns of \(A\) are linearly independent, then the column space dimension the number of columns in \(A\text{,}\) i.e. \(\dim\col(A)=n\text{.}\) Since we always have \(\dim\row(A) = \dim\col(A)\text{,}\) this implies \(m=n\text{,}\) i.e. \(A\) is square. Thus, the answer is no.
Thus, if \(\nullity(A)=2\text{,}\) then the rank of the matrix would be \(4\text{.}\) Since \(\rank(A) = \dim\row(A)\text{,}\) this would imply that \(A\) must have \(4\) independent rows; this is impossible since \(A\) only has \(3\) rows.
Suppose that \(A \) is \(5 \times 4 \) and \(\NullSp(A) = \SpanS(\vec{x}) \) for some column vector \(\vec{x}\ne \mathbf{0} \text{.}\) Can \(\dim(\col(A))=2? \)Solution.
No. The given information tells us that \(\dim(\NullSp(A)) = 1\text{,}\) i.e., \(\nullity(A) = 1\text{.}\) By Theorem 4.6.10 we have \(\rank(A) = \dim(\col(A))\text{,}\) and so by Theorem 4.6.16 we have