Skip to main content
Logo image

Section 3.7 Exercises (with solutions)

Exercises Exercises

1.

Let \(W =\left\{\ba{r}x_1\\x_2\\x_3\\x_4\ea \in \R^4\ \Bigg|\ x_1+2x_2+3x_3+4x_4=0\right\}\text{.}\)
(a)
Find bases for \(W\) and \(W^\perp.\)
Solution.
\(W\) is the solution space to \(Ax=0\) where \(A\) is the \(1\times 4\) matrix \(A = \ba{rrrr}1\amp2\amp3\amp4\ea\text{;}\) it is a hyperplane in \(\R^4.\) We easily read a set of independent solutions from the matrix \(A\) which is already in reduced row-echelon form. Taking \(x_2, x_3, x_4\) as free variables, we may take as a basis:
\begin{equation*} \{w_1, w_2, w_3\} = \left\{\ba{r}-2\\1\\0\\0\ea, \ba{r}-3\\0\\1\\0\ea, \ba{r}-4\\0\\0\\1\ea\right\}. \end{equation*}
Thinking the four fundamental subspaces (Theorem 3.2.17), we know that the
\begin{equation*} W^\perp = (\ker A)^\perp = C(A^*)=\Span\left\{\ba{r}1\\2\\3\\4\ea\right\}. \end{equation*}
If you did not recall that fact, it is clear that this vector is in \(W^\perp,\) but since
\begin{equation*} 4 = \dim \R^4 = \dim W + \dim W^\perp, \end{equation*}
we see we already have a spanning set.
(b)
Find orthogonal bases for \(W\) and \(W^\perp.\)
Solution.
Since \(W^\perp = \Span\left\{\ba{r}1\\2\\3\\4\ea\right\}\) is one-dimensional, the given basis is automatically an orthogonal basis.
For \(W,\) we use Gram-Schmidt: We take \(v_1 = w_1 = \ba{r}-2\\0\\0\\1\ea,\) and compute
\begin{equation*} v_2 = w_2 - \frac{\la w_2, v_1\ra}{\la v_1,v_1\ra}v_1 = \ba{r}-3/5\\-6/5\\1\\0\ea \end{equation*}
and
\begin{equation*} v_3 = w_3 - \cdots = \ba{r}-2/7\\-4/7\\-6/7\\1\ea. \end{equation*}
(c)
Find the orthogonal projection of \(b= \ba{r}1\\1\\1\\1\ea\) onto the subspace \(W\text{.}\)
Hint.
It is definitely worth noting that \(\R^4 = W \boxplus W^\perp\text{.}\) The question is, how to leverage that fact.
Solution.
The issue we want to leverage is that
\begin{equation*} \proj_W = I_V -\proj_{W^\perp}. \end{equation*}
Since we know that \(W^\perp =\Span\left\{e\right\}\) where \(e= \ba{r}1\\2\\3\\4\ea\text{,}\) we compute
\begin{equation*} \proj_{W^\perp}(b) = \frac{\la b,e\ra}{\la e,e \ra} e = \frac{10}{30}\ba{r}1\\2\\3\\4\ea= \frac{1}{3}\ba{r}1\\2\\3\\4\ea. \end{equation*}
Now using the observation, we compute
\begin{equation*} \proj_W(b) = b- \proj_{W^\perp}(b) = \ba{r} 1\\1\\1\\1 \ea- \frac{1}{3}\ba{r}1\\2\\3\\4\ea = \frac{1}{3}\ba{r}2\\1\\0\\-1\ea. \end{equation*}
One alternative is that having gone to the trouble of finding an orthogonal basis for \(W,\) we could brute force the answer from Definition 3.2.13.
Other alternatives: if we made our orthogonal basis for \(W\) into an orthonormal one, we could use Corollary 3.3.2. Or perhaps with a bit less fuss, we could simply take advantage of Proposition 3.3.3 as follows: Let
\begin{equation*} A = \ba{rrr}-2\amp-3\amp-4\\1\amp0\amp0\\0\amp1\amp0\\0\amp0\amp1\ea. \end{equation*}
Then
\begin{equation*} A (A^*A)^{-1}A^* = \ba{rrrr}\frac{29}{30} \amp -\frac{1}{15} \amp -\frac{1}{10} \amp -\frac{2}{15} \\ -\frac{1}{15} \amp \frac{13}{15} \amp -\frac{1}{5} \amp -\frac{4}{15} \\ -\frac{1}{10} \amp -\frac{1}{5} \amp \frac{7}{10} \amp -\frac{2}{5} \\ -\frac{2}{15} \amp -\frac{4}{15} \amp -\frac{2}{5} \amp \frac{7}{15} \ea \end{equation*}
is the matrix of the projection map \([\proj_W]\) with respect to the standard basis, so that
\begin{equation*} \proj_W(b) = A\ba{r}1\\1\\1\\1\ea = \frac{1}{3}\ba{r}2\\1\\0\\-1\ea. \end{equation*}
I am pretty sure which method I prefer!

2.

Let \(A = \ba{rrr} 3\amp0\amp0\\0\amp1\amp2\\0\amp2\amp1 \ea\in M_3(\R)\text{.}\)
(a)
What observation tells you that \(A\) is diagonalizable without any computation?
Solution.
It is a real, symmetric matrix so not only is it diagonalizable, it is orthogonally diagonalizable.
(b)
Compute the characteristic polynomial.
Solution.
\begin{align*} \chi_A \amp = \det(xI-A) = \det\left(\ba{rrr}x-3\amp0\amp0\\0\amp x-1\amp-2\\0\amp-2\amp x-1\ea\right)=(x-3)[(x-1)^2-4]\\ \amp = (x-3)(x^2-2x-3)=(x-3)^3(x+1). \end{align*}
(c)
Determine a basis for each eigenspace.
Solution.
\begin{align*} A+I \amp= \ba{rrr} 4\amp0\amp0\\0\amp2\amp2\\0\amp2\amp2 \ea\mapsto \ba{rrr} 1\amp0\amp0\\0\amp1\amp1\\0\amp0\amp0 \ea\mapsto v_1 =\ba{r}0\\-1\\1\ea\\ A-3I\amp= \ba{rrr} 0\amp0\amp0\\0\amp-2\amp2\\0\amp2\amp-2 \ea\mapsto \ba{rrr} 0\amp1\amp-1\\0\amp0\amp0\\0\amp0\amp 0 \ea\mapsto v_2 =\ba{r}1\\0\\0\ea v_3=\ba{r}0\\1\\1\ea \end{align*}
Note that \(v_3^\prime = \ba{r}1\\1\\1\ea\) is another obvious choice for an independent eigenvector, though not as useful for a later part (since \(v_2\) and \(v_3\) are orthogonal).
(d)
Find a matrix \(P\) so that \(P^{-1}AP\) is diagonal.
Solution.
The matrix \(P\) is any matrix with the eigenvectors as columns. For example, if we want the diagonal matrix to be
\begin{equation*} \ba{rrr}3\\\amp3\\\amp\amp-1\ea \text{ choose } P=\ba{rrr}1\amp0\amp0\\0\amp1\amp-1\\0\amp1\amp1\ea, \end{equation*}
or if we want the diagonal matrix to be
\begin{equation*} \ba{rrr}-1\\\amp 3\\\amp\amp 3\ea \text{ choose } P=\ba{rrr}0\amp1\amp0\\-1\amp0\amp1\\1\amp0\amp1\ea. \end{equation*}
Other matrices are certainly possible.
(e)
Determine whether the matrix \(A\) is orthogonally diagonalizable. If not, why; if so, find an orthogonal matrix \(Q\) so that \(Q^TAQ\) is diagonal.
Solution.
Since \(A\) is a real symmetric matrix, we know it is orthogonally diagonalizable. The columns of the matrices \(P\) above have orthogonal columns. We need only normalize the columns, say
\begin{equation*} Q = \ba{rrr}0\amp 1\amp 0\\-1/\sqrt 2\amp0\amp1/\sqrt 2\\1/\sqrt2\amp 0\amp 1/\sqrt 2\ea. \end{equation*}

3.

View \(\R^7\) as an inner product space with the usual inner product.
(a)
let \(T:\R^7 \to \R^7\) be a linear map with the property that \(\la T(v), v\ra = 0\) for all \(v \in \R^7\text{.}\) Show that \(T\) is not invertible.
Hint.
Calculus tells you that a polynomial of degree 7 and real coefficients has at least one real root.
Solution.
Let \(\chi_T\) be the characteristic polynomial of \(T\text{.}\) Since the degree is odd, the hint says \(\chi_T\) has a real root, that is, \(T\) has a real eigenvalue \(\lambda\text{.}\) Let \(v\) be a (nonzero) eigenvector with \(T(v) = \lambda v\text{.}\) We now consider the requirement that \(\la T(v), v\ra = 0\text{.}\)
\begin{equation*} \la T(v), v\ra = \la \lambda v,v\ra = \lambda \la v,v\ra = 0. \end{equation*}
Since \(v \ne 0\text{,}\) we cannot have \(\la v,v\ra = 0\text{,}\) so we must have \(\lambda = 0\text{,}\) which says zero is an eigenvalue, and hence the nullspace is nontrivial. This means that \(T\) is not invertible.
(b)
Show by example that there exist linear maps \(T:\R^2 \to \R^2\) with \(\la T(v), v\ra = 0\) for all \(v \in \R^2\text{,}\) but with \(T\) invertible. Verify that your \(T\) satisfies the required conditions.
Hint.
If we consider the previous part, the dimension only mattered to produce a real eigenvalue, so that provides a direction to look.
Solution.
Let \([T]_\cE = \ba{rr} 0\amp 1\\-1\amp 0\ea\) where \(\cE\) is the standard basis for \(\R^2\text{.}\) Then \(\chi_T=x^2+1\) which has no real roots. In particular, 0 is not an eigenvalue which means the nullspace is zero, so \(T\) is invertible. We claim that \(\la T(v),v\ra = 0\) for all \(v \in V\text{.}\) We can read off the action of \(T\) from the matrix:
\begin{equation*} T(e_1) = -e_2 \text{ and } T(e_2) = e_1, \text{ so } T(ae_1 + be_2 ) = be_1 - a e_2\text{.} \end{equation*}
We check
\begin{equation*} \la T(ae_1+be_2),ae_1+be_2\ra = \la be_1-ae_2,ae_1+be_2\ra= ba -ab = 0, \end{equation*}
for all \(a, b\text{.}\)

4.

Let \(V\) be a finite-dimensional real inner product space, and \(T: V\to V\) a linear operator satisfying \(T^2 = T\text{,}\) that is \(T(T(v)) = T(v)\) for all \(v\in V\text{.}\) To eliminate trivial situations, assume that \(T\) is neither the zero transformation, nor the identity operator.
(a)
Show that the only possible eigenvalues of \(T\) are zero and one.
Solution.
Suppose that \(T(v) = \lambda v\) for some nonzero vector \(v\text{.}\) Then
\begin{equation*} T(v)= T^2(v) = T(T(v)) = T(\lambda v) = \lambda T(v)\text{,} \end{equation*}
so \((\lambda-1)T(v) = 0\text{,}\) which means either \(\lambda = 1\) (so one is an eigenvalue), or \(T(v) = 0\) which means the nullspace is not zero, hence zero is an eigenvalue.
(b)
Let \(E_\lambda\) denote the \(\lambda\)-eigenspace. Show that \(E_0 = N(T)\text{,}\) the nullspace of \(T\text{,}\) and that \(E_1\) is the image of \(T.\)
Solution.
That \(E_0=N(T)\) is the definition of \(E_0 = \{v\in V\mid T(v) = 0 = 0v\}.\)
If \(v \in E_1\text{,}\) then \(T(v) = 1\cdot v\text{,}\) but then \(T(v) = v\) which says that \(v \in R(T)\text{.}\) Conversely if \(w = T(v^\prime) \in R(T)\text{,}\) then \(T(w) = T^2(v^\prime) = T(v^\prime) = w\text{,}\) so \(w \in E_1\text{.}\) Thus the image is precisely \(E_1.\)
(c)
Show that \(T\) is diagonalizable.
Solution.
\(\dim E_0\) equals the nullity of \(T\text{,}\) and from above \(\dim E_1\) is the rank, so by rank-nullity, the sum of the sizes of the eigenspaces (which have trivial intersection) is the dimension of the space, so \(V\) has a basis of eigenvectors for \(T.\)
(d)
Let \(W\) be a subspace of \(V,\) and let \(S = \proj_W\) be the orthogonal projection onto the subspace \(W.\) Show that \(S^2 = S,\) so that the orthogonal projection is one linear map satisfying the given property.
Solution.
By definition, we take an orthonormal basis for \(W\) (say having dimension \(r\)), and extend it to an orthonomal basis \(\cB = \{v_i\}\) for \(V\text{.}\) Then \(S(v) = \sum_{i=1}^r \la v,v_i\ra v_i = w\) and by Theorem 3.2.10 we know that \(v = w^\perp + w\) for unique \(w^\perp \in W^\perp\text{.}\) Since \(S(v) = w \in W\) and \(w = w + 0\text{,}\) \(S(w) = w\) (Corollary 3.2.14), that is \(S^2(v) = S(v)\text{.}\)

5.

Let \(A = \ba{rrr} 1 \amp 0 \amp -1 \\ -4 \amp 1 \amp 6 \\ 0 \amp -5 \amp -9 \\ 1 \amp 5 \amp 8 \ea\) and \(b = \ba{r} 1\\2\\3\\4\ea.\)
In answering the questions below, you may find some of the information below of use. By \(\rref(X)\) we mean the reduced row-echelon form of the matrix \(X\text{.}\)
\begin{equation*} \rref(A) = \ba{rrr} 1 \amp 0 \amp 0 \\ 0 \amp 1 \amp 0 \\ 0 \amp 0 \amp 1 \\ 0 \amp 0 \amp 0 \ea \qquad \rref(A|b) = \ba{rrrr}1 \amp 0 \amp 0 \amp 0 \\ 0 \amp 1 \amp 0 \amp 0 \\ 0 \amp 0 \amp 1 \amp 0 \\ 0 \amp 0 \amp 0 \amp 1 \ea \end{equation*}
\begin{equation*} A^TA = \ba{rrr} 18 \amp 1 \amp -17 \\ 1 \amp 51 \amp 91 \\ -17 \amp 91 \amp 182 \ea,\quad A^Tb= \ba{r} -3\\7\\16 \ea, \quad\rref(A^TA|A^Tb)= \ba{rrrr} 1 \amp 0 \amp 0 \amp 74 \\ 0 \amp 1 \amp 0 \amp -128 \\ 0 \amp 0 \amp 1 \amp 71 \ea \end{equation*}
\begin{equation*} AA^T=\ba{rrrr} 2 \amp -10 \amp 9 \amp -7 \\ -10 \amp 53 \amp -59 \amp 49 \\ 9 \amp -59 \amp 106 \amp -97 \\ -7 \amp 49 \amp -97 \amp 90 \ea, \ AA^Tb= \ba{r} -19 \\ 115 \\ -179 \\ 160 \ea, \ \rref(AA^T|AA^Tb)= \ba{rrrrr} 1 \amp 0 \amp 0 \amp 1 \amp 5 \\ 0 \amp 1 \amp 0 \amp 0 \amp 2 \\ 0 \amp 0 \amp 1 \amp -1 \amp -1 \\ 0 \amp 0 \amp 0 \amp 0 \amp 0 \ea \end{equation*}
(a)
Show that the system \(Ax=b\) is inconsistent.
Solution.
We see that \(\rref(A|b)\) has a pivot in the augmented column, meaning the system is inconsistent.
(b)
Find a least squares solution to the system \(Ax=b\text{.}\)
Solution.
A least squares solution to \(Ax=b\) is obtained by solving the consistent system \(A^TAx=A^Tb\text{.}\) From the work above, we read off the solution \(x=\ba{r}74\\ -128\\ 71\ea.\)

6.

Suppose a real matrix has SVD given by \(A=U\Sigma V^T\text{:}\)
\begin{equation*} A = \ba{rrr} 0\amp1\amp0\\1\amp0\amp0\\0\amp0\amp1 \ea \ba{rrrr} \sqrt 3\amp0\amp0\amp0\\0\amp\sqrt2\amp0\amp0\\0\amp0\amp0\amp0 \ea \ba{rrrr} 1/\sqrt3\amp1/\sqrt3\amp1/\sqrt3\amp0\\1/\sqrt2\amp0\amp-1/\sqrt2\amp0\\1/\sqrt6\amp-2/\sqrt6\amp1/\sqrt6\amp0\\0\amp0\amp0\amp1 \ea. \end{equation*}
(a)
Using only your knowledge of the SVD (and no compuation), determine \(\rank A\text{.}\)
Solution.
The rank is two since there are precisely two nonzero singular values, \(\sqrt 3\) and \(\sqrt2.\)
(b)
Using only your knowledge of the SVD, give a basis for the kernel (nullspace) of \(A\text{;}\) explain your process.
Solution.
The SVD process begins by finding an orthonormal basis \(\{v_1, \dots, v_4\}\) for \(A^TA\text{.}\) With \(\sigma_i = \|Av_i\|\) and the rank of \(A\) equaling 2, we know the nullity of \(A\) is also two, and since \(Av_3 = Av_4=0\text{,}\) \(\{v_3, v_3\}\) gives an orthogonal basis for the kernel.
(c)
Using only your knowledge of the SVD, give a basis for the column space of \(A\text{,}\) explaining your process.
Solution.
The column space is spanned by \(\{Av_1, \dots, Av_r\}\) where \(r = \rank A=2\text{,}\) so \(\{Av_1, Av_2\}\) is an (orthogonal) basis for the column space.

7.

Let \(A\) have singular value decomposition
\begin{equation*} A = U\Sigma V^T = \ba{rr}2/\sqrt5\amp1/\sqrt5\\1/\sqrt5\amp-2/\sqrt5\ea \ba{rr}8\amp0\\0\amp2\ea \ba{rr}1/\sqrt5\amp2/\sqrt5\\2/\sqrt5\amp-1/\sqrt5\ea. \end{equation*}
(a)
Prove that \(A\) is invertible.
Solution.
\(A\) is a \(2\times 2\) matrix with two nonzero singular values, so has rank 2, and so is invertible. Alternatively, it is easy to show that \(\det A \ne 0\text{.}\)
(b)
Using the given SVD, find an expression for \(A^{-1}\text{.}\)
Solution.
\(A = U\Sigma V^T\) implies that \(A^{-1} = (V^T)^{-1}\Sigma^{-1}U^{-1}= V \ba{rr} 1/8\amp0\\0\amp1/2\ea U^T\) since both \(U\) and \(V\)are orthogonal matrices.
(c)
The goal of this part is to find an SVD for \(A^{-1}\text{.}\) You should express your answer (confidently) as an appropriate product of matrices without multiplying things out, though you should explain why the expression you write represents an SVD for \(A^{-1}\text{.}\) In particular, a couple of warm up exercises will help in this endeavor, and no, the answer in part b is not the correct answer.
  • First show that the product of two orthogonal matrices in \(M_n(\R)\) is orthogonal.
  • Next show that the diagonal matrices (with real entries) \(\ba{rr} \lambda_1\amp0\\0\amp\lambda_2 \ea\) and \(\ba{rr} \lambda_2\amp0\\0\amp\lambda_1 \ea\) are orthogonally equivalent, i.e., that there exists an orthogonal matrix \(P\) so that
    \begin{equation*} \ba{rr} \lambda_1\amp0\\0\amp\lambda_2 \ea = P\ba{rr} \lambda_2\amp0\\0\amp\lambda_1 \ea P^T. \end{equation*}
  • Now you should be able to proceed using your answer from part b as a starting point.
Solution.
  • For the first warm up, suppose that \(AA^T=I_n=BB^T\text{.}\) Then
    \begin{equation*} (AB)(AB)^T = ABB^TA^T = AI_nA^T =AA^T = I_n. \end{equation*}
  • For the second warm up, one can choose
    \begin{equation*} P=\ba{rr}0\amp 1\\1\amp 0\ea, \end{equation*}
    but an explanation would be nice. It should be clear that the standard basis vectors \(e_1, e_2\) for \(\R^2\) are eigenvectors for the matrix. \(\ba{rr} \lambda_1\amp0\\0\amp\lambda_2 \ea\text{.}\) It follows that the matrix \(P\) with columns \(e_2, e_1\) is also a matrix of eigenvectors, but which reverses the order of appearance of the eigenvalues.
  • Now for the main event: The expression for \(A^{-1}\) in the previous part would be an SVD for \(A^{-1}\) but for the fact that the singular values do not satisfy \(\sigma_1>\sigma_2\text{.}\) Fortunately the warm up exercises come to the rescue! We see that
    \begin{equation*} \ba{rr} 1/2\amp0\\0\amp1/8 \ea = \ba{rr} 0\amp1\\1\amp0 \ea \ba{rr} 1/8\amp0\\0\amp1/2 \ea \ba{rr} 0\amp1\\1\amp0 \ea, \end{equation*}
    and that \(Q=\ba{rr} 0\amp1\\1\amp0 \ea\) is an orthogonal matrix with \(Q^T = Q\text{,}\) hence by the warm ups, so are the matrices \(QU^T = (UQ^T)^T= (UQ)^T\) and \(VQ\text{.}\) Thus
    \begin{equation*} A^{-1} = (V^T)^{-1}\Sigma^{-1}U^{-1}= V \ba{rr} 1/8\amp0\\0\amp1/2 \ea U^T = (VQ)\ba{rr} 1/2\amp0\\0\amp1/8 \ea (UQ)^T \end{equation*}
    is an SVD for \(A^{-1}\text{.}\)