Skip to main content
Logo image

Section 4.1 Basic Definitions and Examples

Listed in alphabetical order.

Definition 4.1.1.

Given an \(n\times n\) matrix \(A\) with eigenvalue \(\lambda\text{,}\) the algebraic multiplicity of the eigenvalue is the degree \(d\) to which the term \((x-\lambda)^d\) occurs in the factorization of the characteristic polynomial for \(A.\)

Example 4.1.3. Some standard bases for familiar vector spaces.

  • The standard basis for \(F^n\) is \(\cB = \{\e_1, \dots, \e_n\}\) where \(\e_i\) is the column vector in \(F^n\) with a 1 in the \(i\)th coordinate and zeroes in the remaining coordinates.
  • A standard basis for \(M_{m\times n}(F)\) is
    \begin{equation*} \cB = \{\E_{ij}\mid 1\le i\le m, 1\le j \le n\}, \end{equation*}
    where \(\E_{ij}\) is the \(m\times n\) matrix with a 1 in row \(i\) and column \(j\text{,}\) and zeroes in all other entries.
  • A standard basis for \(P_n(F)\) is \(\cB = \{1, x, x^2, \dots, x^n\},\) and a standard basis for \(P(F) = F[x]\) is \(\cB = \{1, x, x^2, x^3,\dots\}.\)

Definition 4.1.4.

The characteristic polynomial of a square matrix \(A\in M_n(F)\) is \(\chi_A = \det(xI_n-A)\text{.}\) One can show that \(\chi_A\) is a monic polynomial of degree \(n\) with coefficients in the field \(F.\)
Note that some authors define the characteristic polynomial as \(\det(A-xI_n)\) in which case the leading coefficient is \((-1)^n,\) but since the interest is only in the factorization of \(\chi_A\) (in particular any roots it may have), it does not really matter which definition one uses.

Definition 4.1.5.

The column space of an \(m\times n\) matrix \(A\) is the span of the columns of \(A.\) As such, it is a subspace of \(F^m.\)

Definition 4.1.6.

Given an \(m\times n\) matrix \(A\) with complex entries, the conjugate transpose of \(A\) is the \(n\times m\) matrix \(A^*\) whose \(ij\)-entry is given by
\begin{equation*} (A^*)_{ij} = \overline{A_{ji}} = \overline{(A^T)_{ij}}. \end{equation*}

Definition 4.1.7.

The dimension of a vector space is the cardinality (size) of any basis for the vector space.
Implicit in the definition of dimension are theorems which prove that every vector space has a basis, and that any two bases for a given vector space have the same cardinality. In other words, the dimension is a well-defined term not depending upon which basis is chosen to consider. When a vector space has a basis with a finite number of elements, it is called finite-dimensional.

Definition 4.1.8.

An elementary matrix is a matrix obtained by performing a single elementary row (or column) operation to an identity matrix.

Definition 4.1.9.

Elementary row (respectively column) operations on a matrix are one of the following:
  • Interchange two rows (resp. columns) of \(A.\)
  • Multiply a row (resp. column) of \(A\) by a nonzero scalar.
  • Replace a given row (resp. column) of \(A\) by the sum of the given row (resp. column) and a multiple of a different row (resp. column).

Definition 4.1.10.

Given an \(n\times n\) matrix \(A\) with eigenvalue \(\lambda\text{,}\) the geometric multiplicity of the eigenvalue is the dimension of the eigenspace associated to \(\lambda.\)

Definition 4.1.11.

A complex matrix \(A\) is called Hermitian if \(A = A^*.\) Necessarily the matrix needs to be square.

Definition 4.1.12.

The image of a linear map \(T:V\to W\) is
\begin{equation*} \Im(T) := \{w\in W\mid w=T(v)\text{ for some } v\in V\}. \end{equation*}
The image of \(T\) is a subspace of \(W\text{;}\) \(T\) is surjective if and only if \(W=\Im(T).\)

Definition 4.1.13.

A function \(f:X\to Y\) between sets \(X\) and \(Y\) is injective if for every \(x, x' \in X\text{,}\) \(f(x) = f(x')\) implies \(x=x'\text{.}\)

Definition 4.1.14.

Let \(F\) denote the field of real or complex numbers. For \(z = a + bi \in \C\) (\(a,b \in \R\) and \(i^2 = -1\)), we have the notion of the complex conjugate of \(z\text{,}\) denoted \(\overline z = a-bi.\) Note that when \(z\in \R\text{,}\) that is \(z = a = a + 0i \in \C\text{,}\) we have \(z = \overline z.\) The magnitude (norm, \(absolute value\)) of \(z=a+bi\) is \(|z| = \sqrt{a^2 + b^2}.\)
Let \(V\) be a vector space over the field \(F.\) An inner product is a function:
\begin{equation*} \la \cdot, \cdot \ra : V\times V \to F \end{equation*}
so that for all \(u,v,w \in V\) and \(\lambda \in F:\)
  1. \(\displaystyle \la u+v,w\ra = \la u,w\ra + \la v,w\ra\)
  2. \(\displaystyle \la \lambda v,w\ra = \lambda\la v,w\ra\)
  3. \(\overline{\la v,w\ra} = \la w,v\ra\text{,}\) where the bar denotes complex conjugate.
  4. \(\la v,v\ra\) is a positive real number for all \(v\ne 0.\)

Definition 4.1.15.

An inner product space is a vector space \(V\) defined over a field \(F = \R\) or \(\C\) to which is associated an inner product. If \(F=\R\text{,}\) \(V\) is called a real inner product space, and if \(F=\C\text{,}\) then \(V\) is called a complex inner product space.

Definition 4.1.16.

An isomorphism is a linear map which is bijective (one-to-one and onto; injective and surjective).

Definition 4.1.17.

The Kronecker delta is defined by
\begin{equation*} \delta_{ij} = \begin{cases} 1\amp \text{if }i=j\\ 0\amp \text{otherwise.} \end{cases} \end{equation*}

Definition 4.1.18.

A linear combination of vectors \(v_1, \dots, v_r \in V\) is any vector of the form \(a_1v_1 + \cdots + a_rv_r\) for scalars \(a_i \in F.\)

Definition 4.1.19.

Let \(S \subseteq V\) be a subset of vectors in a vector space \(V\) (finite or infinite). The set \(S\) is a linearly dependent subset of \(V\) if it is not linearly independent, that is there exists a finite subset \(\{v_1, \dots, v_r\} \subseteq S\text{,}\) and nonzero scalars \(a_1, \dots, a_r\) so that
\begin{equation*} a_1v_1 + \cdots + a_rv_r = \0. \end{equation*}

Definition 4.1.20.

Let \(S \subseteq V\) be a subset of vectors in a vector space \(V\) (finite or infinite). The set \(S\) is a linearly independent subset of \(V\) if for every finite subset \(\{v_1, \dots, v_r\} \subseteq S\text{,}\) a linear dependence relation of the form
\begin{equation*} a_1v_1 + \cdots + a_rv_r = \0 \end{equation*}
forces all the scalars \(a_i = 0.\)

Definition 4.1.21.

Given two vector spaces \(V\) and \(W\) (defined over the same field \(F\)), a linear map (or linear transformation) from \(V\) to \(W\) is a function \(T: V\to W\) which is
  • additive: \(T(v+v') = T(v) + T(v')\) for all \(v, v' \in V\text{,}\) and
  • preserves scalar multiplication: \(T(\lambda v) = \lambda T(v)\) for all vectors \(v\in V\) and scalars \(\lambda\text{.}\)

Definition 4.1.22.

The minimal polynomial of a square matrix \(A\in M_n(F)\) is the monic polynomial, \(\mu_A\text{,}\) of least degree with coefficients in the field \(F\) so that \(\mu_A(A) = 0\text{.}\) The Cayley-Hamilton theorem implies that the minimal polynomial divides the characteristic polynomial.

Definition 4.1.23.

A matrix \(A\in M_n(\C)\) is normal if it commutes with its conjuate transpose: \(AA^* = A^*A.\)

Definition 4.1.24.

The nullity of a linear transformation \(T:V\to W\) is the dimension of \(ker(T),\) that is, the dimension of its nullspace.
If \(T:F^n\to F^m\) is given by \(T(x) = Ax\) for an \(m\times n\) matrix \(A\text{,}\) then the nullity of \(T\) is the dimension of the set of solutions of \(Ax=0.\)

Definition 4.1.25.

The nullspace of a linear transformation \(T:V\to W\) is the kernel of \(T\) that is,
\begin{equation*} ker(T) =\{v\in V\mid T(v) = \0_W\}. \end{equation*}
If \(T:F^n\to F^m\) is given by \(T(x) = Ax\) for an \(m\times n\) matrix \(A\text{,}\) then the nullspace of \(T\) is often called the nullspace of \(A\), the set of solutions of \(Ax=0.\)

Definition 4.1.26.

A matrix \(A \in M_n(\R)\) is an orthogonal matrix if
\begin{equation*} A^TA = AA^T = I_n. \end{equation*}
Note that the condition \(A^TA=I_n\) is equivalent to saying that the columns of \(A\) form an orthonormal basis for \(\R^n\text{,}\) while the condition \(AA^T\) makes the analogous statement about the rows of \(A.\)

Definition 4.1.27.

The pivot positions of a matrix are the positions (row,column) which correspond to a leading one in the reduced row-echelon form of the matrix. The pivots are the actual entry of the given matrix at the pivot position.
The pivot columns are the columns of the original matrix corresponding to the columns of the RREF containing a leading one.

Definition 4.1.28.

The rank of a linear transformation \(T:V\to W\) is the dimension of its image, \(\Im(T).\)
If \(T:F^n\to F^m\) is given by \(T(x) = Ax\) for an \(m\times n\) matrix \(A\text{,}\) then the rank of \(T\) is the dimension of the column space of \(A.\)
By theorem, it is also equal to the dimension of the row space which is the number of nonzero rows in the RREF form of the matrix \(A.\)

Definition 4.1.29.

The row space of an \(m\times n\) matrix \(A\) is the span of the rows of \(A.\) As such, it is a subspace of \(F^n.\)

Definition 4.1.30.

Let \(A, B \in M_n(F)\text{.}\) The matrix \(B\) is said to be similar (or conjugate) to \(A\) if there exists an invertible matrix \(P\in M_n(F)\) so that \(B= P^{-1}A P.\) Note that if we put \(Q=P^{-1}\text{,}\) then \(B=QAQ^{-1}\text{,}\) so it does not matter which side carries the inverse. Also note that this is a symmetric relationship, so that \(B\) is similar to \(A\) if and only if \(A\) is similar to \(B.\) Indeed similarity (conjugacy) is an equivalence relation.

Definition 4.1.31.

Let \(S \subseteq V\) be a subset of vectors in a vector space \(V\) (finite or infinite). The span of the set \(S\text{,}\) denoted \(\Span(S)\text{,}\) is the set of all finite linear combinations of the elements of \(S\text{.}\) That is to say
\begin{equation*} \Span(S) = \{ a_1 v_1 + \cdots + a_r v_r \mid r\ge 1, a_i \in F, v_i \in S\} \end{equation*}

Definition 4.1.32.

Let \(V\) be a vector space over a field \(F,\) and let \(W\subseteq V\text{.}\) \(W\) is called a subspace of \(V\) if \(W\) is itself a vector space with the operations of vector addition and scalar multiplication inherited from \(V.\)
Of course checking all the vector space axioms can be quite tedious, but as a theorem you prove much easier criteria to check. Recall that you already know that \(V\) is a vector space, so many of the axioms (associativity, distributive laws etc) are inherited from \(V.\) Indeed, you prove that to show that \(W\) is a subspace of \(V\text{,}\) it is enough to show that the additive identity of \(V\) is in \(W,\) and that \(W\) is closed under the inherited operations of vector addition and scalar multiplication, i.e, whenever \(w, w' \in W\) and \(\lambda \in F\text{,}\) we must have \(w+w'\in W\text{,}\) and \(\lambda w\in W.\)

Definition 4.1.33.

A function \(f:X\to Y\) between sets \(X\) and \(Y\) is surjective if for every \(y\in Y\text{,}\) there exists an \(x \in X\) such that \(f(x)=y\text{.}\)

Definition 4.1.34.

A matrix \(A\) is called symmetric if \(A = A^T.\) Necessarily the matrix needs to be square.

Definition 4.1.35.

Given a square matrix \(A\in M_n(F)\text{,}\) we define its trace to be the scalar
\begin{equation*} \tr(A) := \sum_{i=1}^n A_{ii}. \end{equation*}

Definition 4.1.36.

A matrix \(A \in M_n(\C)\) is an unitary matrix if
\begin{equation*} A^*A = AA^* = I_n. \end{equation*}
Note that the condition \(A^*A=I_n\) is equivalent to saying that the columns of \(A\) form an orthonormal basis for \(\C^n\text{,}\) while the condition \(AA^*\) makes the analogous statement about the rows of \(A.\)

Definition 4.1.37.

A vector space is a non-empty set \(V\) and an associated field of scalars \(F\text{,}\) having operations of vector addition, denoted \(+\text{,}\) and scalar multiplication, denoted by juxtaposition, satisfying the following properties: For all vectors \(u, v, w \in V\text{,}\) and scalars \(\lambda, \mu \in F\)
  • closure under vector addition
  • \(\displaystyle u+v \in V\)
  • addition is commutative
  • \(\displaystyle u+v = v+u \)
  • addition is associative
  • \(\displaystyle (u+v)+w = u+(v+w) \)
  • additive identity
  • There is a vector \(\mathbf 0\in V\) so that \(\mathbf 0 + u = u.\)
  • additive inverses
  • For each \(u\in V\text{,}\) there is a vector denoted \(-u\in V\) so that \(u+ -u = \mathbf 0.\)
  • closure under scalar multiplication
  • \(\displaystyle \lambda u \in V. \)
  • scalar multiplication distributes across vector addition
  • \(\displaystyle \lambda(u+v) = \lambda u+\lambda v \)
  • distributes over scalar addition
  • \(\displaystyle (\lambda + \mu)v = \lambda v + \mu v \)
  • scalar associativity
  • \(\displaystyle (\lambda \mu)v = \lambda(\mu v) \)
  • \(V\)is unital
  • The field element \(1 \in F\) satisfies \(1 v = v.\)