## SectionILTInjective Linear Transformations

Some linear transformations possess one, or both, of two key properties, which go by the names injective and surjective. We will see that they are closely related to ideas like linear independence and spanning, and subspaces like the null space and the column space. In this section we will define an injective linear transformation and analyze the resulting consequences. The next section will do the same for the surjective property. In the final section of this chapter we will see what happens when we have the two properties simultaneously.

### SubsectionILTInjective Linear Transformations

As usual, we lead with a definition.

#### DefinitionILT.Injective Linear Transformation.

Suppose $$\ltdefn{T}{U}{V}$$ is a linear transformation. Then $$T$$ is injective if whenever $$\lteval{T}{\vect{x}}=\lteval{T}{\vect{y}}\text{,}$$ then $$\vect{x}=\vect{y}\text{.}$$

Given an arbitrary function, it is possible for two different inputs to yield the same output (think about the function $$f(x)=x^2$$ and the inputs $$x=3$$ and $$x=-3$$). For an injective function, this never happens. If we have equal outputs ($$\lteval{T}{\vect{x}}=\lteval{T}{\vect{y}}$$) then we must have achieved those equal outputs by employing equal inputs ($$\vect{x}=\vect{y}$$). Some authors prefer the term one-to-one where we use injective, and we will sometimes refer to an injective linear transformation as an injection.

### SubsectionEILTExamples of Injective Linear Transformations

It is perhaps most instructive to examine a linear transformation that is not injective first.

Archetype Q is the linear transformation

\begin{equation*} \ltdefn{T}{\complex{5}}{\complex{5}},\quad \lteval{T}{\colvector{x_1\\x_2\\x_3\\x_4\\x_5}}= \colvector{-2 x_1 + 3 x_2 + 3 x_3 - 6 x_4 + 3 x_5\\ -16 x_1 + 9 x_2 + 12 x_3 - 28 x_4 + 28 x_5\\ -19 x_1 + 7 x_2 + 14 x_3 - 32 x_4 + 37 x_5\\ -21 x_1 + 9 x_2 + 15 x_3 - 35 x_4 + 39 x_5\\ -9 x_1 + 5 x_2 + 7 x_3 - 16 x_4 + 16 x_5}\text{.} \end{equation*}

Notice that for

\begin{align*} \vect{x}&=\colvector{1\\3\\-1\\2\\4}& \vect{y}&=\colvector{4\\7\\0\\5\\7}\\ \end{align*}

we have

\begin{align*} \lteval{T}{\colvector{1\\3\\-1\\2\\4}}&=\colvector{4\\55\\72\\77\\31}& \lteval{T}{\colvector{4\\7\\0\\5\\7}}&=\colvector{4\\55\\72\\77\\31}\text{.} \end{align*}

So we have two vectors from the domain, $$\vect{x}\neq\vect{y}\text{,}$$ yet $$\lteval{T}{\vect{x}}=\lteval{T}{\vect{y}}\text{,}$$ in violation of Definition ILT. This is another example where you should not concern yourself with how $$\vect{x}$$ and $$\vect{y}$$ were selected, as this will be explained shortly. However, do understand why these two vectors provide enough evidence to conclude that $$T$$ is not injective.

Here is a cartoon of a non-injective linear transformation. Notice that the central feature of this cartoon is that $$\lteval{T}{\vect{u}}=\vect{v}=\lteval{T}{\vect{w}}\text{.}$$ Even though this happens again with some unnamed vectors, it only takes one occurrence to destroy the possibility of injectivity. Note also that the two vectors displayed in the bottom of $$V$$ have no bearing, either way, on the injectivity of $$T\text{.}$$

To show that a linear transformation is not injective, it is enough to find a single pair of inputs that get sent to the identical output, as in Example NIAQ. However, to show that a linear transformation is injective we must establish that this coincidence of outputs never occurs. Here is an example that shows how to establish this.

Archetype R is the linear transformation

\begin{equation*} \ltdefn{T}{\complex{5}}{\complex{5}},\quad \lteval{T}{\colvector{x_1\\x_2\\x_3\\x_4\\x_5}}= \colvector{-65 x_1 + 128 x_2 + 10 x_3 - 262 x_4 + 40 x_5\\ 36 x_1 - 73 x_2 - x_3 + 151 x_4 - 16 x_5\\ -44 x_1 + 88 x_2 + 5 x_3 - 180 x_4 + 24 x_5\\ 34 x_1 - 68 x_2 - 3 x_3 + 140 x_4 - 18 x_5\\ 12 x_1 - 24 x_2 - x_3 + 49 x_4 - 5 x_5}\text{.} \end{equation*}

To establish that $$R$$ is injective we must begin with the assumption that $$\lteval{T}{\vect{x}}=\lteval{T}{\vect{y}}$$ and somehow arrive at the conclusion that $$\vect{x}=\vect{y}\text{.}$$ Here we go,

\begin{align*} \colvector{0\\0\\0\\0\\0} &=\lteval{T}{\vect{x}}-\lteval{T}{\vect{y}}\\ &=\lteval{T}{\colvector{x_1\\x_2\\x_3\\x_4\\x_5}}-\lteval{T}{\colvector{y_1\\y_2\\y_3\\y_4\\y_5}}\\ &= \colvector{-65 x_1 + 128 x_2 + 10 x_3 - 262 x_4 + 40 x_5\\ 36 x_1 - 73 x_2 - x_3 + 151 x_4 - 16 x_5\\ -44 x_1 + 88 x_2 + 5 x_3 - 180 x_4 + 24 x_5\\ 34 x_1 - 68 x_2 - 3 x_3 + 140 x_4 - 18 x_5\\ 12 x_1 - 24 x_2 - x_3 + 49 x_4 - 5 x_5}\\ &\quad\quad- \colvector{-65 y_1 + 128 y_2 + 10 y_3 - 262 y_4 + 40 y_5\\ 36 y_1 - 73 y_2 - y_3 + 151 y_4 - 16 y_5\\ -44 y_1 + 88 y_2 + 5 y_3 - 180 y_4 + 24 y_5\\ 34 y_1 - 68 y_2 - 3 y_3 + 140 y_4 - 18 y_5\\ 12 y_1 - 24 y_2 - y_3 + 49 y_4 - 5 y_5}\\ &= \colvector{-65 (x_1-y_1) + 128 (x_2-y_2) + 10 (x_3-y_3) - 262 (x_4-y_4) + 40 (x_5-y_5)\\ 36 (x_1-y_1) - 73 (x_2-y_2) - (x_3-y_3) + 151 (x_4-y_4) - 16 (x_5-y_5)\\ -44 (x_1-y_1) + 88 (x_2-y_2) + 5 (x_3-y_3) - 180 (x_4-y_4) + 24 (x_5-y_5)\\ 34 (x_1-y_1) - 68 (x_2-y_2) - 3 (x_3-y_3) + 140 (x_4-y_4) - 18 (x_5-y_5)\\ 12 (x_1-y_1) - 24 (x_2-y_2) - (x_3-y_3) + 49 (x_4-y_4) - 5 (x_5-y_5)}\\ &= \begin{bmatrix} -65&128&10&-262&40\\ 36&-73&-1&151&-16\\ -44&88&5&-180&24\\ 34&-68&-3&140&-18\\ 12&-24&-1&49&-5 \end{bmatrix} \colvector{x_1-y_1\\x_2-y_2\\x_3-y_3\\x_4-y_4\\x_5-y_5}\text{.} \end{align*}

Now we recognize that we have a homogeneous system of 5 equations in 5 variables (the terms $$x_i-y_i$$ are the variables), so we row-reduce the coefficient matrix to

So the only solution is the trivial solution

\begin{align*} x_1-y_1&=0&x_2-y_2&=0&x_3-y_3&=0&x_4-y_4&=0&x_5-y_5&=0 \end{align*}

and we conclude that indeed $$\vect{x}=\vect{y}\text{.}$$ By Definition ILT, $$T$$ is injective.

Here is the cartoon for an injective linear transformation. It is meant to suggest that we never have two inputs associated with a single output. Again, the two lonely vectors at the bottom of $$V$$ have no bearing either way on the injectivity of $$T\text{.}$$

Let us now examine an injective linear transformation between abstract vector spaces.

Archetype V is defined by

\begin{equation*} \ltdefn{T}{P_3}{M_{22}},\quad\lteval{T}{a+bx+cx^2+dx^3}= \begin{bmatrix} a+b & a-2c\\ d & b-d \end{bmatrix}\text{.} \end{equation*}

To establish that the linear transformation is injective, begin by supposing that two polynomial inputs yield the same output matrix,

\begin{equation*} \lteval{T}{a_1+b_1x+c_1x^2+d_1x^3}=\lteval{T}{a_2+b_2x+c_2x^2+d_2x^3}\text{.} \end{equation*}

Then

\begin{align*} \zeromatrix &=\begin{bmatrix} 0&0\\0&0 \end{bmatrix}\\ &=\lteval{T}{a_1+b_1x+c_1x^2+d_1x^3}-\lteval{T}{a_2+b_2x+c_2x^2+d_2x^3}&& \text{Hypothesis}\\ &=\lteval{T}{(a_1+b_1x+c_1x^2+d_1x^3)-(a_2+b_2x+c_2x^2+d_2x^3)}&& \knowl{./knowl/definition-LT.html}{\text{Definition LT}}\\ &=\lteval{T}{(a_1-a_2)+(b_1-b_2)x+(c_1-c_2)x^2+(d_1-d_2)x^3}&& \text{Operations in }P_3\\ &= \begin{bmatrix} (a_1-a_2)+(b_1-b_2) & (a_1-a_2)-2(c_1-c_2)\\ (d_1-d_2) & (b_1-b_2)-(d_1-d_2) \end{bmatrix}&& \text{Definition of }T\text{.} \end{align*}

This single matrix equality translates to the homogeneous system of equations in the variables $$a_i-b_i\text{,}$$

\begin{align*} (a_1-a_2)+(b_1-b_2)&=0\\ (a_1-a_2)-2(c_1-c_2)&=0\\ (d_1-d_2)&=0\\ (b_1-b_2)-(d_1-d_2)&=0\text{.} \end{align*}

This system of equations can be rewritten as the matrix equation

\begin{equation*} \begin{bmatrix} 1&1&0&0\\1&0&-2&0\\0&0&0&1\\0&1&0&-1 \end{bmatrix} \colvector{(a_1-a_2)\b_1-b_2)\\(c_1-c_2)\\(d_1-d_2)}=\colvector{0\\0\\0\\0}\text{.} \end{equation*} Since the coefficient matrix is nonsingular (check this) the only solution is trivial, i.e. \begin{align*} a_1-a_2&=0&b_1-b_2&=0&c_1-c_2&=0&d_1-d_2&=0\\ \end{align*} so that \begin{align*} a_1&=a_2&b_1&=b_2&c_1&=c_2&d_1&=d_2 \end{align*} so the two inputs must be equal polynomials. By Definition ILT, \(T is injective.

### SubsectionKLTKernel of a Linear Transformation

For a linear transformation $$\ltdefn{T}{U}{V}\text{,}$$ the kernel is a subset of the domain $$U\text{.}$$ Informally, it is the set of all inputs that the transformation sends to the zero vector of the codomain. It will have some natural connections with the null space of a matrix, so we will keep the same notation, and if you think about your objects, then there should be little confusion. Here is the careful definition.

#### DefinitionKLT.Kernel of a Linear Transformation.

Suppose $$\ltdefn{T}{U}{V}$$ is a linear transformation. Then the kernel of $$T$$ is the set

\begin{equation*} \krn{T}=\setparts{\vect{u}\in U}{\lteval{T}{\vect{u}}=\zerovector}\text{.} \end{equation*}

Notice that the kernel of $$T$$ is just the preimage of $$\zerovector\text{,}$$ $$\preimage{T}{\zerovector}$$ (Definition PI). Here is an example.

Archetype O is the linear transformation

\begin{equation*} \ltdefn{T}{\complex{3}}{\complex{5}},\quad \lteval{T}{\colvector{x_1\\x_2\\x_3}}= \colvector{-x_1 + x_2 - 3 x_3\\ -x_1 + 2 x_2 - 4 x_3\\ x_1 + x_2 + x_3\\ 2 x_1 + 3 x_2 + x_3\\ x_1 + 2 x_3 }\text{.} \end{equation*}

To determine the elements of $$\complex{3}$$ in $$\krn{T}\text{,}$$ find those vectors $$\vect{u}$$ such that $$\lteval{T}{\vect{u}}=\zerovector\text{,}$$ that is,

\begin{align*} \lteval{T}{\vect{u}}&=\zerovector\\ \colvector{-u_1 + u_2 - 3 u_3\\ -u_1 + 2 u_2 - 4 u_3\\ u_1 + u_2 + u_3\\ 2 u_1 + 3 u_2 + u_3\\ u_1 + 2 u_3 } &= \colvector{0\\0\\0\\0\\0}\text{.} \end{align*}

Vector equality (Definition CVE) leads us to a homogeneous system of 5 equations in the variables $$u_i\text{,}$$

\begin{align*} -u_1 + u_2 - 3 u_3&=0\\ -u_1 + 2 u_2 - 4 u_3&=0\\ u_1 + u_2 + u_3&=0\\ 2 u_1 + 3 u_2 + u_3&=0\\ u_1 + 2 u_3&=0\text{.} \end{align*}

Row-reducing the coefficient matrix gives

\begin{equation*} \begin{bmatrix} \leading{1} & 0 & 2\\ 0 & \leading{1} & -1\\ 0 & 0 & 0\\ 0 & 0 & 0\\ 0 & 0 & 0 \end{bmatrix}\text{.} \end{equation*}

The kernel of $$T$$ is the set of solutions to this homogeneous system of equations, which by Theorem BNS can be expressed as

\begin{equation*} \krn{T}=\spn{\set{\colvector{-2\\1\\1}}}\text{.} \end{equation*}

We know that the span of a set of vectors is always a subspace (Theorem SSS), so the kernel computed in Example NKAO is also a subspace. This is no accident, the kernel of a linear transformation is always a subspace.

We can apply the three-part test of Theorem TSS. First $$\lteval{T}{\zerovector_U}=\zerovector_V$$ by Theorem LTTZZ, so $$\zerovector_U\in\krn{T}$$ and we know that the kernel is nonempty.

Suppose we assume that $$\vect{x},\,\vect{y}\in\krn{T}\text{.}$$ Is $$\vect{x}+\vect{y}\in\krn{T}\text{?}$$ We have

\begin{align*} \lteval{T}{\vect{x}+\vect{y}}&=\lteval{T}{\vect{x}}+\lteval{T}{\vect{y}}&& \knowl{./knowl/definition-LT.html}{\text{Definition LT}}\\ &=\zerovector+\zerovector&&\vect{x},\,\vect{y}\in\krn{T}\\ &=\zerovector&& \knowl{./knowl/property-Z.html}{\text{Property Z}}\text{.} \end{align*}

This qualifies $$\vect{x}+\vect{y}$$ for membership in $$\krn{T}\text{.}$$ So we have additive closure.

Suppose we assume that $$\alpha\in\complexes$$ and $$\vect{x}\in\krn{T}\text{.}$$ Is $$\alpha\vect{x}\in\krn{T}\text{?}$$ We have

\begin{align*} \lteval{T}{\alpha\vect{x}}&=\alpha\lteval{T}{\vect{x}}&& \knowl{./knowl/definition-LT.html}{\text{Definition LT}}\\ &=\alpha\zerovector&&\vect{x}\in\krn{T}\\ &=\zerovector&& \knowl{./knowl/theorem-ZVSM.html}{\text{Theorem ZVSM}}\text{.} \end{align*}

This qualifies $$\alpha\vect{x}$$ for membership in $$\krn{T}\text{.}$$ So we have scalar closure and Theorem TSS tells us that $$\krn{T}$$ is a subspace of $$U\text{.}$$

Let us compute another kernel, now that we know in advance that it will be a subspace.

Archetype P is the linear transformation

\begin{equation*} \ltdefn{T}{\complex{3}}{\complex{5}},\quad \lteval{T}{\colvector{x_1\\x_2\\x_3}}= \colvector{-x_1 + x_2 + x_3\\ -x_1 + 2 x_2 + 2 x_3\\ x_1 + x_2 + 3 x_3\\ 2 x_1 + 3 x_2 + x_3\\ -2 x_1 + x_2 + 3 x_3}\text{.} \end{equation*}

To determine the elements of $$\complex{3}$$ in $$\krn{T}\text{,}$$ find those vectors $$\vect{u}$$ such that $$\lteval{T}{\vect{u}}=\zerovector\text{,}$$ that is,

\begin{align*} \lteval{T}{\vect{u}}&=\zerovector\\ \colvector{ -u_1 + u_2 + u_3\\ -u_1 + 2 u_2 + 2 u_3\\ u_1 + u_2 + 3 u_3\\ 2 u_1 + 3 u_2 + u_3\\ -2 u_1 + u_2 + 3 u_3 } &= \colvector{0\\0\\0\\0\\0}\text{.} \end{align*}

Vector equality (Definition CVE) leads us to a homogeneous system of 5 equations in the variables $$u_i\text{,}$$

\begin{align*} -u_1 + u_2 + u_3&=0\\ -u_1 + 2 u_2 + 2 u_3&=0\\ u_1 + u_2 + 3 u_3&=0\\ 2 u_1 + 3 u_2 + u_3&=0\\ -2 u_1 + u_2 + 3 u_3&=0\text{.} \end{align*}

Row-reducing the coefficient matrix gives

\begin{equation*} \begin{bmatrix} \leading{1} & 0 & 0\\ 0 & \leading{1} & 0\\ 0 & 0 & \leading{1}\\ 0 & 0 & 0\\ 0 & 0 & 0 \end{bmatrix}\text{.} \end{equation*}

The kernel of $$T$$ is the set of solutions to this homogeneous system of equations, which is simply the trivial solution $$\vect{u}=\zerovector\text{,}$$ so

\begin{equation*} \krn{T}=\set{\zerovector}=\spn{\set{\ }}\text{.} \end{equation*}

Our next theorem says that if a preimage is a nonempty set then we can construct it by picking any one element and adding on elements of the kernel.

Let $$M=\setparts{\vect{u}+\vect{z}}{\vect{z}\in\krn{T}}\text{.}$$ First, we show that $$M\subseteq\preimage{T}{\vect{v}}\text{.}$$ Suppose that $$\vect{w}\in M\text{,}$$ so $$\vect{w}$$ has the form $$\vect{w}=\vect{u}+\vect{z}\text{,}$$ where $$\vect{z}\in\krn{T}\text{.}$$ Then

\begin{align*} \lteval{T}{\vect{w}}&=\lteval{T}{\vect{u}+\vect{z}}\\ &=\lteval{T}{\vect{u}}+\lteval{T}{\vect{z}}&& \knowl{./knowl/definition-LT.html}{\text{Definition LT}}\\ &=\vect{v}+\zerovector&&\vect{u}\in\preimage{T}{\vect{v}},\ \vect{z}\in\krn{T}\\ &=\vect{v}&& \knowl{./knowl/property-Z.html}{\text{Property Z}} \end{align*}

which qualifies $$\vect{w}$$ for membership in the preimage of $$\vect{v}\text{,}$$ $$\vect{w}\in\preimage{T}{\vect{v}}\text{.}$$

For the opposite inclusion, suppose $$\vect{x}\in\preimage{T}{\vect{v}}\text{.}$$ Then,

\begin{align*} \lteval{T}{\vect{x}-\vect{u}}&=\lteval{T}{\vect{x}}-\lteval{T}{\vect{u}}&& \knowl{./knowl/definition-LT.html}{\text{Definition LT}}\\ &=\vect{v}-\vect{v}&&\vect{x},\,\vect{u}\in\preimage{T}{\vect{v}}\\ &=\zerovector\text{.} \end{align*}

This qualifies $$\vect{x}-\vect{u}$$ for membership in the kernel of $$T\text{,}$$ $$\krn{T}\text{.}$$ So there is a vector $$\vect{z}\in\krn{T}$$ such that $$\vect{x}-\vect{u}=\vect{z}\text{.}$$ Rearranging this equation gives $$\vect{x}=\vect{u}+\vect{z}$$ and so $$\vect{x}\in M\text{.}$$ So $$\preimage{T}{\vect{v}}\subseteq M$$ and we see that $$M=\preimage{T}{\vect{v}}\text{,}$$ as desired.

This theorem, and its proof, should remind you very much of Theorem PSPHS. Additionally, you might go back and review Example SPIAS. Can you tell now which is the only preimage to be a subspace?

Here is the cartoon which describes the “many-to-one” behavior of a typical linear transformation. Presume that $$\lteval{T}{\vect{u}_i}=\vect{v}_i\text{,}$$ for $$i=1,2,3\text{,}$$ and as guaranteed by Theorem LTTZZ, $$\lteval{T}{\zerovector_U}=\zerovector_V\text{.}$$ Then four pre-images are depicted, each labeled slightly different. $$\preimage{T}{\vect{v}_2}$$ is the most general, employing Theorem KPI to provide two equal descriptions of the set. The most unusual is $$\preimage{T}{\zerovector_V}$$ which is equal to the kernel, $$\krn{T}\text{,}$$ and hence is a subspace (by Theorem KLTS). The subdivisions of the domain, $$U\text{,}$$ are meant to suggest the partioning of the domain by the collection of pre-images. It also suggests that each pre-image is of similar size or structure, since each is a “shifted” copy of the kernel. Notice that we cannot speak of the dimension of a pre-image, since it is almost never a subspace. Also notice that $$\vect{x},\,\vect{y}\in V$$ are elements of the codomain with empty pre-images.

The next theorem is one we will cite frequently, as it characterizes injections by the size of the kernel.

##### (⇒)

We assume $$T$$ is injective and we need to establish that two sets are equal (Definition SE). Since the kernel is a subspace (Theorem KLTS), $$\set{\zerovector}\subseteq\krn{T}\text{.}$$ To establish the opposite inclusion, suppose $$\vect{x}\in\krn{T}\text{.}$$ We have

\begin{align*} \lteval{T}{\vect{x}} &=\zerovector&& \knowl{./knowl/definition-KLT.html}{\text{Definition KLT}}\\ &=\lteval{T}{\zerovector}&& \knowl{./knowl/theorem-LTTZZ.html}{\text{Theorem LTTZZ}}\text{.} \end{align*}

We can apply Definition ILT to conclude that $$\vect{x}=\zerovector\text{.}$$ Therefore $$\krn{T}\subseteq\set{\zerovector}$$ and by Definition SE, $$\krn{T}=\set{\zerovector}\text{.}$$

##### (⇐)

To establish that $$T$$ is injective, appeal to Definition ILT and begin with the assumption that $$\lteval{T}{\vect{x}}=\lteval{T}{\vect{y}}\text{.}$$ Then

\begin{align*} \lteval{T}{\vect{x}-\vect{y}} &=\lteval{T}{\vect{x}}-\lteval{T}{\vect{y}}&& \knowl{./knowl/definition-LT.html}{\text{Definition LT}}\\ &=\zerovector&& \text{Hypothesis}\text{.} \end{align*}

So $$\vect{x}-\vect{y}\in\krn{T}$$ by Definition KLT and with the hypothesis that the kernel is trivial we conclude that $$\vect{x}-\vect{y}=\zerovector\text{.}$$ Then

\begin{gather*} \vect{y} =\vect{y}+\zerovector =\vect{y}+\left(\vect{x}-\vect{y}\right) =\vect{x} \end{gather*}

thus establishing that $$T$$ is injective by Definition ILT.

You might begin to think about how Figure KPI would change if the linear transformation is injective, which would make the kernel trivial by Theorem KILT.

We are now in a position to revisit our first example in this section, Example NIAQ. In that example, we showed that Archetype Q is not injective by constructing two vectors, which when used to evaluate the linear transformation provided the same output, thus violating Definition ILT. Just where did those two vectors come from?

The key is the vector

\begin{equation*} \vect{z}=\colvector{3\\4\\1\\3\\3} \end{equation*}

which you can check is an element of $$\krn{T}$$ for Archetype Q. Choose a vector $$\vect{x}$$ at random, and then compute $$\vect{y}=\vect{x}+\vect{z}$$ (verify this computation back in Example NIAQ). Then

\begin{align*} \lteval{T}{\vect{y}}&=\lteval{T}{\vect{x}+\vect{z}}\\ &=\lteval{T}{\vect{x}}+\lteval{T}{\vect{z}}&& \knowl{./knowl/definition-LT.html}{\text{Definition LT}}\\ &=\lteval{T}{\vect{x}}+\zerovector&&\vect{z}\in\krn{T}\\ &=\lteval{T}{\vect{x}}&& \knowl{./knowl/property-Z.html}{\text{Property Z}}\text{.} \end{align*}

Whenever the kernel of a linear transformation is nontrivial, we can employ this device and conclude that the linear transformation is not injective. This is another way of viewing Theorem KILT. For an injective linear transformation, the kernel is trivial and our only choice for $$\vect{z}$$ is the zero vector, which will not help us create two different inputs for $$T$$ that yield identical outputs. For every one of the archetypes that is not injective, there is an example presented of exactly this form.

In Example NKAO the kernel of Archetype O was determined to be

\begin{equation*} \spn{\set{\colvector{-2\\1\\1}}} \end{equation*}

a subspace of $$\complex{3}$$ with dimension 1. Since the kernel is not trivial, Theorem KILT tells us that $$T$$ is not injective.

In Example TKAP it was shown that the linear transformation in Archetype P has a trivial kernel. So by Theorem KILT, $$T$$ is injective.

#### SageILT.Injective Linear Transformations.

By now, you have probably already figured out how to determine if a linear transformation is injective, and what its kernel is. You may also now begin to understand why Sage calls the null space of a matrix a kernel. Here are two examples, first a reprise of Example NKAO.

U = QQ^3
V = QQ^5
x1, x2, x3 = var('x1, x2, x3')
outputs = [ -x1 +   x2 - 3*x3,
-x1 + 2*x2 - 4*x3,
x1 +   x2 +   x3,
2*x1 + 3*x2 +   x3,
x1        + 2*x3]
T_symbolic(x1, x2, x3) = outputs
T = linear_transformation(U, V, T_symbolic)
T.is_injective()

T.kernel()


So we have a concrete demonstration of one half of Theorem KILT. Here is the second example, a do-over for Example TKAP, but renamed as S.

U = QQ^3
V = QQ^5
x1, x2, x3 = var('x1, x2, x3')
outputs = [  -x1 +   x2 +   x3,
-x1 + 2*x2 + 2*x3,
x1 +   x2 + 3*x3,
2*x1 + 3*x2 +   x3,
-2*x1 +   x2 + 3*x3]
S_symbolic(x1, x2, x3) = outputs
S = linear_transformation(U, V, S_symbolic)
S.is_injective()

S.kernel()

S.kernel() == U.subspace([])


And so we have a concrete demonstration of the other half of Theorem KILT.

Now that we have Theorem KPI, we can return to our discussion from Sage PI. The .preimage_representative()method of a linear transformation will give us a single element of the pre-image, with no other guarantee about the nature of that element. That is fine, since this is all Theorem KPI requires (in addition to the kernel). Remember that not every element of the codomain may have a nonempty pre-image (as indicated in the hypotheses of Theorem KPI). Here is an example using Tfrom above, with a choice of a codomain element that has a nonempty pre-image.

TK = T.kernel()
v = vector(QQ, [2, 3, 0, 1, -1])
u = T.preimage_representative(v)
u


Now the following will create random elements of the preimage of v, which can be verified by the test always returning True. Use the compute cell just below if you are curious what plooks like.

p = u + TK.random_element()
T(p) == v

p                 # random


As suggested, some choices of vcan lead to empty pre-images, in which case Theorem KPI does not even apply.

v = vector(QQ, [4, 6, 3, 1, -2])
u = T.preimage_representative(v)


The situation is less interesting for an injective linear transformation. Still, pre-images may be empty, but when they are nonempty, they are just singletons (a single element) since the kernel is empty. So a repeat of the above example, with Srather than T, would not be very informative.

### SubsectionILTLIInjective Linear Transformations and Linear Independence

There is a connection between injective linear transformations and linearly independent sets that we will make precise in the next two theorems. However, more informally, we can get a feel for this connection when we think about how each property is defined. A set of vectors is linearly independent if the only relation of linear dependence is the trivial one. A linear transformation is injective if the only way two input vectors can produce the same output is in the trivial way, when both input vectors are equal.

Begin with a relation of linear dependence on $$R$$ (Definition RLD, Definition LI),

\begin{align*} a_1\lteval{T}{\vect{u}_1}+a_2\lteval{T}{\vect{u}_2}+a_3\lteval{T}{\vect{u}_3}+\ldots+a_t\lteval{T}{\vect{u}_t}&=\zerovector\\ \lteval{T}{\lincombo{a}{u}{t}}&=\zerovector&& \knowl{./knowl/theorem-LTLC.html}{\text{Theorem LTLC}}\\ \lincombo{a}{u}{t}&\in\krn{T}&& \knowl{./knowl/definition-KLT.html}{\text{Definition KLT}}\\ \lincombo{a}{u}{t}&\in\set{\zerovector}&& \knowl{./knowl/theorem-KILT.html}{\text{Theorem KILT}}\\ \lincombo{a}{u}{t}&=\zerovector&& \knowl{./knowl/definition-SET.html}{\text{Definition SET}}\text{.} \end{align*}

Since this is a relation of linear dependence on the linearly independent set $$S\text{,}$$ we can conclude that

\begin{align*} a_1&=0&a_2&=0&a_3&=0&\ldots&&a_t&=0 \end{align*}

and this establishes that $$R$$ is a linearly independent set.

##### (⇒)

Assume $$T$$ is injective. Since $$B$$ is a basis, we know $$B$$ is linearly independent (Definition B). Then Theorem ILTLI says that $$C$$ is a linearly independent subset of $$V\text{.}$$

##### (⇐)

Assume that $$C$$ is linearly independent. To establish that $$T$$ is injective, we will show that the kernel of $$T$$ is trivial (Theorem KILT). Suppose that $$\vect{u}\in\krn{T}\text{.}$$ As an element of $$U\text{,}$$ we can write $$\vect{u}$$ as a linear combination of the basis vectors in $$B$$ (uniquely). So there are are scalars, $$\scalarlist{a}{m}\text{,}$$ such that

\begin{equation*} \vect{u}=\lincombo{a}{u}{m}\text{.} \end{equation*}

Then,

\begin{align*} \zerovector &=\lteval{T}{\vect{u}}&& \knowl{./knowl/definition-KLT.html}{\text{Definition KLT}}\\ &=\lteval{T}{\lincombo{a}{u}{m}}&& \knowl{./knowl/definition-SSVS.html}{\text{Definition SSVS}}\\ &=a_1\lteval{T}{\vect{u}_1}+a_2\lteval{T}{\vect{u}_2}+a_3\lteval{T}{\vect{u}_3}+\cdots+a_m\lteval{T}{\vect{u}_m}&& \knowl{./knowl/theorem-LTLC.html}{\text{Theorem LTLC}}\text{.} \end{align*}

This is a relation of linear dependence (Definition RLD) on the linearly independent set $$C\text{,}$$ so the scalars are all zero: $$a_1=a_2=a_3=\cdots=a_m=0\text{.}$$ Then

\begin{align*} \vect{u}&=\lincombo{a}{u}{m}\\ &=0\vect{u}_1+0\vect{u}_2+0\vect{u}_3+\cdots+0\vect{u}_m&& \knowl{./knowl/theorem-ZSSM.html}{\text{Theorem ZSSM}}\\ &=\zerovector+\zerovector+\zerovector+\cdots+\zerovector&& \knowl{./knowl/theorem-ZSSM.html}{\text{Theorem ZSSM}}\\ &=\zerovector&& \knowl{./knowl/property-Z.html}{\text{Property Z}}\text{.} \end{align*}

Since $$\vect{u}$$ was chosen as an arbitrary vector from $$\krn{T}\text{,}$$ we have $$\krn{T}=\set{\zerovector}$$ and Theorem KILT tells us that $$T$$ is injective.

### SubsectionILTDInjective Linear Transformations and Dimension

Suppose to the contrary that $$m=\dimension{U}\gt\dimension{V}=t\text{.}$$ Let $$B$$ be a basis of $$U\text{,}$$ which will then contain $$m$$ vectors. Apply $$T$$ to each element of $$B$$ to form a set $$C$$ that is a subset of $$V\text{.}$$ By Theorem ILTB, $$C$$ is linearly independent and therefore must contain $$m$$ distinct vectors. So we have found a set of $$m$$ linearly independent vectors in $$V\text{,}$$ a vector space of dimension $$t\text{,}$$ with $$m\gt t\text{.}$$ However, this contradicts Theorem G, so our assumption is false and $$\dimension{U}\leq\dimension{V}\text{.}$$

The linear transformation in Archetype U is

\begin{equation*} \ltdefn{T}{M_{23}}{\complex{4}},\quad \lteval{T}{\begin{bmatrix}a&b&c\\d&e&f\end{bmatrix}}= \colvector{a+2b+12c-3d+e+6f\\2a-b-c+d-11f\\a+b+7c+2d+e-3f\\a+2b+12c+5e-5f}\text{.} \end{equation*}

Since $$\dimension{M_{23}}=6\gt 4=\dimension{\complex{4}}\text{,}$$ $$T$$ cannot be injective for then $$T$$ would violate Theorem ILTD.

Notice that the previous example made no use of the actual formula defining the function. Merely a comparison of the dimensions of the domain and codomain is enough to conclude that the linear transformation is not injective. Archetype M and Archetype N are two more examples of linear transformations that have “big” domains and “small” codomains, resulting in “collisions” of outputs and thus are non-injective linear transformations.

### SubsectionCILTComposition of Injective Linear Transformations

In Subsection LT.NLTFO we saw how to combine linear transformations to build new linear transformations, specifically, how to build the composition of two linear transformations (Definition LTC). It will be useful later to know that the composition of injective linear transformations is again injective, so we prove that here.

That the composition is a linear transformation was established in Theorem CLTLT, so we need only establish that the composition is injective. Applying Definition ILT, choose $$\vect{x}\text{,}$$ $$\vect{y}$$ from $$U\text{.}$$ Then if $$\lteval{\left(\compose{S}{T}\right)}{\vect{x}}=\lteval{\left(\compose{S}{T}\right)}{\vect{y}}\text{,}$$

\begin{align*} &\Rightarrow&\lteval{S}{\lteval{T}{\vect{x}}}&=\lteval{S}{\lteval{T}{\vect{y}}}&& \knowl{./knowl/definition-LTC.html}{\text{Definition LTC}}\\ &\Rightarrow&\lteval{T}{\vect{x}}&=\lteval{T}{\vect{y}}&& \knowl{./knowl/definition-ILT.html}{\text{Definition ILT}}\text{ for }S\\ &\Rightarrow&\vect{x}&=\vect{y}&& \knowl{./knowl/definition-ILT.html}{\text{Definition ILT}}\text{ for }T\text{.} \end{align*}

#### SageCILT.Composition of Injective Linear Transformations.

One way to use Sage is to construct examples of theorems and verify the conclusions. Sometimes you will get this wrong: you might build an example that does not satisfy the hypotheses, or your example may not satisfy the conclusions. This may be because you are not using Sage properly, or because you do not understand a definition or a theorem, or in very limited cases you may have uncovered a bug in Sage (which is always the preferred explanation!). But in the process of trying to understand a discrepancy or unexpected result, you will learn much more, both about linear algebra and about Sage. And Sage is incredibly patient — it will stay up with you all night to help you through a rough patch.

Let us illustrate the above in the context of Theorem CILTI. The hypotheses indicate we need two injective linear transformations. Where will get two such linear transformations? Well, the contrapositive of Theorem ILTD tells us that if the dimension of the domain exceeds the dimension of the codomain, we will never be injective. So we should at a minimum avoid this scenario. We can build two linear transformations from matrices created randomly, and just hope that they lead to injective linear transformations. Here is an example of how we create examples like this. The random matrix has single-digit entries, and almost always will lead to an injective linear transformation, though we cannot be absolutely certain. Evaluate this cell repeatedly, to see how rarely the result is not injective.

E = random_matrix(ZZ, 3, 2, x=-9, y=9)
T = linear_transformation(QQ^2, QQ^3, E, side='right')
T.is_injective()                              # random


Our concrete example below was created this way, so here we go.

U = QQ^2
V = QQ^3
W = QQ^4
A = matrix(QQ, 3, 2, [[-4, -1],
[-3,  3],
[ 7, -6]])
B = matrix(QQ, 4, 3, [[ 7,  0, -1],
[ 6,  2,  7],
[ 3, -1,  2],
[-6, -1,  1]])
T = linear_transformation(U, V, A, side='right')
T.is_injective()

S = linear_transformation(V, W, B, side='right')
S.is_injective()

C = S*T
C.is_injective()


#### 1.

Suppose $$\ltdefn{T}{\complex{8}}{\complex{5}}$$ is a linear transformation. Why is $$T$$ not injective?

#### 2.

Describe the kernel of an injective linear transformation.

### ExercisesILTExercises

#### C10.

Each archetype below is a linear transformation. Compute the kernel for each.

Archetype M, Archetype N, Archetype O, Archetype P, Archetype Q, Archetype R, Archetype S, Archetype T, Archetype U, Archetype V, Archetype W, Archetype X

#### C20.

The linear transformation $$\ltdefn{T}{\complex{4}}{\complex{3}}$$ is not injective. Find two inputs $$\vect{x},\,\vect{y}\in\complex{4}$$ that yield the same output (that is $$\lteval{T}{\vect{x}}=\lteval{T}{\vect{y}}$$).

\begin{equation*} \lteval{T}{\colvector{x_1\\x_2\\x_3\\x_4}}= \colvector{ 2x_1+x_2+x_3\\ -x_1+3x_2+x_3-x_4\\ 3x_1+x_2+2x_3-2x_4 }\text{.} \end{equation*}
Solution.

A linear transformation that is not injective will have a nontrivial kernel (Theorem KILT), and this is the key to finding the desired inputs. We need one nontrivial element of the kernel, so suppose that $$\vect{z}\in\complex{4}$$ is an element of the kernel,

\begin{equation*} \colvector{0\\0\\0} =\zerovector =\lteval{T}{\vect{z}} =\colvector{ 2z_1+z_2+z_3\\ -z_1+3z_2+z_3-z_4\\ 3z_1+z_2+2z_3-2z_4 }\text{.} \end{equation*}

Vector equality Definition CVE leads to the homogeneous system of three equations in four variables,

\begin{align*} 2z_1+z_2+z_3&=0\\ -z_1+3z_2+z_3-z_4&=0\\ 3z_1+z_2+2z_3-2z_4&=0\text{.} \end{align*}

The coefficient matrix of this system row-reduces as

\begin{equation*} \begin{bmatrix} 2 & 1 & 1 & 0 \\ -1 & 3 & 1 & -1 \\ 3 & 1 & 2 & -2 \end{bmatrix} \rref \begin{bmatrix} \leading{1} & 0 & 0 & 1 \\ 0 & \leading{1} & 0 & 1 \\ 0 & 0 & \leading{1} & -3 \end{bmatrix}\text{.} \end{equation*}

From this we can find a solution (we only need one), that is an element of $$\krn{T}\text{,}$$

\begin{equation*} \vect{z}=\colvector{-1\\-1\\3\\1}\text{.} \end{equation*}

Now, we choose a vector $$\vect{x}$$ at random and set $$\vect{y}=\vect{x}+\vect{z}\text{,}$$

\begin{align*} \vect{x} &=\colvector{2\\3\\4\\-2} & \vect{y}&=\vect{x}+\vect{z}= \colvector{2\\3\\4\\-2}+\colvector{-1\\-1\\3\\1} =\colvector{1\\2\\7\\-1} \end{align*}

and you can check that

\begin{equation*} \lteval{T}{\vect{x}} =\colvector{11\\13\\21} =\lteval{T}{\vect{y}}\text{.} \end{equation*}

A quicker solution is to take two elements of the kernel (in this case, scalar multiples of $$\vect{z}$$) which both get sent to $$\zerovector$$ by $$T\text{.}$$ Quicker yet, take $$\zerovector$$ and $$\vect{z}$$ as $$\vect{x}$$ and $$\vect{y}\text{,}$$ which also both get sent to $$\zerovector$$ by $$T\text{.}$$

#### C25.

Define the linear transformation

\begin{equation*} \ltdefn{T}{\complex{3}}{\complex{2}},\quad \lteval{T}{\colvector{x_1\\x_2\\x_3}}=\colvector{2x_1-x_2+5x_3\\-4x_1+2x_2-10x_3}\text{.} \end{equation*}

Find a basis for the kernel of $$T\text{,}$$ $$\krn{T}\text{.}$$ Is $$T$$ injective?

Solution.

To find the kernel, we require all $$\vect{x}\in\complex{3}$$ such that $$\lteval{T}{\vect{x}}=\zerovector\text{.}$$ This condition is

\begin{equation*} \colvector{2x_1-x_2+5x_3\\-4x_1+2x_2-10x_3}=\colvector{0\\0}\text{.} \end{equation*}

This leads to a homogeneous system of two linear equations in three variables, whose coefficient matrix row-reduces to

\begin{equation*} \begin{bmatrix} \leading{1} & -\frac{1}{2} & \frac{5}{2}\\ 0 & 0 & 0 \end{bmatrix}\text{.} \end{equation*}

With two free variables Theorem BNS yields the basis for the null space

\begin{equation*} \set{ \colvector{-\frac{5}{2}\\0\\1},\, \colvector{\frac{1}{2}\\1\\0} }\text{.} \end{equation*}

With $$\nullity{T}\neq 0\text{,}$$ $$\krn{T}\neq\set{\zerovector}\text{,}$$ so Theorem KILT says $$T$$ is not injective.

#### C26.

Let

\begin{equation*} A = \begin{bmatrix} 1 & 2 & 3 & 1 & 0\\ 2 & -1 & 1 & 0 & 1\\ 1 & 2 & -1 & -2 & 1\\ 1 & 3 & 2 & 1 & 2 \end{bmatrix} \end{equation*}

and let $$\ltdefn{T}{\complex{5}}{\complex{4}}$$ be given by $$\lteval{T}{\vect{x}}=A\vect{x}\text{.}$$ Is $$T$$ injective? (Hint: No calculation is required.)

Solution.

By Theorem ILTD, if a linear transformation $$\ltdefn{T}{U}{V}$$ is injective, then $$\dim(U)\le\dim(V)\text{.}$$ In this case, $$\ltdefn{T}{\complex{5}}{\complex{4}}\text{,}$$ and $$5=\dimension{\complex{5}}\gt\dimension{\complex{4}}=4\text{.}$$ Thus, $$T$$ cannot possibly be injective.

#### C27.

Let $$\ltdefn{T}{\complex{3}}{\complex{3}}$$ be given by $$\lteval{T}{\colvector{x\\y\\z}} = \colvector{2x + y + z\\ x - y + 2z\\ x + 2y - z}\text{.}$$ Find $$\krn{T}\text{.}$$ Is $$T$$ injective?

Solution.

If $$\lteval{T}{\colvector{x\\y\\z}} = \zerovector\text{,}$$ then $$\colvector{2x + y + z\\x - y + 2z\\x + 2y - z} = \zerovector\text{.}$$ Thus, we have the system

\begin{align*} 2x + y + z &= 0\\ x - y + 2z &= 0\\ x + 2y - z &= 0\text{.} \end{align*}

Thus, we are looking for the null space of the matrix

\begin{equation*} A_T = \begin{bmatrix} 2& 1 & 1\\ 1 & -1 & 2\\ 1 & 2 & -1 \end{bmatrix}\text{.} \end{equation*}

Since $$A_T$$ row-reduces to

\begin{equation*} \begin{bmatrix} \leading{1} & 0 & 1\\ 0 & \leading{1} & -1 \\ 0 & 0 & 0 \end{bmatrix}\text{,} \end{equation*}

the kernel of $$T$$ is all vectors where $$x = -z$$ and $$y = z\text{.}$$ Thus, $$\krn{T} = \spn{\set{\colvector{ -1\\1\\1}}}\text{.}$$

Since the kernel is not trivial, Theorem KILT tells us that $$T$$ is not injective.

#### C28.

Let

\begin{equation*} A = \begin{bmatrix} 1 & 2 & 3 & 1 \\ 2 & -1 & 1 & 0 \\ 1 & 2 & -1 & -2 \\ 1 & 3 & 2 & 1 \end{bmatrix} \end{equation*}

and let $$\ltdefn{T}{\complex{4}}{\complex{4}}$$ be given by $$\lteval{T}{\vect{x}}=A\vect{x}\text{.}$$ Find $$\krn{T}\text{.}$$ Is $$T$$ injective?

Solution.

Since $$T$$ is given by matrix multiplication, $$\krn{T} = \nsp{A}\text{.}$$ We have

\begin{align*} \begin{bmatrix} 1 & 2 & 3 & 1\\ 2 & -1 & 1 & 0\\ 1 & 2 & -1 & -2 \\ 1 & 3 & 2 & 1 \end{bmatrix} &\rref \begin{bmatrix} \leading{1} & 0 & 0 & 0\\ 0 & \leading{1} & 0 & 0\\ 0 & 0 & \leading{1} & 0\\ 0 & 0 & 0 & \leading{1} \end{bmatrix}\text{.} \end{align*}

The null space of $$A$$ is $$\set{\zerovector}\text{,}$$ so the kernel of $$T$$ is also trivial: $$\krn{T} = \set{\zerovector}\text{.}$$

#### C29.

Let

\begin{equation*} A = \begin{bmatrix} 1 & 2 & 1 & 1 \\ 2 & 1 & 1 & 0 \\ 1 & 2 & 1 & 2 \\ 1 & 2 & 1 & 1 \end{bmatrix} \end{equation*}

and let $$\ltdefn{T}{\complex{4}}{\complex{4}}$$ be given by $$\lteval{T}{\vect{x}}=A\vect{x}\text{.}$$ Find $$\krn{T}\text{.}$$ Is $$T$$ injective?

Solution.

Since $$T$$ is given by matrix multiplication, $$\krn{T} = \nsp{A}\text{.}$$ We have

\begin{align*} \begin{bmatrix} 1 & 2 & 1 & 1\\ 2 & 1 & 1 & 0\\ 1 & 2 & 1 & 2 \\ 1 & 2 & 1 & 1 \end{bmatrix} &\rref \begin{bmatrix} \leading{1} & 0 & 1/3 & 0\\ 0 & \leading{1} & 1/3 & 0\\ 0 & 0 & 0 & \leading{1}\\ 0 & 0 & 0 & 0 \end{bmatrix}\text{.} \end{align*}

Thus, a basis for the null space of $$A$$ is $$\set{\colvector{-1\\-1\\3\\0}}\text{,}$$ and the kernel is $$\krn{T} = \spn{\set{\colvector{-1\\-1\\3\\0}}}\text{.}$$ Since the kernel is nontrivial, this linear transformation is not injective.

#### C30.

Let $$T : M_{22} \rightarrow P_2$$ be given by $$T\left(\begin{bmatrix} a & b \\ c & d \end{bmatrix}\right) = (a + b) + (a + c)x + (a + d)x^2\text{.}$$ Is $$T$$ injective? Find $$\krn{T}\text{.}$$

Solution.

We can see without computing that $$T$$ is not injective, since the dimension of $$M_{22}$$ is larger than the dimension of $$P_2\text{.}$$ However, that does not address the question of the kernel of $$T\text{.}$$ We need to find all matrices $$\begin{bmatrix} a & b \\ c & d \end{bmatrix}$$ so that $$(a + b) + (a + c)x + (a + d)x^2 = 0\text{.}$$ This means $$a + b = 0\text{,}$$ $$a + c = 0\text{,}$$ and $$a + d = 0\text{,}$$ or equivalently, $$b = d = c = -a\text{.}$$ Thus, the kernel is a one-dimensional subspace of $$M_{22}$$ spanned by $$\begin{bmatrix} 1 & -1\\-1&-1 \end{bmatrix}\text{.}$$ Symbolically, we have $$\krn{T} = \spn{\set{\begin{bmatrix} 1 & -1\\-1&-1 \end{bmatrix}}}\text{.}$$

#### C31.

Given that the linear transformation $$\ltdefn{T}{\complex{3}}{\complex{3}}\text{,}$$ $$\lteval{T}{\colvector{x\\y\\z}} = \colvector{2x + y\\2y + z\\x + 2z}$$ is injective, show directly that

\begin{equation*} \set{ \lteval{T}{\vect{e}_1},\, \lteval{T}{\vect{e}_2},\, \lteval{T}{\vect{e}_3} } \end{equation*}

is a linearly independent set.

Solution.

We have

\begin{align*} \lteval{T}{\vect{e}_1} &= \colvector{2\\0\\1} & \lteval{T}{\vect{e}_2} &= \colvector{1\\2\\0} & \lteval{T}{\vect{e}_3} &= \colvector{0\\1\\2}\text{.} \end{align*}

Let us put these vectors into a matrix and row reduce to test their linear independence.

\begin{align*} \begin{bmatrix} 2 & 1 & 0\\ 0 & 2 & 1\\ 1 & 0 & 2 \end{bmatrix} &\rref \begin{bmatrix} \leading{1} & 0 & 0\\ 0 & \leading{1} & 0\\ 0 & 0 & \leading{1} \end{bmatrix} \end{align*}

so the set of vectors $$\set{\lteval{T}{\vect{e}_1},\, \lteval{T}{\vect{e}_2},\,\lteval{T}{\vect{e}_3}}$$ is linearly independent.

#### C32.

Given that the linear transformation $$\ltdefn{T}{\complex{2}}{\complex{3}}\text{,}$$ $$\lteval{T}{\colvector{x\\y}} = \colvector{x+y\\2x + y\\x + 2y}$$ is injective, show directly that

\begin{equation*} \set{ \lteval{T}{\vect{e}_1},\, \lteval{T}{\vect{e}_2} } \end{equation*}

is a linearly independent set.

Solution.

We have $$\lteval{T}{\vect{e}_1} = \colvector{1\\2\\1}$$ and $$\lteval{T}{\vect{e}_2} = \colvector{1\\1\\2}\text{.}$$ Putting these into a matrix as columns and row-reducing, we have

\begin{align*} \begin{bmatrix} 1 & 1\\ 2 & 1\\ 1 & 2 \end{bmatrix} &\rref \begin{bmatrix} \leading{1} & 0 \\ 0 & \leading{1}\\ 0 & 0 \end{bmatrix}\text{.} \end{align*}

Thus, the set of vectors $$\set{\lteval{T}{\vect{e}_1},\,\lteval{T}{\vect{e}_2}}$$ is linearly independent.

#### C33.

Given that the linear transformation $$\ltdefn{T}{\complex{3}}{\complex{5}}\text{,}$$

\begin{equation*} \lteval{T}{\colvector{x\\y\\z}} = \begin{bmatrix} 1 & 3 & 2\\ 0 & 1 & 1\\ 1 & 2 & 1\\ 1 & 0 & 1\\ 3 & 1 & 2 \end{bmatrix} \colvector{x\\y\\z} \end{equation*}

is injective, show directly that

\begin{equation*} \set{ \lteval{T}{\vect{e}_1},\, \lteval{T}{\vect{e}_2},\, \lteval{T}{\vect{e}_3} } \end{equation*}

is a linearly independent set.

Solution.

We have

\begin{align*} \lteval{T}{\vect{e}_1} &= \colvector{1\\0\\1\\1\\3} & \lteval{T}{\vect{e}_2} &= \colvector{3\\1\\2\\0\\1} & \lteval{T}{\vect{e}_3} &= \colvector{2\\1\\1\\1\\2}\text{.} \end{align*}

Apply Theorem LIVRN to test the linear independence of $$T\text{.}$$

\begin{align*} \begin{bmatrix} 1 & 3 & 2\\ 0 & 1 & 1\\ 1 & 2 & 1\\ 1 & 0 & 1\\ 3 & 1 & 2 \end{bmatrix} &\rref \begin{bmatrix} \leading{1} & 0 & 0\\ 0 & \leading{1} & 0\\ 0 & 0 & \leading{1} \\ 0 & 0 & 0\\ 0 & 0 & 0 \end{bmatrix} \end{align*}

so since $$r = 3 = n\text{,}$$ the set of vectors $$\set{\lteval{T}{\vect{e}_1},\,\lteval{T}{\vect{e}_2},\,\lteval{T}{\vect{e}_3}}$$ is linearly independent.

#### C40.

Show that the linear transformation $$R$$ is not injective by finding two different elements of the domain, $$\vect{x}$$ and $$\vect{y}\text{,}$$ such that $$\lteval{R}{\vect{x}}=\lteval{R}{\vect{y}}\text{.}$$ ($$S_{22}$$ is the vector space of symmetric $$2\times 2$$ matrices.)

\begin{equation*} \ltdefn{R}{S_{22}}{P_1}\quad \lteval{R}{\begin{bmatrix}a&b\\b&c\end{bmatrix}}=(2a-b+c)+(a+b+2c)x\text{.} \end{equation*}
Solution.

We choose $$\vect{x}$$ to be any vector we like. A particularly cocky choice would be to choose $$\vect{x}=\zerovector\text{,}$$ but we will instead choose

\begin{equation*} \vect{x}= \begin{bmatrix} 2 & -1 \\ -1 & 4 \end{bmatrix}\text{.} \end{equation*}

Then $$\lteval{R}{\vect{x}}=9+9x\text{.}$$ Now compute the kernel of $$R\text{,}$$ which by Theorem KILT we expect to be nontrivial. Setting $$\lteval{R}{\begin{bmatrix}a&b\\b&c\end{bmatrix}}$$ equal to the zero vector, $$\zerovector=0+0x\text{,}$$ and equating coefficients leads to a homogeneous system of equations. Row-reducing the coefficient matrix of this system will allow us to determine the values of $$a\text{,}$$ $$b$$ and $$c$$ that create elements of the null space of $$R\text{,}$$

\begin{equation*} \begin{bmatrix} 2 & -1 & 1 \\ 1 & 1 & 2 \end{bmatrix} \rref \begin{bmatrix} \leading{1} & 0 & 1 \\ 0 & \leading{1} & 1 \end{bmatrix}\text{.} \end{equation*}

We only need a single element of the null space of this coefficient matrix, so we will not compute a precise description of the whole null space. Instead, choose the free variable $$c=2\text{.}$$ Then

\begin{equation*} \vect{z}=\begin{bmatrix} -2 & -2 \\ -2 & 2\end{bmatrix} \end{equation*}

is the corresponding element of the kernel. We compute the desired $$\vect{y}$$ as

\begin{equation*} \vect{y}=\vect{x}+\vect{z}= \begin{bmatrix} 2 & -1 \\ -1 & 4 \end{bmatrix} + \begin{bmatrix} -2 & -2 \\ -2 & 2\end{bmatrix} = \begin{bmatrix} 0 & -3 \\ -3 & 6 \end{bmatrix}\text{.} \end{equation*}

Then check that $$\lteval{R}{\vect{y}}=9+9x\text{.}$$

#### M60.

Suppose $$U$$ and $$V$$ are vector spaces. Define the function $$\ltdefn{Z}{U}{V}$$ by $$\lteval{Z}{\vect{u}}=\zerovector_{V}$$ for every $$\vect{u}\in U\text{.}$$ Then by Exercise LT.M60, $$Z$$ is a linear transformation. Formulate a condition on $$U$$ that is equivalent to $$Z$$ being an injective linear transformation. In other words, fill in the blank to complete the following statement (and then give a proof): $$Z$$ is injective if and only if $$U$$ is . (See Exercise SLT.M60, Exercise IVLT.M60.)

#### T10.

Suppose $$\ltdefn{T}{U}{V}$$ is a linear transformation. For which vectors $$\vect{v}\in V$$ is $$\preimage{T}{\vect{v}}$$ a subspace of $$U\text{?}$$

Solution.

Suppose that $$\preimage{T}{\vect{v}}$$ is a subspace of $$U\text{.}$$ Then $$\preimage{T}{\vect{v}}$$ is nonempty so we can apply Theorem KPI, and assert the existence of a vector $$\vect{u}\in\preimage{T}{\vect{v}}$$ so that $$\preimage{T}{\vect{v}}=\vect{u}+\krn{T}\text{.}$$ Furthermore, if $$\preimage{T}{\vect{v}}$$ is a subspace, then $$\zerovector\in\preimage{T}{\vect{v}}\text{,}$$ so there exists a vector $$\vect{z}\in\krn{T}$$ such that $$\zerovector=\vect{u}+\vect{z}\text{.}$$ Now

\begin{align*} \vect{v}&=\vect{v}+\zerovector&& \knowl{./knowl/property-Z.html}{\text{Property Z}}\\ &=\vect{v}+\lteval{T}{\vect{z}}&& \knowl{./knowl/definition-KLT.html}{\text{Definition KLT}}\\ &=\lteval{T}{\vect{u}}+\lteval{T}{\vect{z}}&& \knowl{./knowl/definition-PI.html}{\text{Definition PI}}\\ &=\lteval{T}{\vect{u}+\vect{z}}&& \knowl{./knowl/definition-LT.html}{\text{Definition LT}}\\ &=\lteval{T}{\zerovector}\\ &=\zerovector&& \knowl{./knowl/theorem-LTTZZ.html}{\text{Theorem LTTZZ}}\text{.} \end{align*}

So we our hypothesis that the preimage is a subspace has lead to the conclusion that $$\vect{v}$$ could only be one vector, the zero vector. We still need to verify that $$\preimage{T}{\zerovector}$$ is indeed a subspace, but since $$\preimage{T}{\zerovector}=\krn{T}$$ this is just Theorem KLTS.

#### T15.

Suppose that $$\ltdefn{T}{U}{V}$$ and $$\ltdefn{S}{V}{W}$$ are linear transformations. Prove the following relationship between kernels.

\begin{equation*} \krn{T}\subseteq\krn{\compose{S}{T}}\text{.} \end{equation*}
Solution.

We are asked to prove that $$\krn{T}$$ is a subset of $$\krn{\compose{S}{T}}\text{.}$$ Employing Definition SSET, choose $$\vect{x}\in\krn{T}\text{.}$$ Then we know that $$\lteval{T}{\vect{x}}=\zerovector\text{.}$$ So

\begin{align*} \lteval{\left(\compose{S}{T}\right)}{\vect{x}}&=\lteval{S}{\lteval{T}{\vect{x}}}&& \knowl{./knowl/definition-LTC.html}{\text{Definition LTC}}\\ &=\lteval{S}{\zerovector}&&\vect{x}\in\krn{T}\\ &=\zerovector&& \knowl{./knowl/theorem-LTTZZ.html}{\text{Theorem LTTZZ}}\text{.} \end{align*}

This qualifies $$\vect{x}$$ for membership in $$\krn{\compose{S}{T}}\text{.}$$

#### T20.

Suppose that $$A$$ is an $$m\times n$$ matrix. Define the linear transformation $$T$$ by

\begin{equation*} \ltdefn{T}{\complex{n}}{\complex{m}},\quad \lteval{T}{\vect{x}}=A\vect{x}\text{.} \end{equation*}

Prove that the kernel of $$T$$ equals the null space of $$A\text{,}$$ $$\krn{T}=\nsp{A}\text{.}$$

Solution.

This is an equality of sets, so we want to establish two subset conditions (Definition SE).

First, show $$\nsp{A}\subseteq\krn{T}\text{.}$$ Choose $$\vect{x}\in\nsp{A}\text{.}$$ Check to see if $$\vect{x}\in\krn{T}\text{,}$$

\begin{align*} \lteval{T}{\vect{x}} &=A\vect{x}&& \text{Definition of }T\\ &=\zerovector&& \vect{x}\in\nsp{A}\text{.} \end{align*}

So by Definition KLT, $$\vect{x}\in\krn{T}$$ and thus $$\nsp{A}\subseteq\krn{T}\text{.}$$

Now, show $$\krn{T}\subseteq\nsp{A}\text{.}$$ Choose $$\vect{x}\in\krn{T}\text{.}$$ Check to see if $$\vect{x}\in\nsp{A}\text{,}$$

\begin{align*} A\vect{x} &=\lteval{T}{\vect{x}}&& \text{Definition of }T\\ &=\zerovector&& \vect{x}\in\krn{T}\text{.} \end{align*}

So by Definition NSM, $$\vect{x}\in\nsp{A}$$ and thus $$\krn{T}\subseteq\nsp{A}\text{.}$$