## ChapterRRepresentations

Previous work with linear transformations may have convinced you that we can convert most questions about linear transformations into questions about systems of equations or properties of subspaces of $$\complex{m}\text{.}$$ In this section we begin to make these vague notions precise. We have used the word â€śrepresentationâ€ť prior, but it will get a heavy workout in this chapter. In many ways, everything we have studied so far was in preparation for this chapter.

## Annotated Acronyms R.

### DefinitionÂ VR.

Matrix representations build on vector representations, so this is the definition that gets us started. A representation depends on the choice of a single basis for the vector space. TheoremÂ VRRB is what tells us this idea might be useful.

### TheoremÂ VRILT.

As an invertible linear transformation, vector representation allows us to translate, back and forth, between abstract vector spaces ($$V$$) and concrete vector spaces ($$\complex{n}$$). This is key to all our notions of representations in this chapter.

### TheoremÂ CFDVS.

Every vector space with finite dimension â€ślooks likeâ€ť a vector space of column vectors. Vector representation is the isomorphism that establishes that these vector spaces are isomorphic.

### DefinitionÂ MR.

Building on the definition of a vector representation, we define a representation of a linear transformation, determined by a choice of two bases, one for the domain and one for the codomain. Notice that vectors are represented by columnar lists of scalars, while linear transformations are represented by rectangular tables of scalars. Building a matrix representation is as important a skill as row-reducing a matrix.

### TheoremÂ FTMR.

DefinitionÂ MR is not really very interesting until we have this theorem. The second form tells us that we can compute outputs of linear transformations via matrix multiplication, along with some bookkeeping for vector representations. Searching forward through the text on â€śFTMRâ€ť is an interesting exercise. You will find reference to this result buried inside many key proofs at critical points, and it also appears in numerous examples and solutions to exercises.

### TheoremÂ MRCLT.

Turns out that matrix multiplication is really a very natural operation, it is just the chaining together (composition) of functions (linear transformations). Beautiful. Even if you do not try to work the problem, study SolutionÂ T80.1 for more insight.

### TheoremÂ KNSI.

Kernels â€śareâ€ť null spaces. For this reason you will see these terms used interchangeably.

### TheoremÂ RCSI.

Ranges â€śareâ€ť column spaces. For this reason you will see these terms used interchangeably.

### TheoremÂ IMR.

Invertible linear transformations are represented by invertible (nonsingular) matrices.

### TheoremÂ NME9.

The NMEx series has always been important, but we have held off saying so until now. This is the end of the line for this one, so it is a good time to contemplate all that it means.

### TheoremÂ SCB.

Diagonalization back in SectionÂ SD was really a change of basis to achieve a diagonal matrix repesentation. Maybe we should be highlighting the more general TheoremÂ MRCB here, but its overly technical description just is not as appealing. However, it will be important in some of the matrix decompositions you will encounter in a future course in linear algebra.

### TheoremÂ EER.

This theorem, with the companion definition, DefinitionÂ EELT, tells us that eigenvalues, and eigenvectors, are fundamentally a characteristic of linear transformations (not matrices). If you study matrix decompositions in a future course in linear algebra you will come to appreciate that almost all of a matrix's secrets can be unlocked with knowledge of the eigenvalues and eigenvectors.

### TheoremÂ OD.

Can you imagine anything nicer than an orthonormal diagonalization? A basis of pairwise orthogonal, unit norm, eigenvectors that provide a diagonal representation for a matrix? Here we learn just when this can happen â€” precisely when a matrix is normal, which is a disarmingly simple property to define.