Linear Algebra

Transformations

7/13 in Linear Algebra. See all.

Linear Transformation
A transformation is a function that takes in a vector and outputs another vector — not at all different from our typical AP Calculus BC-style vector-valued functions. Now, however, we're working with vectors that could be anything, not just Cartesian coordinates. To denote that a transformation TT has domain V\mathbb{V} and range W\mathbb{W}, we write T:VWT:\mathbb{V}\to \mathbb{W}. A transformation T:VWT:\mathbb{V}\to \mathbb{W} (mathematically) is linear if for all scalars v1,v2v_1, v_2 and scalars α,β\alpha,\beta: T(αv1+βv2)=αT(v1)+βT(v2).T(\alpha v_1 + \beta v_2) = \alpha T(v_1)+\beta T(v_2). For any vector space V\mathbb{V} with basis BB, the transformation ϕ:VRn\phi:\mathbb{V}\to \mathbb{R}^n defined by ϕ(v)=(v)B\phi(v)=(v)_B is linear, one-to-one, and onto, where (v)B(v)_B is the coordinate vector of vv with respect to BB.

How do we describe these numerically? An interesting observation is that we only need to follow where the basis vectors end up — since the linear transformation preserves addition and multiplication (linear combination), a vector like v=2i^+3j^\vec{v}=2\hat{\textbf{i}}+3\hat{\textbf{j}} transformed is the same linear combination of the transformed i^\hat{\textbf{i}} and j^\hat{\textbf{j}} as the original vector was of i^\hat{\textbf{i}} and j^\hat{\textbf{j}}! So, a linear transformation is completely described by the coordinates where i^\hat{\textbf{i}} and j^\hat{\textbf{j}} land. This can be packaged into a matrix where each column describes the transformed locations of i^\hat{\textbf{i}} and j^\hat{\textbf{j}}. In the most general (2×22\times 2) case:

[abcd][xy]=[ax+bycx+dy]\begin{bmatrix}a & b \\ c & d\end{bmatrix}\cdot\begin{bmatrix}x \\ y\end{bmatrix}=\begin{bmatrix}ax+by\\cx + dy\end{bmatrix}

Ex. For the 2x2 matrix given, find where [57]is transformed to.[3221][57]=5[32]+7[21]=[1510]+[147]=[293]\begin{align*}\text{For the 2x2 matrix given, find where }&\left[\begin{array}{c}5 \\ 7 \end{array}\right] \text{is transformed to.} \newline\begin{bmatrix}3 & 2 \\-2 & 1\end{bmatrix}\cdot\begin{bmatrix}5 \\ 7\end{bmatrix}&=5\cdot\begin{bmatrix}3 \\ -2\end{bmatrix}+7\cdot\begin{bmatrix}2 \\ 1\end{bmatrix}\\&=\begin{bmatrix}15\\-10\end{bmatrix}+\begin{bmatrix}14\\7\end{bmatrix}\\&=\boxed{\begin{bmatrix}29\\-3\end{bmatrix}}\end{align*} If V\mathbb{V} and W\mathbb{W} are vector spaces with bases BB and CC, respectively, and T:VWT:\mathbb{V}\to \mathbb{W} is linear, then there exists an m×nm\times n matrix AA such that for every vector vVv\in \mathbb{V}, (Tv)C=A(v)B(Tv)_C=A(v)_B. This matrix is obviously dependent on the bases BB and CC; to reflect this dependence, we write ABCA_B^C, where the subscript reflects the chosen basis for the domain of the transformation, and the superscript reflects the chosen basis for the range of the basis.

Image of a Transformation
The image of a transformation T:VWT:\mathbb{V}\to \mathbb{W} is denoted image(T)image(T), and is the subset of W\mathbb{W} of all {Tv    vV}\left\{ Tv\; |\; v\in \mathbb{V} \right\}. For any linear transformation T:VWT:\mathbb{V}\to \mathbb{W},

  1. 0image(T)\textbf{0}\in image(T).
  2. image(T)image(T) is a subspace of WW.
    1. TT is onto if and only if r(T)=dim(W)r(T)=\dim(\mathbb{W}).

If dim(V)=dim(W)\dim(\mathbb{V})=\dim(\mathbb{W}) and T:VWT:\mathbb{V}\to \mathbb{W} is linear, then TT is onto if and only if TT is one-to-one.

Rank
Some squishes (*sigh*... transformations) are more "squishy" than others — 3d space squished into a line is more "squished" than 3d space squished into a plane, but both transformations have determinant 00. The rank describes this — it is the number of dimensions in the output. A transformation whose output is a line has a rank of 11, while a transformation whose output is a plane has a rank of 22. Formally, the rank of a linear transformation T:VWT:\mathbb{V}\to \mathbb{W} is r(T)=dim(image(T)).r(T)=\dim(image(T)).

Null Space/Kernel
The set of vectors that get transformed into the origin during a transformation. For a transformation with full rank, the null space is just the zero vector. For a rank zero transformation, the null space is nn-dimensional space. The kernel (or nullspace) of a transformation is: ker(T)={vV    Tv=0}\ker(T)=\left\{v\in \mathbb{V} \;|\; Tv=0\right\} For any transformation T:VWT:\mathbb{V}\to \mathbb{W},

  1. 0ker(T)\textbf{0}\in \ker(T).
  2. ker(T)\ker(T) is a subspace of V\mathbb{V}. If VV is finite dimensional and T:VWT:\mathbb{V}\to \mathbb{W} is linear, then r(T)+null(T)=dim(V)r(T)+null(T)=\dim(\mathbb{V}). A transformation is one-to-one if and only if ker(T)={0}\ker(T)=\left\{\textbf{0}\right\}.

Nullity
The dimension of the nullspace of TT is called the nullity of TT, and is denoted by null(T)null(T).

Injective
A transformation TT is injective (one-to-one) if for all vectors v1,v2Vv_1, v_2\in \mathbb{V}, v1v2v_1\neq v_2 implies Tv1Tv2Tv_1\neq Tv_2.

Surjective
A transformation TT is onto (or surjective) if W=image(T)\mathbb{W}=image(T). So TT is onto W\mathbb{W} if every element in the range is in the image of TT.

Isomorphic
If T:VWT:\mathbb{V}\to \mathbb{W} is linear, onto, and one-to-one, then it is isomorphic, and an isomorphism exists between V\mathbb{V} and W\mathbb{W}, denoted VW\mathbb{V}\cong \mathbb{W}. It is reflective, transitive, and symmetric. Since isomorphism is an equivalence relation, VW\mathbb{V}\cong \mathbb{W} if and only if dim(V)=dim(W)\dim(\mathbb{V})=\dim(\mathbb{W}) (there exists an isomorphism T1:VRnT_1:\mathbb{V}\to \mathbb{R}^n and T2:WRnT_2:\mathbb{W}\to \mathbb{R}^n). This means that any nn-dimensional vector space is just a copy of any other nn-dimensional vector space, and any vector space (up to the names of its elements) is completely determined by its dimension.

Important Linear Transformations

  • 9090^\circ rotation counterclockwise:
    • [0110]\begin{bmatrix}0 & 1\\-1 & 0\end{bmatrix}
  • Shear (i^\hat{\textbf{i}} remains fixed and j^\hat{\textbf{j}} moves to (1,1)(1, 1)):
    • [1101]\begin{bmatrix}1 & 1\\ 0 & 1\end{bmatrix}
  • Counterclockwise rotation by θ\theta around the origin:
    • [cosθsinθsinθcosθ]\begin{bmatrix}\cos\theta & -\sin\theta \\ \sin\theta & \cos\theta\end{bmatrix}

Similar Matrices
Two matrices AA and BB are similar if there exists an invertible matrix PP such that B=P1APB=P^{-1}AP. Similar matrices have the same rank, determinant, and trace. As we'll see when we get to change of basis and determinants, similar matrices are matrices that represent the same linear transformation, but with different bases. Similarity (denoted ABA\cong B) is transitive, symmetric, and reflexive.

Transformation Composition (Matrix Multiplication)
Applying one transformation, then another, is still a linear transformation! So, there is a matrix that describes this composition of transformations, and we can call it the "product" of the two original matrices. However, since linear transformations aren't commutative (and are functions), we read matrix multiplication from right to left (just like we read f(g(x))f(g(x))). So how do we multiply matrices AA and BB? A=[abcd],B=[efgh].A=\begin{bmatrix}a & b \\ c & d\end{bmatrix}, B=\begin{bmatrix}e & f \\ g & h\end{bmatrix}. M=BAM=B\cdot A There are two ways to think about it. The first way is the intuitive way: the first column of AA shows where i^\hat{\textbf{i}} is transformed to, and the second shows where j^\hat{\textbf{j}} is transformed to. Then, just multiply the two vectors by BB to get where the two transformed vectors go: [efgh][ac]=[ae+cfag+ch][efgh][bd]=[be+dfbg+dh]so, the composed matrix is[ae+cfbe+dfag+chbg+dh]\begin{align*}\begin{bmatrix}e & f\\ g & h\end{bmatrix}\cdot\begin{bmatrix}a\\c\end{bmatrix}=\begin{bmatrix}ae + cf\\ ag + ch\end{bmatrix}\\\\\begin{bmatrix}e & f\\g & h\end{bmatrix}\cdot\begin{bmatrix}b\\d\end{bmatrix}=\begin{bmatrix}be+df\\bg+dh\end{bmatrix}\\\\\text{so, the composed matrix is}\\\boxed{\begin{bmatrix}ae+cf & be+df\\ag+ch&bg+dh\end{bmatrix}}\end{align*} And this shows us the second way of thinking about multiplying matrices: memorize the formula!

P.S. I do not support the second way.


If you recall, in an earlier section we talked about the derivative as a linear transformation, denoted by a matrix with respect to the standard basis of Pn\mathbb{P}^n. Do you see how we arrived at that matrix now? We just followed where each basis vector went!