Ch 1.8 - Introduction to Linear Transformations

Linear Dependence

  • Definition of Linear Dependence Refresher: If a set of vectors V<em>1,V</em>2,V<em>3{V<em>1, V</em>2, V<em>3} is linearly dependent, it means there exist scalars C</em>1C</em>1, C<em>2C<em>2, and C</em>3C</em>3, not all zero, such that C<em>1V</em>1+C<em>2V</em>2+C<em>3V</em>3=0C<em>1V</em>1 + C<em>2V</em>2 + C<em>3V</em>3 = 0.

  • Extension to a Larger Set: If C<em>1V</em>1+C<em>2V</em>2+C<em>3V</em>3=0C<em>1V</em>1 + C<em>2V</em>2 + C<em>3V</em>3 = 0 holds, then adding 0extV<em>4=00 ext{ V}<em>4 = 0 to both sides results in C</em>1V<em>1+C</em>2V<em>2+C</em>3V<em>3+0V</em>4=0C</em>1V<em>1 + C</em>2V<em>2 + C</em>3V<em>3 + 0V</em>4 = 0. Since not all scalars (certainly C<em>1,C</em>2,C<em>3C<em>1, C</em>2, C<em>3 and also 00 for V</em>4V</em>4) are zero, the set V<em>1,V</em>2,V<em>3,V</em>4{V<em>1, V</em>2, V<em>3, V</em>4} is also linearly dependent.

Introduction to Linear Transformations

Dynamic View of Matrix Multiplication

  • Notation Equivalence: The difference between a matrix equation Ax=bAx = b and its associated vector equation x<em>1a</em>1+ext+x<em>na</em>n=bx<em>1a</em>1 + ext{…} + x<em>na</em>n = b is primarily notational.

  • Transformational Perspective: In various applications (e.g., computer graphics, signal processing), a matrix equation Ax=bAx = b can be viewed dynamically. The matrix AA is considered an entity that "acts" on a vector xx through multiplication, resulting in a new vector called AxAx.

    • Example: The equations [1amp;1 0amp;5 1amp;3][1 3]=[4 15 10]\begin{bmatrix} 1 &amp; 1 \ 0 &amp; 5 \ 1 &amp; 3 \end{bmatrix} \begin{bmatrix} 1 \ 3 \end{bmatrix} = \begin{bmatrix} 4 \ 15 \ 10 \end{bmatrix} and [1amp;1 0amp;5 1amp;3][3 3]=[0 0 0]\begin{bmatrix} 1 &amp; 1 \ 0 &amp; 5 \ 1 &amp; 3 \end{bmatrix} \begin{bmatrix} -3 \ 3 \end{bmatrix} = \begin{bmatrix} 0 \ 0 \ 0 \end{bmatrix} show that multiplication by AA transforms x=[1 3]x = \begin{bmatrix} 1 \ 3 \end{bmatrix} into b=[4 15 10]b = \begin{bmatrix} 4 \ 15 \ 10 \end{bmatrix} and transforms u=[3 3]u = \begin{bmatrix} -3 \ 3 \end{bmatrix} into the zero vector.

  • Solving Ax=bAx = b from this Viewpoint: Finding all vectors xx in RnR^n that are transformed into the vector bb in RmR^m under the action of multiplication by AA. For instance, in Figure 1, a vector from R3R^3 is transformed to a vector in R2R^2.

Definition of a Transformation (Function or Mapping)

  • A transformation (or function or mapping) TT from RnR^n to RmR^m is a rule that assigns to each vector in RnR^n a unique vector in RmR^m.

    • Domain: The set RnR^n is called the domain of TT.

    • Codomain: The set RmR^m is called the codomain of TT.

    • Notation: T:RnRmT: R^n \rightarrow R^m indicates that RnR^n is the domain and RmR^m is the codomain.

    • Image: For a vector xx in RnR^n, the vector T(x)T(x) in RmR^m is called the image of xx (under the action of TT).

    • Range: The set of all possible images T(x)xRn{T(x) \mid x \in R^n} is called the range of TT. The range is a subset of the codomain. (See Figure 2 for an illustration of domain, codomain, and range).

  • Importance of Dynamic View: This dynamic perspective of matrix-vector multiplication is crucial for understanding linear algebra concepts and building mathematical models of physical systems that evolve over time (e.g., in Sections 1.10, 4.8, and Chapter 5).

Matrix Transformations

  • Definition: A matrix transformation is a mapping where for each xx in RnR^n, T(x)T(x) is computed as AxAx, where AA is an m×nm \times n matrix.

    • Simplified Notation: Sometimes denoted by xAxx \mapsto Ax.

  • Domain and Codomain of Matrix Transformations:

    • The domain of TT is RnR^n when the matrix AA has nn columns.

    • The codomain of TT is RmR^m when each column of AA has mm entries.

  • Range of a Matrix Transformation: The range of TT is precisely the set of all linear combinations of the columns of AA, because every image T(x)T(x) is of the form AxAx, which is by definition a linear combination of the columns of AA.

Example 1: Image, Pre-image, Uniqueness, and Range Membership

Let A=[1amp;3 3amp;5 1amp;7]A = \begin{bmatrix} 1 &amp; -3 \ 3 &amp; 5 \ -1 &amp; 7 \end{bmatrix}.
Define a transformation T:R2R3T: R^2 \rightarrow R^3 by T(x)=AxT(x) = Ax, so that T([x<em>1 x</em>2])=[1amp;3 3amp;5 1amp;7][x<em>1 x</em>2]=[x<em>13x</em>2 3x<em>1+5x</em>2 x<em>1+7x</em>2]T(\begin{bmatrix} x<em>1 \ x</em>2 \end{bmatrix}) = \begin{bmatrix} 1 &amp; -3 \ 3 &amp; 5 \ -1 &amp; 7 \end{bmatrix} \begin{bmatrix} x<em>1 \ x</em>2 \end{bmatrix} = \begin{bmatrix} x<em>1 - 3x</em>2 \ 3x<em>1 + 5x</em>2 \ -x<em>1 + 7x</em>2 \end{bmatrix}.

a. Finding the image of u=[3 2]u = \begin{bmatrix} 3 \ 2 \end{bmatrix} under TT (i.e., T(u)T(u)):
T(u)=Au=[1amp;3 3amp;5 1amp;7][3 2]=[(1)(3)+(3)(2) (3)(3)+(5)(2) (1)(3)+(7)(2)]=[36 9+10 3+14]=[3 19 11]T(u) = A u = \begin{bmatrix} 1 &amp; -3 \ 3 &amp; 5 \ -1 &amp; 7 \end{bmatrix} \begin{bmatrix} 3 \ 2 \end{bmatrix} = \begin{bmatrix} (1)(3) + (-3)(2) \ (3)(3) + (5)(2) \ (-1)(3) + (7)(2) \end{bmatrix} = \begin{bmatrix} 3 - 6 \ 9 + 10 \ -3 + 14 \end{bmatrix} = \begin{bmatrix} -3 \ 19 \ 11 \end{bmatrix}

b. Finding an xx in R2R^2 whose image under TT is b=[3 2 5]b = \begin{bmatrix} 3 \ 2 \ -5 \end{bmatrix} (i.e., solving Ax=bAx = b):
This requires solving the matrix equation Ax=bAx = b:
[1amp;3 3amp;5 1amp;7][x<em>1 x</em>2]=[3 2 5]\begin{bmatrix} 1 &amp; -3 \ 3 &amp; 5 \ -1 &amp; 7 \end{bmatrix} \begin{bmatrix} x<em>1 \ x</em>2 \end{bmatrix} = \begin{bmatrix} 3 \ 2 \ -5 \end{bmatrix}
We form the augmented matrix and row reduce it:
[1amp;3amp;3 3amp;5amp;2 1amp;7amp;5][1amp;3amp;3 0amp;14amp;7 0amp;4amp;2](R<em>23R</em>1R<em>2,R</em>3+R<em>1R</em>3)\begin{bmatrix} 1 &amp; -3 &amp; 3 \ 3 &amp; 5 &amp; 2 \ -1 &amp; 7 &amp; -5 \end{bmatrix} \sim \begin{bmatrix} 1 &amp; -3 &amp; 3 \ 0 &amp; 14 &amp; -7 \ 0 &amp; 4 &amp; -2 \end{bmatrix} \quad (R<em>2 - 3R</em>1 \rightarrow R<em>2, R</em>3 + R<em>1 \rightarrow R</em>3)
[1amp;3amp;3 0amp;1amp;.5 0amp;4amp;2](R<em>2/14R</em>2)\sim \begin{bmatrix} 1 &amp; -3 &amp; 3 \ 0 &amp; 1 &amp; -.5 \ 0 &amp; 4 &amp; -2 \end{bmatrix} \quad (R<em>2 / 14 \rightarrow R</em>2)
[1amp;3amp;3 0amp;1amp;.5 0amp;0amp;0](R<em>34R</em>2R<em>3)\sim \begin{bmatrix} 1 &amp; -3 &amp; 3 \ 0 &amp; 1 &amp; -.5 \ 0 &amp; 0 &amp; 0 \end{bmatrix} \quad (R<em>3 - 4R</em>2 \rightarrow R<em>3) From the reduced echelon form: x</em>2=.5x</em>2 = -.5
x<em>13x</em>2=3x<em>13(.5)=3x</em>1+1.5=3x1=1.5x<em>1 - 3x</em>2 = 3 \Rightarrow x<em>1 - 3(-.5) = 3 \Rightarrow x</em>1 + 1.5 = 3 \Rightarrow x_1 = 1.5
Thus, the vector is x=[1.5 .5]x = \begin{bmatrix} 1.5 \ -.5 \end{bmatrix}. The image of this xx under TT is the given vector bb.

c. Uniqueness of xx: From the row reduction in part (b), the system Ax=bAx=b has a unique solution (no free variables). Therefore, there is exactly one xx whose image under TT is bb.

d. Determining if c=[3 2 5]c = \begin{bmatrix} 3 \ 2 \ -5 \end{bmatrix} (this seems to be a typo in the original text, as 'c' also represented the third column of the original augmented matrix in 'b'. Let's use the actual definition provided in 'd') c=[3 2 8]c = \begin{bmatrix} 3 \ 2 \ 8 \end{bmatrix} is in the range of TT:
This asks if cc is the image of some xx in R2R^2, i.e., if c=T(x)c = T(x) for some xx. This means checking if the system Ax=cAx=c is consistent. We row reduce the augmented matrix:
[1amp;3amp;3 3amp;5amp;2 1amp;7amp;8][1amp;3amp;3 0amp;14amp;7 0amp;4amp;11](R<em>23R</em>1R<em>2,R</em>3+R<em>1R</em>3)\begin{bmatrix} 1 &amp; -3 &amp; 3 \ 3 &amp; 5 &amp; 2 \ -1 &amp; 7 &amp; 8 \end{bmatrix} \sim \begin{bmatrix} 1 &amp; -3 &amp; 3 \ 0 &amp; 14 &amp; -7 \ 0 &amp; 4 &amp; 11 \end{bmatrix} \quad (R<em>2 - 3R</em>1 \rightarrow R<em>2, R</em>3 + R<em>1 \rightarrow R</em>3)
[1amp;3amp;3 0amp;1amp;.5 0amp;4amp;11](R<em>2/14R</em>2)\sim \begin{bmatrix} 1 &amp; -3 &amp; 3 \ 0 &amp; 1 &amp; -.5 \ 0 &amp; 4 &amp; 11 \end{bmatrix} \quad (R<em>2 / 14 \rightarrow R</em>2)
[1amp;3amp;3 0amp;1amp;.5 0amp;0amp;13](R<em>34R</em>2R<em>3)\sim \begin{bmatrix} 1 &amp; -3 &amp; 3 \ 0 &amp; 1 &amp; -.5 \ 0 &amp; 0 &amp; 13 \end{bmatrix} \quad (R<em>3 - 4R</em>2 \rightarrow R<em>3) The third row corresponds to the equation 0x</em>1+0x2=130x</em>1 + 0x_2 = 13, which simplifies to 0=130=13. This is a contradiction, so the system is inconsistent. Therefore, cc is not in the range of TT.

  • Summary of Example 1: Example 1c is a uniqueness problem (is bb the image of a unique xx?), and Example 1d is an existence problem (does there exist an xx whose image is cc?).

Geometric Matrix Transformations

These examples illustrate the dynamic view of matrices as operators that transform vectors.

  • Example 2: Projection Transformation

    • Let A=[1amp;0amp;0 0amp;1amp;0 0amp;0amp;0]A = \begin{bmatrix} 1 &amp; 0 &amp; 0 \ 0 &amp; 1 &amp; 0 \ 0 &amp; 0 &amp; 0 \end{bmatrix}. The transformation T(x)=AxT(x) = Ax projects points in R3R^3 onto the x<em>1x</em>2x<em>1x</em>2-plane.

    • T([x<em>1 x</em>2 x<em>3])=[1amp;0amp;0 0amp;1amp;0 0amp;0amp;0][x</em>1 x<em>2 x</em>3]=[x<em>1 x</em>2 0]T(\begin{bmatrix} x<em>1 \ x</em>2 \ x<em>3 \end{bmatrix}) = \begin{bmatrix} 1 &amp; 0 &amp; 0 \ 0 &amp; 1 &amp; 0 \ 0 &amp; 0 &amp; 0 \end{bmatrix} \begin{bmatrix} x</em>1 \ x<em>2 \ x</em>3 \end{bmatrix} = \begin{bmatrix} x<em>1 \ x</em>2 \ 0 \end{bmatrix}.

  • Example 3: Shear Transformation

    • Let A=[1amp;1 0amp;1]A = \begin{bmatrix} 1 &amp; 1 \ 0 &amp; 1 \end{bmatrix}. The transformation T:R2R2T: R^2 \rightarrow R^2 defined by T(x)=AxT(x) = Ax is called a shear transformation.

    • It deforms a shape (e.g., a 2×22 \times 2 square) into a sheared parallelogram (See Figure 4).

    • Key Idea: TT maps line segments onto line segments. By checking the images of the corners of the square, one can see the transformation.

      • Image of [0 2]\begin{bmatrix} 0 \ 2 \end{bmatrix} is A[0 2]=[1amp;1 0amp;1][0 2]=[2 2]A \begin{bmatrix} 0 \ 2 \end{bmatrix} = \begin{bmatrix} 1 &amp; 1 \ 0 &amp; 1 \end{bmatrix} \begin{bmatrix} 0 \ 2 \end{bmatrix} = \begin{bmatrix} 2 \ 2 \end{bmatrix}.

      • Image of [2 2]\begin{bmatrix} 2 \ 2 \end{bmatrix} is A[2 2]=[1amp;1 0amp;1][2 2]=[4 2]A \begin{bmatrix} 2 \ 2 \end{bmatrix} = \begin{bmatrix} 1 &amp; 1 \ 0 &amp; 1 \end{bmatrix} \begin{bmatrix} 2 \ 2 \end{bmatrix} = \begin{bmatrix} 4 \ 2 \end{bmatrix}.

    • Geometric Effect: The transformation deforms the square as if its top were pushed to the right while the base is held fixed. Shear transformations are observed in physics, geology, and crystallography.

Linear Transformations

Formal Definition

  • Theorem 5 in Section 1.4 established properties for matrix transformations: A(u+v)=Au+AvA(u+v) = Au+Av and A(cu)=cAuA(cu) = cAu. These properties define the most important class of transformations in linear algebra.

  • A transformation (or mapping) TT is linear if:

    • (i) Additivity: T(u+v)=T(u)+T(v)T(u+v) = T(u) + T(v) for all vectors u,vu, v in the domain of TT.

    • (ii) Homogeneity of Degree 1: T(cu)=cT(u)T(cu) = cT(u) for all scalars cc and all vectors uu in the domain of TT.

Important Notes on Linear Transformations

  • Matrix Transformations are Linear: Every matrix transformation (T(x)=AxT(x)=Ax) is a linear transformation. (However, not all linear transformations are matrix transformations; this will be explored in Chapters 4 and 5).

  • Preservation of Operations: Linear transformations preserve the operations of vector addition and scalar multiplication. Property (i) means that applying TT after adding vectors uu and vv is the same as applying TT to uu and vv separately and then adding their images.

Useful Facts derived from Linearity

If TT is a linear transformation, then:

  1. Image of the Zero Vector: T(0)=0T(0) = 0

    • Proof: From condition (ii), T(0)=T(0u)=0T(u)=0T(0) = T(0u) = 0T(u) = 0 (where 00 on the left is the scalar, and 00 on the right is the zero vector in the codomain).

  2. Linear Combination Property: T(cu+dv)=cT(u)+dT(v)T(cu+dv) = cT(u) + dT(v)

    • Proof: This property requires both (i) and (ii):
      T(cu+dv)=T(cu)+T(dv)(by property (i))T(cu+dv) = T(cu) + T(dv) \quad \text{(by property (i))}
      =cT(u)+dT(v)(by property (ii))= cT(u) + dT(v) \quad \text{(by property (ii))}

    • Implication: If a transformation satisfies T(cu+dv)=cT(u)+dT(v)T(cu+dv) = cT(u) + dT(v) for all u,vu, v and c,dc, d, it must be linear. (Setting c=d=1c=d=1 gives additivity, and setting d=0d=0 gives scalar multiplication property).

  • Superposition Principle (Generalization): Repeated application of the linear combination property leads to a useful generalization: T(c<em>1v</em>1+c<em>2v</em>2++c<em>pv</em>p)=c<em>1T(v</em>1)+c<em>2T(v</em>2)++c<em>pT(v</em>p)T(c<em>1v</em>1 + c<em>2v</em>2 + \text{…} + c<em>pv</em>p) = c<em>1T(v</em>1) + c<em>2T(v</em>2) + \text{…} + c<em>pT(v</em>p)

    • Significance in Science/Engineering: This is widely known as the superposition principle in engineering and physics. If v<em>1,,v</em>pv<em>1, \text{…}, v</em>p are inputs (signals) to a system, and T(v<em>1),,T(v</em>p)T(v<em>1), \text{…}, T(v</em>p) are the corresponding responses, then a system satisfies the superposition principle if its response to a linear combination of inputs is the same linear combination of the responses to the individual inputs. This concept is fundamental in many fields.

Examples of Linear Transformations

Example 4: Contraction and Dilation Transformation
  • Given a scalar rr, define T:R2R2T: R^2 \rightarrow R^2 by T(x)=rxT(x) = rx.

    • This is a contraction when 0 < r < 1 (shrinks vectors).

    • This is a dilation when r > 1 (stretches vectors).

  • Proof of Linearity (for r=3r=3):
    Let u,vu, v be vectors in R2R^2 and c,dc, d be scalars.
    T(cu+dv)=r(cu+dv)(by definition of T)T(cu+dv) = r(cu+dv) \quad \text{(by definition of } T)
    =c(ru)+d(rv)(by vector arithmetic)= c(ru) + d(rv) \quad \text{(by vector arithmetic)}
    =cT(u)+dT(v)(by definition of T)= cT(u) + dT(v) \quad \text{(by definition of } T)
    Since T(cu+dv)=cT(u)+dT(v)T(cu+dv) = cT(u)+dT(v), the transformation TT is linear. (See Figure 5 for a dilation example).

Example 5: Rotation Transformation
  • Define a linear transformation T:R2R2T: R^2 \rightarrow R^2 by T(x)=[0amp;1 1amp;0]xT(x) = \begin{bmatrix} 0 &amp; -1 \ 1 &amp; 0 \end{bmatrix}x.

  • Let u=[4 1]u = \begin{bmatrix} 4 \ 1 \end{bmatrix}, v=[1 3]v = \begin{bmatrix} 1 \ 3 \end{bmatrix}, and u+v=[5 4]u+v = \begin{bmatrix} 5 \ 4 \end{bmatrix}.

    • T(u)=[0amp;1 1amp;0][4 1]=[1 4]T(u) = \begin{bmatrix} 0 &amp; -1 \ 1 &amp; 0 \end{bmatrix} \begin{bmatrix} 4 \ 1 \end{bmatrix} = \begin{bmatrix} -1 \ 4 \end{bmatrix}

    • T(v)=[0amp;1 1amp;0][1 3]=[3 1]T(v) = \begin{bmatrix} 0 &amp; -1 \ 1 &amp; 0 \end{bmatrix} \begin{bmatrix} 1 \ 3 \end{bmatrix} = \begin{bmatrix} -3 \ 1 \end{bmatrix}

    • T(u+v)=[0amp;1 1amp;0][5 4]=[4 5]T(u+v) = \begin{bmatrix} 0 &amp; -1 \ 1 &amp; 0 \end{bmatrix} \begin{bmatrix} 5 \ 4 \end{bmatrix} = \begin{bmatrix} -4 \ 5 \end{bmatrix}

  • Verification of Additivity: Notice that T(u)+T(v)=[1 4]+[3 1]=[4 5]T(u)+T(v) = \begin{bmatrix} -1 \ 4 \end{bmatrix} + \begin{bmatrix} -3 \ 1 \end{bmatrix} = \begin{bmatrix} -4 \ 5 \end{bmatrix}, which is indeed equal to T(u+v)T(u+v).

  • Geometric Interpretation: This transformation rotates vectors (e.g., u,v,extandu+vu, v, ext{ and } u+v) counterclockwise about the origin through 9090^\circ (See Figure 6). It transforms the entire parallelogram determined by uu and vv into the parallelogram determined by T(u)T(u) and T(v)T(v).

Example 6: Cost Transformation for Manufacturing Production
  • Scenario: A company manufactures two products, B and C.

  • Unit Cost Matrix UU: From Section 1.3, we construct a