跳到主要内容

Linear Transformations and Matrices

Linear transformation

Let VV and WW be vector spaces over a field FF. A function T:VWT : V \to W is called a linear transformation from VV to WW if, for all x,yVx, y \in V and cFc \in F, the following conditions hold:

(a)T(x+y)=T(x)+T(y)\begin{equation*} \text{(a)}\quad T(x + y) = T(x) + T(y) \end{equation*} (b)T(cx)=cT(x)\begin{equation*} \text{(b)}\quad T(cx) = c\,T(x) \end{equation*}

A transformation T:VWT : V \to W is linear if and only if

T(cx+y)=cT(x)+T(y)\begin{equation*} T(cx + y) = c\,T(x) + T(y) \end{equation*}

for all x,yVx, y \in V and all scalars cFc \in F.

Identity transformation

The identity transformation IV:VVI_V : V \to V is defined by

IV(x)=xfor all xV.\begin{equation*} I_V(x) = x \quad \text{for all } x \in V. \end{equation*}

We often write II instead of IVI_V.

Zero transformation

The zero transformation T0:VWT_0 : V \to W is defined by

T0(x)=0for all xV.\begin{equation*} T_0(x) = 0 \quad \text{for all } x \in V. \end{equation*}

It is clear that both of these transformations are linear.

Let VV and WW be vector spaces, and let T:VWT : V \to W be linear.

Null space (Kernel)

We define the null space (or kernel) of TT, denoted N(T)N(T), to be the set of all vectors xVx \in V such that T(x)=0T(x) = 0. That is,

N(T)={xV:T(x)=0}.\begin{equation*} N(T) = \{\, x \in V : T(x) = 0 \,\}. \end{equation*}

Range (Image)

We define the range (or image) of TT, denoted R(T)R(T), to be the subset of WW consisting of all images T(x)T(x) of vectors xVx \in V. That is,

R(T)={T(x):xV}.\begin{equation*} R(T) = \{\, T(x) : x \in V \,\}. \end{equation*}

Let VV and WW be vector spaces, and let T:VWT : V \to W be linear. If N(T)N(T) and R(T)R(T) are finite-dimensional, then we define the nullity of TT, denoted nullity(T)\operatorname{nullity}(T), and the rank of TT, denoted rank(T)\operatorname{rank}(T), to be the dimensions of N(T)N(T) and R(T)R(T), respectively.

Theorems

  1. Let VV and WW be vector spaces and T:VWT : V \to W be linear. Then N(T)N(T) and R(T)R(T) are subspaces of VV and WW, respectively.

  2. Let VV and WW be vector spaces, and let T:VWT : V \to W be linear. If β={v1,v2,,vn}\beta = \{v_1, v_2, \ldots, v_n\} is a basis for VV, then

    R(T)=span(T(β))=span({T(v1),T(v2),,T(vn)}).\begin{equation*} R(T) = \operatorname{span}(T(\beta)) = \operatorname{span}(\{T(v_1), T(v_2), \ldots, T(v_n)\}). \end{equation*}
  3. (Dimension Theorem) Let VV and WW be vector spaces, and let T:VWT : V \to W be linear. If VV is finite-dimensional, then

    nullity(T)+rank(T)=dim(V).\begin{equation*} \operatorname{nullity}(T) + \operatorname{rank}(T) = \dim(V). \end{equation*}
  4. Let VV and WW be vector spaces, and let T:VWT: V \to W be linear. Then TT is one-to-one if and only if N(T)={0}N(T) = \{0\}.

  5. Let VV and WW be vector spaces of equal (finite) dimension, and let T:VWT: V \to W be linear. Then the following are equivalent:

    • TT is one-to-one.

    • TT is onto.

    • rank(T)=dim(V)\operatorname{rank}(T) = \dim(V).

  6. Let VV and WW be vector spaces over FF, and suppose that {v1,v2,,vn}\{v_1, v_2, \ldots, v_n\} is a basis for VV. For w1,w2,,wnw_1, w_2, \ldots, w_n in WW, there exists exactly one linear transformation T:VWT: V \to W such that T(vi)=wiT(v_i) = w_i for i=1,2,,ni = 1, 2, \ldots, n.

    Corollary: Let VV and WW be vector spaces, and suppose that VV has a finite basis {v1,v2,,vn}\{v_1, v_2, \ldots, v_n\}. If U,T:VWU, T : V \to W are linear and U(vi)=T(vi)U(v_i) = T(v_i) for i=1,2,,ni = 1, 2, \ldots, n, then U=TU = T.

The matrix representation of a linear transformation

Let VV be a finite-dimensional vector space. An ordered basis for VV is a basis for VV endowed with a specific order; that is, an ordered basis for VV is a finite sequence of linearly independent vectors in VV that generates VV.

Coordinate vector

Let β={u1,u2,,un}\beta = \{u_1, u_2, \dots, u_n\} be an ordered basis for a finite-dimensional vector space VV. For xVx \in V, let a1,a2,,ana_1, a_2, \dots, a_n be the unique scalars such that

x=i=1naiui.\begin{equation*} x = \sum_{i=1}^{n} a_i u_i. \end{equation*}

We define the coordinate vector of xx relative to β\beta, denoted [x]β[x]_\beta, by

[x]β=(a1a2an).\begin{equation*} [x]_\beta = \begin{pmatrix} a_1 \\ a_2 \\ \vdots \\ a_n \end{pmatrix}. \end{equation*}

Let us now proceed with the promised matrix representation of a linear transformation.

Suppose that VV and WW are finite-dimensional vector spaces with ordered bases β={v1,v2,,vn}\beta = \{v_1, v_2, \dots, v_n\} and γ={w1,w2,,wm}\gamma = \{w_1, w_2, \dots, w_m\}, respectively. Let T:VWT : V \to W be linear. Then for each jj, 1jn1 \le j \le n, there exist unique scalars aijFa_{ij} \in F, 1im1 \le i \le m, such that

T(vj)=i=1maijwifor 1jn.\begin{equation*} T(v_j) = \sum_{i=1}^{m} a_{ij} w_i \qquad \text{for } 1 \le j \le n. \end{equation*}

Using the notation above, we call the m×nm \times n matrix AA defined by Aij=aijA_{ij} = a_{ij} the matrix representation of TT in the ordered bases β\beta and γ\gamma and write

A=[T]βγ.\begin{equation*} A = [T]_{\beta}^{\gamma}. \end{equation*}

If V=WV = W and β=γ\beta = \gamma, then we write simply

A=[T]β.\begin{equation*} A = [T]_{\beta}. \end{equation*}

Notice that the jjth column of AA is simply [T(vj)]γ[\,T(v_j)\,]_{\gamma}. Also observe that if U:VWU : V \to W is a linear transformation such that [U]βγ=[T]βγ[U]_{\beta}^{\gamma} = [T]_{\beta}^{\gamma}, then U=TU = T.

Let T,U:VWT, U : V \to W be arbitrary functions, where VV and WW are vector spaces over FF, and let aFa \in F. We define T+U:VWT + U : V \to W by

(T+U)(x)=T(x)+U(x)for all xV,\begin{equation*} (T + U)(x) = T(x) + U(x) \qquad \text{for all } x \in V, \end{equation*}

and aT:VWaT : V \to W by

(aT)(x)=aT(x)for all xV.\begin{equation*} (aT)(x) = a\,T(x) \qquad \text{for all } x \in V. \end{equation*}

Let VV and WW be vector spaces over FF. We denote the vector space of all linear transformations from VV into WW by L(V,W)\mathcal{L}(V, W). In the case that V=WV = W, we write L(V)\mathcal{L}(V) instead of L(V,W)\mathcal{L}(V, W).

Theorems

  1. Let VV and WW be finite-dimensional vector spaces with ordered bases β\beta and γ\gamma, respectively, and let T,U:VWT, U : V \to W be linear transformations. Then

    (a)

    [T+U]βγ=[T]βγ+[U]βγ\begin{equation*} [T + U]_{\beta}^{\gamma} = [T]_{\beta}^{\gamma} + [U]_{\beta}^{\gamma} \end{equation*}

    and

    (b)

    [aT]βγ=a[T]βγfor all scalars a.\begin{equation*} [aT]_{\beta}^{\gamma} = a [T]_{\beta}^{\gamma} \qquad \text{for all scalars } a. \end{equation*}

Composition of linear transformations and matrix multiplication

Let VV, WW, and ZZ be vector spaces over the same field FF, and let T:VWT : V \to W and U:WZU : W \to Z be linear. Then UT:VZUT : V \to Z is linear.

Matrix multiplication

Let T:VWT : V \to W and U:WZU : W \to Z be linear transformations, and let A=[U]βγA = [U]_{\beta}^{\gamma} and B=[T]αβB = [T]_{\alpha}^{\beta}, where α={v1,v2,,vn}\alpha = \{v_1, v_2, \dots, v_n\}, β={w1,w2,,wm}\beta = \{w_1, w_2, \dots, w_m\}, and γ={z1,z2,,zp}\gamma = \{z_1, z_2, \dots, z_p\} are ordered bases for VV, WW, and ZZ, respectively.

We would like to define the product ABAB of two matrices so that

AB=[UT]αγ.\begin{equation*} AB = [UT]_{\alpha}^{\gamma}. \end{equation*}

Consider the matrix [UT]αγ[UT]_{\alpha}^{\gamma}. For 1jn1 \le j \le n, we have

(UT)(vj)=U(T(vj))=U ⁣(k=1mBkjwk)=k=1mBkjU(wk)\begin{equation*} (UT)(v_j) = U(T(v_j)) = U \!\left( \sum_{k=1}^{m} B_{kj} w_k \right) = \sum_{k=1}^{m} B_{kj} \, U(w_k) \end{equation*} =k=1mBkj(i=1pAikzi)=i=1p(k=1mAikBkj)zi\begin{equation*} = \sum_{k=1}^{m} B_{kj} \left( \sum_{i=1}^{p} A_{ik} z_i \right) = \sum_{i=1}^{p} \left( \sum_{k=1}^{m} A_{ik} B_{kj} \right) z_i \end{equation*} =i=1pCijzi,\begin{equation*} = \sum_{i=1}^{p} C_{ij} \, z_i, \end{equation*}

where

Cij=k=1mAikBkj.\begin{equation*} C_{ij} = \sum_{k=1}^{m} A_{ik} B_{kj}. \end{equation*}

This computation motivates the following definition of matrix multiplication.

We define the Kronecker delta δij\delta_{ij} by δij=1\delta_{ij} = 1 if i=ji = j and δij=0\delta_{ij} = 0 if iji \ne j. The n×nn \times n identity matrix InI_n is defined by (In)ij=δij(I_n)_{ij} = \delta_{ij}.

Thus, for example,

I1=(1),I2=(1001),andI3=(100010001).\begin{equation*} I_1 = (1), \qquad I_2 = \begin{pmatrix} 1 & 0 \\ 0 & 1 \end{pmatrix}, \qquad \text{and} \qquad I_3 = \begin{pmatrix} 1 & 0 & 0 \\ 0 & 1 & 0 \\ 0 & 0 & 1 \end{pmatrix}. \end{equation*}

Theorems

Let AA be an m×nm \times n matrix, BB and CC be n×pn \times p matrices, and DD and EE be q×mq \times m matrices. Then

  1. A(B+C)=AB+ACA(B + C) = AB + AC and (D+E)A=DA+EA(D + E)A = DA + EA.

  2. a(AB)=(aA)B=A(aB)a(AB) = (aA)B = A(aB) for any scalar aa.

  3. ImA=A=AInI_m A = A = A I_n.

  4. If VV is an nn-dimensional vector space with an ordered basis β\beta, then [IV]β=In[I_V]_{\beta} = I_n.

Let VV and WW be finite-dimensional vector spaces having ordered bases β\beta and γ\gamma, respectively, and let T:VWT : V \to W be linear. Then, for each uVu \in V, we have

[T(u)]γ=[T]βγ[u]β.\begin{equation*} [T(u)]_{\gamma} = [T]_{\beta}^{\gamma} [u]_{\beta}. \end{equation*}

Left-multiplication transformation

Let AA be an m×nm \times n matrix with entries from a field FF. We denote by LAL_A the mapping LA:FnFmL_A : F^n \to F^m defined by LA(x)=AxL_A(x) = Ax (the matrix product of AA and xx) for each column vector xFnx \in F^n. We call LAL_A a left-multiplication transformation.

Theorems

Let AA be an m×nm \times n matrix with entries from FF. Then the left-multiplication transformation LA:FnFmL_A : F^n \to F^m is linear. Furthermore, if BB is any other m×nm \times n matrix (with entries from FF) and β\beta and γ\gamma are the standard ordered bases for FnF^n and FmF^m, respectively, then we have the following properties.

  1. [LA]βγ=A[L_A]_{\beta}^{\gamma} = A.

  2. LA=LBL_A = L_B if and only if A=BA = B.

  3. LA+B=LA+LBL_{A+B} = L_A + L_B and LaA=aLAL_{aA} = a L_A for all aFa \in F.

  4. If T:FnFmT : F^n \to F^m is linear, then there exists a unique m×nm \times n matrix CC such that T=LCT = L_C. In fact, C=[T]βγC = [T]_{\beta}^{\gamma}.

  5. If EE is an n×pn \times p matrix, then LAE=LALEL_{AE} = L_A L_E.

  6. If m=nm = n, then LIn=IFnL_{I_n} = I_{F^n}.

Invertibility and isomorphisms

Invertibility

Let VV and WW be vector spaces, and let T:VWT : V \to W be linear. A function U:WVU : W \to V is said to be an inverse of TT if TU=IWTU = I_W and UT=IVUT = I_V. If TT has an inverse, then TT is said to be invertible. As noted in Appendix B, if TT is invertible, then the inverse of TT is unique and is denoted by T1T^{-1}.

We often use the fact that a function is invertible if and only if it is both one-to-one and onto.

Let VV and WW be vector spaces, and let T:VWT : V \to W be linear and invertible. Then T1:WVT^{-1} : W \to V is linear.

Let AA be an n×nn \times n matrix. Then AA is invertible if there exists an n×nn \times n matrix BB such that AB=BA=IAB = BA = I.

Theorems

  1. Let TT be an invertible linear transformation from VV to WW. Then VV is finite-dimensional if and only if WW is finite-dimensional. In this case, dim(V)=dim(W)\dim(V) = \dim(W).

  2. Let VV and WW be finite-dimensional vector spaces with ordered bases β\beta and γ\gamma, respectively. Let T:VWT : V \to W be linear. Then TT is invertible if and only if [T]βγ[T]_{\beta}^{\gamma} is invertible. Furthermore,[T1]γβ=([T]βγ)1[T^{-1}]_{\gamma}^{\beta} = ([T]_{\beta}^{\gamma})^{-1}.

Isomorphisms

Let VV and WW be vector spaces. We say that VV is isomorphic to WW if there exists a linear transformation T:VWT : V \to W that is invertible. Such a linear transformation is called an isomorphism from VV onto WW.

Theorems

  1. Let VV and WW be finite-dimensional vector spaces (over the same field). Then VV is isomorphic to WW if and only if dim(V)=dim(W)\dim(V) = \dim(W).

The change of coordinate matrix

Let β\beta and β\beta' be two ordered bases for a finite-dimensional vector space VV, and let Q=[IV]ββQ = [\,I_V\,]_{\beta}^{\beta'}. Then

(a) QQ is invertible.

(b) For any vVv \in V, [v]β=Q[v]β[v]_{\beta} = Q [v]_{\beta'}.

The matrix Q=[IV]ββQ = [\,I_V\,]_{\beta}^{\beta'} defined in Theorem 2.22 is called a change of coordinate matrix.

Let TT be a linear operator on a finite-dimensional vector space VV, and let β\beta and β\beta' be ordered bases for VV. Suppose that QQ is the change of coordinate matrix that changes β\beta'-coordinates into β\beta-coordinates. Then

[T]β=Q1[T]βQ.\begin{equation*} [T]_{\beta'} = Q^{-1} [T]_{\beta} Q. \end{equation*}

Let AA and BB be matrices in Mn×n(F)M_{n \times n}(F). We say that BB is similar to AA if there exists an invertible matrix QQ such that

B=Q1AQ.\begin{equation*} B = Q^{-1} A Q. \end{equation*}

Dual spaces

For a vector space VV over FF, we define the dual space of VV to be the vector space L(V,F)\mathcal{L}(V, F), denoted by VV^{*}.We also define the double dual VV^{**} of VV to be the dual of VV^{*}.

Suppose that VV is a finite-dimensional vector space with the ordered basis β={x1,x2,,xn}\beta = \{x_1, x_2, \dots, x_n\}. Let fif_i (1in1 \le i \le n) be the iith coordinate function with respect to β\beta as just defined, and let β={f1,f2,,fn}\beta^{*} = \{f_1, f_2, \dots, f_n\}. Then β\beta^{*} is an ordered basis for VV^{*}, and, for any fVf \in V^{*}, we have

f=i=1nf(xi)fi.\begin{equation*} f = \sum_{i=1}^{n} f(x_i)\, f_i. \end{equation*}

we call the ordered basis β={f1,f2,,fn}\beta^{*} = \{f_1, f_2, \dots, f_n\} of VV^{*} that satisfies fi(xj)=δijf_i(x_j) = \delta_{ij} (1i,jn1 \le i, j \le n) the dual basis of β\beta.

Transpose

Let VV and WW be finite-dimensional vector spaces over FF with ordered bases β\beta and γ\gamma, respectively. For any linear transformation T:VWT : V \to W, the mapping Tt:WVT^{t} : W^{*} \to V^{*} defined by Tt(g)=gTT^{t}(g) = gT for all gWg \in W^{*} is a linear transformation with the property that

[Tt]γβ=([T]βγ)t.\begin{equation*} [T^{t}]_{\gamma^{*}}^{\beta^{*}} = ([T]_{\beta}^{\gamma})^{t}. \end{equation*}

For a vector xVx \in V, we define x^:VF\widehat{x} : V^{*} \to F by x^(f)=f(x)\widehat{x}(f) = f(x) for every fVf \in V^{*}. It is easy to verify that x^\widehat{x} is a linear functional on VV^{*}, so x^V\widehat{x} \in V^{**}.

Let VV be a finite-dimensional vector space, and let xVx \in V. If x^(f)=0\widehat{x}(f) = 0 for all fVf \in V^{*}, then x=0x = 0.

Theorems

  1. Let VV be a finite-dimensional vector space, and define ψ:VV\psi : V \to V^{**} by ψ(x)=x^\psi(x) = \widehat{x}. Then ψ\psi is an isomorphism.

  2. Let VV be a finite-dimensional vector space with dual space VV^{*}. Then every ordered basis for VV^{*} is the dual basis for some basis for VV.