7D Isometries, Unitary Operators, and QR Factorization


Topics

  • Isometries
  • Unitary Operators
  • QR Factorization


Isometries


Content

  • Definition: Isometry
  • Theorem: Charicterizations of Isometries

Definition: Isometry

A linear map L(V,W)\mathscr{L}(V,W) is an isometry if S(v)=v\Vert S(v) \Vert = \Vert v \Vert for every vVv \in V. In words, SS preserves the norm of VV, or equivalently distances in VV remain the same in WW through SS.

Theorem: Charicterizations of Isometries

Let SL(V,W)S \in \mathscr{L}(V,W), (e1,,en)(e_1, \ldots, e_n) an orthonormal basis for VV and (f1,,fm)(f_1, \ldots, f_m) an orthonormal basis for WW. Then the following are equivalent: a) SS is an isometry. b) SS=I.S^*S = I. c) S(u),S(v)=u,v\langle S(u), S(v) \rangle = \langle u, v \rangle for every u,vVu, v \in V. d) (S(e1),,S(en))(S(e_1), \ldots, S(e_n)) is an orthonormal list in WW. e) The columns of M(S,(e1,,en),(f1,,fm))\mathcal{M}(S, (e_1, \ldots, e_n), (f_1, \ldots, f_m)) form an orthonormal list of vectors in Fm.\mathbb{F}^m.

Proof:

a) \Rightarrow b): We will use the fact that ISSI - S^*S is self-adjoint to show that it is the zero operator as follows:

(ISS)(v),v=v,vSS(v),v \langle (I-S^*S)(v), v \rangle = \langle v, v \rangle - \langle S^*S(v), v \rangle

=v2S(v),S(v)=v2v2=0. = \Vert v \Vert^2 - \langle S(v), S(v) \rangle = \Vert v \Vert^2 - \Vert v \Vert^2 = 0.

Therefore, ISSI - S^*S must be the zero operator. b) \Rightarrow c):

S(u),S(v)=SS(u),v=u,v. \langle S(u), S(v) \rangle = \langle S^*S(u), v \rangle = \langle u, v \rangle.

c) \Rightarrow d): Because the inner product does not change after applying SS, (S(e1),,S(en))(S(e_1), \ldots, S(e_n)) must also be orthonormal. d) \Rightarrow e): This follows from the last result because the coordinates of the columns are from an orthonormal basis. So the inner product becomes the standard inner product on Fm.\mathbb{F}^m. e) \Rightarrow a): Because we are working with an orthonormal basis for WW, the fact that the columns are orthonormal mean the images of e1e_1 through ene_n form an orthonormal list in WW. Now, because the inner product is biliniar, we only needed to show that this property held for the basis vectors, so we're done.



Unitary Operators


Content

  • Definition: Unitary Operator
  • Theorem: Characterizations of Unitary Operators
  • Lemma: Eigenvalues of Unitary Operators have Absolute Value 11.
  • Theorem: Description of Unitary Operators on Complex Inner Product Spaces

Definition: Unitary Operator

An operator SL(V)S \in \mathscr{L}(V) is unitary if it is an invertible isometry.

Theorem: Characterizations of Unitary Operators

Suppose SL(V)S \in \mathscr{L}(V). Suppose (e1,,en)(e_1,\ldots, e_n) is an orthonormal basis of VV. Then the following are equivalent: a) SS is a unitary operator. b) SS=SS=IS^*S = SS^* = I c) SS is invertible with inverse SS^*. d) (S(e1),,S(en))(S(e_1), \ldots, S(e_n)) is an orthonormal basis for VV. e) The rows of M(S,(e1,,en))\mathcal{M}(S, (e_1, \ldots, e_n)) form an orthonormal basis for Fn\mathbb{F}^n with respect to the standard inner product on Fn.\mathbb{F}^n. f) SS^* is a unitary operator.

Proof:

a) \Rightarrow b): Since SS is unitary, it is invertible and an isometry (and hence SS=IS^*S = I). So

S=SSS1=S1. S^* = S^* S S^{-1} = S^{-1}.

Therefore SS=SS1=I.SS^* = SS^{-1} = I. b) \Rightarrow c): Because ker(S)ker(SS)=ker(I)={0}\ker(S) \subseteq \ker(S^*S) = \ker(I) = \{ 0 \}, SS must be injective. And because it is an operator, SS must therefore be invertible. Furthermore,

S=SSS1=S1. S^* = S^* S S^{-1} = S^{-1}.

c) \Rightarrow d): We know from the last result in the previous section that SS=IS^*S = I is equivalent to (S(e1),,S(en))\left( S(e_1), \ldots, S(e_n) \right) being an orthonormal basis (parts b and d in particular). So this just follows from that equivalence. d) \Rightarrow e): The main idea behind this proof is that the adjoint is the conjugate transpose and thus has columns that are conjugates of an orthonormal basis, which also form an orthonormal basis. The full proof is longer: Suppose (S(e1),,S(en))(S(e_1), \ldots, S(e_n)) is an orthonormal basis for VV. Then from the last result of the previous section, this is equivalent to SS being an isometry. Because SS is an injective operator on a finite-dimensional vector space, SS must be invertible, so SS is a unitary operator (an invertible isometry). Now, from part e) of the last result in the previous section, this means that the columns of our matrix with respect to this basis must form an orthonormal basis for Fn\mathbb{F}^n. This finally means that the rows also form an orthonormal basis, since they are just the conjugates of the matrix of the adjoint, which means they are also an orthonormal basis. e) \Rightarrow f): Suppose the rows of M(T)\mathcal{M}(T) form an orthonormal basis for Fn\mathbb{F}^n. Then the columns of M(S)\mathcal{M}(S) also form an orthonormal basis for Fn\mathbb{F}^n and thus SS is an isometry. So SS and SS^* are inverses of each other. f) \Rightarrow a): Suppose SS^* is unitary and apply all of the previous implications to SS^*, showing that (S)(S^*)^* is also unitary. And since (S)=S(S^*)^*=S, we are done.

Lemma: Eigenvalues of Unitary Operators have Absolute Value 1.

Suppose λ\lambda is an eigenvalue of a unitary operator. Then λ=1| \lambda | = 1.

Proof:

This follows from teh preservation of the inner product, and hence the norm: If vVv \in V is an eigenvector with eigenvalue λF\lambda \in \mathbb{F}, then

v=S(v)=λv=λ v \Vert v \Vert = \Vert S(v) \Vert = \Vert \lambda v \Vert = \vert \lambda \vert \ \Vert v \Vert

λ=1. \Rightarrow | \lambda | = 1.

Theorem: Description of Unitary Operators on Complex Inner Product Spaces

Suppose F=C\mathbb{F} = \mathbb{C} and SL(V)S \in \mathscr{L}(V). Then the following are equivalent: a) SS is a unitary operator. b) There is an orthonormal basis of VV consisting of eigenvectors of SS whose corresponding eigenvalues all have absolute value 1.

Proof:

\Rightarrow: If SS is unitary then it is normal. So by the complex spectral theorem, it is diagonalizable with respect to an orthonormal basis. We also know that its eigenvalues must have an absolute value of 1, so we're done. \Leftarrow: Let (e1,,en)(e_1, \ldots, e_n) be an orthonormal basis of eigenvectors of SS with eigenvalues with absolute value 1. Then the SS-images of these vectors are also orthonormal, as

S(ei),S(ej)=λiλjei,ej. \langle S(e_i), S(e_j) \rangle = \lambda_i \overline{\lambda}_j \langle e_i, e_j \rangle.

Now, this means that SS is unitary from the first result of this chapter.



QR Factorization


Content

  • Definition: Unitary Matrix
  • Lemma: Characterizations of Unitary Matrices
  • Lemma: QR Factorization

Definition: Unitary Matrix

A square matrix is unitary if its columns form an orthonormal basis in Fn.\mathbb{F}^n.

Lemma: Characterizations of Unitary Matrices

Suppose QQ is an n×nn \times n matrix. Then the following are equivalent. a) QQ is a unitary matrix. b) The rows of QQ form tan orthonormal list in Fn\mathbb{F}^n. c) Q(v)=v\Vert Q(v) \Vert = \Vert v \Vert for every vFn.v \in \mathbb{F}^n. d) QQ=QQ=IQ^*Q = QQ^* = I, where II represents the identity matrix.

Proof:

(Proof idea) The matrix QQ is a linear operator from Fn\mathbb{F}^n to Fn\mathbb{F}^n. All of these conditions have been proven to be equivalent for operators.

Theorem: QR Factorization

Suppose AA is a square matrix with linearly independent columns. Then there exist unique matrices QQ and RR, where QQ is unitary, RR is upper-triangular, with only positive numbers on its diagonal, and A=QRA = QR.

Proof:

Apply the Gram-Schmidt procedure to the columns of AA. QQ's columns consist of these orthonormal basis vectors. Now, RR is the right-inverse of the matrix whose columns would apply the linear combinations resulting in the Gram-Schmidt procedure on AA to result in QQ. Specifically, each entry of RR can be found as Rj,k=vk,ej,R_{j,k} = \langle v_k, e_j \rangle, where vkv_k is the kkth column of AA and eje_j is the jjth column of QQ. Then QQ is unitary because its columns are orthonormal. Furthermore, RR is upper-triangular because all of the changes in the Gram-Schmidt procedure add scalar multiples of previous vectors only, making values below the diagonal all zero. Furthermore, the diagonals are positive because the last step is dividing by the norm, which is positive.