Two non-zero vectors x and y are said to be orthogonal if xTy=0. A set {v1,v2,…,vm} of vectors is said to be orthogonal if vTjvi=0 for all i≠j.
Vectors {v1,v2,…,vm} are said to be linearly independent if for scalars c1,c2,…,cm
c1v1+c2v2+⋯+cmvm=0if and only if cj=0 for all j.
Let {x1,x2,…,xm} be a set of linearly independent vectors in Rn, m≤n. Then {v1,v2,…,vm} defined by
v1=x1,v2=x2−(vT1x2vT1v1)v1,v2=x3−(vT1x3vT1v1)v1−(vT2x3vT2v2)v2,⋮vk=xk−m−1∑i−1(vTixkvTivi)vi.is an orthogonal set of vectors.
Usually, we then define qi=vi/‖ to get unit vectors. The vectors \{q_i\} form an orthonormal set.
Consider the n \times m matrix
Q = [q_1,q_2, \ldots, q_m],whose columns are the orthonormal vectors q_j. It the follows that
(Q^TQ)_{ij} = q_i^Tq_j = \begin{cases} 1, & i = j,\\ 0, & i \neq j.\end{cases}Or, Q^TQ = I, the m\times m identity matrix.
A matrix Q is said to be orthogonal if its columns form an orthonormal set in \mathbb R^n.
For an n\times n orthogonal matrix, Q^TQ = I, and the following theorem is a direct consequence of this.
If Q is an n \times n orthogonal matrix then
This last point shows that Q preserves l_2 distances.
If \lambda is an eigenvalue of A then Av= \lambda v for a non-zero vector v. If we multiply this equation by S^{-1} we obtain
S^{-1}A v = \lambda S^{-1} v, S^{-1}AS S^{-1} v = \lambda S^{-1} v, B w = \lambda w, ~~ w = S^{-1}v.It follows that w = S^{-1}v \neq 0 and \lambda is an eigenvalue of B. Conversely, if \lambda is an eigenvalue of B, Bw = \lambda w for a non-zero vector w. We multiply this equation by S
SB w = \lambda S w, ASw = \lambda S w, A v = \lambda v, ~~ v = Sw.And so, \lambda is an eigenvalue of A. This shows that the eigenvalues coincide.
An n\times n matrix is similar to a diagonal matrix D if and only if A has n linearly independent eigenvectors. In this case, the matrix S in A = S D S^{-1} has its columns being the eigenvectors of A.
The transformation A \mapsto S A S^{-1} is called a similarity transformation of A.
Recall that a triangular matrix has its eigenvalues on the diagonal. The following theorem is of great importance.
For any matrix A there exits a non-singular matrix U such that
T = U^{-1} A U,where T is an upper-triangular matrix.
Note that U will be, in general, a complex matrix. Define U^* = \bar{U}^T where the \bar U denotes complex conjugation. It a component of Schur's theorem is that U^{-1}= U^* and U is called unitary (the complex analogue of an orthogonal matrix).
Recall that the Spectral Theorem (from Lecture 17) states that T can be chosen to be diagonal.