Diagonalizable matrices •
Involutions are diagonalizable over the reals (and indeed any field of characteristic not 2), with ±1 on the diagonal. • Finite order
endomorphisms are diagonalizable over \mathbb{C} (or any algebraically closed field where the characteristic of the field does not divide the order of the endomorphism) with
roots of unity on the diagonal. This follows since the minimal polynomial is
separable, because the roots of unity are distinct. •
Projections are diagonalizable, with 0s and 1s on the diagonal. • Real
symmetric matrices are diagonalizable by
orthogonal matrices; i.e., given a real symmetric matrix Q^{\mathrm T}AQ is diagonal for some orthogonal matrix More generally, matrices are diagonalizable by
unitary matrices if and only if they are
normal. In the case of the real symmetric matrix, we see that {{nowrap|A=A^{\mathrm T},}} so clearly AA^{\mathrm T} = A^{\mathrm T}A holds. Examples of normal matrices are real symmetric (or
skew-symmetric) matrices (e.g. covariance matrices) and
Hermitian matrices (or skew-Hermitian matrices). See
spectral theorems for generalizations to infinite-dimensional vector spaces.
Matrices that are not diagonalizable In general, a
rotation matrix is not diagonalizable over the reals, but all
rotation matrices are diagonalizable over the complex field. Even if a matrix is not diagonalizable, it is always possible to "do the best one can", and find a matrix with the same properties consisting of eigenvalues on the leading diagonal, and either ones or zeroes on the superdiagonal – known as
Jordan normal form. Some matrices are not diagonalizable over any field, most notably nonzero
nilpotent matrices. This happens more generally if the
algebraic and geometric multiplicities of an eigenvalue do not coincide. For instance, consider : C = \begin{bmatrix} 0 & 1 \\ 0 & 0 \end{bmatrix}. This matrix is not diagonalizable: there is no matrix U such that U^{-1}CU is a diagonal matrix. Indeed, C has one eigenvalue (namely zero) and this eigenvalue has algebraic multiplicity 2 and geometric multiplicity 1. Some real matrices are not diagonalizable over the reals. Consider for instance the matrix : B = \left[\begin{array}{rr} 0 & 1 \\ \!-1 & 0 \end{array}\right]. The matrix B does not have any real eigenvalues, so there is no
real matrix Q such that Q^{-1}BQ is a diagonal matrix. However, we can diagonalize B if we allow complex numbers. Indeed, if we take : Q = \begin{bmatrix} 1 & i \\ i & 1 \end{bmatrix}, then Q^{-1}BQ is diagonal. It is easy to find that B is the rotation matrix which rotates counterclockwise by angle \theta = -\frac{\pi}{2} Note that the above examples show that the sum of diagonalizable matrices need not be diagonalizable.
How to diagonalize a matrix Diagonalizing a matrix is the same process as finding its
eigenvalues and eigenvectors, in the case that the eigenvectors form a basis. For example, consider the matrix :A=\left[\begin{array}{rrr} 0 & 1 & \!\!\!-2\\ 0 & 1 & 0\\ 1 & \!\!\!-1 & 3 \end{array}\right]. The roots of the
characteristic polynomial p(\lambda)=\det(\lambda I-A) are the eigenvalues Solving the linear system \left(1I-A\right) \mathbf{v} = \mathbf{0} gives the eigenvectors \mathbf{v}_1 = (1,1,0) and {{nowrap|\mathbf{v}_2 = (0,2,1),}} while \left(2I-A\right)\mathbf{v} = \mathbf{0} gives {{nowrap|\mathbf{v}_3 = (1,0,-1);}} that is, A \mathbf{v}_i = \lambda_i \mathbf{v}_i for These vectors form a basis of {{nowrap|V = \mathbb{R}^3,}} so we can assemble them as the column vectors of a
change-of-basis matrix P to get: P^{-1}AP = \left[\begin{array}{rrr} 1 & 0 & 1\\ 1 & 2 & 0\\ 0 & 1 & \!\!\!\!-1 \end{array}\right]^{-1} \left[\begin{array}{rrr} 0 & 1 & \!\!\!-2\\ 0 & 1 & 0\\ 1 & \!\!\!-1 & 3 \end{array}\right] \left[\begin{array}{rrr} 1 & \,0 & 1\\ 1 & 2 & 0\\ 0 & 1 & \!\!\!\!-1 \end{array}\right] = \begin{bmatrix} 1 & 0 & 0 \\ 0 & 1 & 0 \\ 0 & 0 & 2 \end{bmatrix} = D . We may see this equation in terms of transformations: P takes the standard basis to the eigenbasis, {{nowrap|P \mathbf{e}_i = \mathbf{v}_i,}} so we have: P^{-1} AP \mathbf{e}_i = P^{-1} A \mathbf{v}_i = P^{-1} (\lambda_i\mathbf{v}_i) = \lambda_i\mathbf{e}_i, so that P^{-1} AP has the standard basis as its eigenvectors, which is the defining property of Note that there is no preferred order of the eigenvectors in changing the order of the
eigenvectors in P just changes the order of the
eigenvalues in the diagonalized form of == Application to matrix functions ==