Hubbry Logo
Diagonalizable matrixDiagonalizable matrixMain
Open search
Diagonalizable matrix
Community hub
Diagonalizable matrix
logo
7 pages, 0 posts
0 subscribers
Be the first to start a discussion here.
Be the first to start a discussion here.
Diagonalizable matrix
Diagonalizable matrix
from Wikipedia

In linear algebra, a square matrix  is called diagonalizable or non-defective if it is similar to a diagonal matrix. That is, if there exists an invertible matrix  and a diagonal matrix such that . This is equivalent to . (Such , are not unique.) This property exists for any linear map: for a finite-dimensional vector space , a linear map  is called diagonalizable if there exists an ordered basis of  consisting of eigenvectors of . These definitions are equivalent: if  has a matrix representation as above, then the column vectors of  form a basis consisting of eigenvectors of , and the diagonal entries of  are the corresponding eigenvalues of ; with respect to this eigenvector basis,  is represented by .

Diagonalization is the process of finding the above  and and makes many subsequent computations easier. One can raise a diagonal matrix  to a power by simply raising the diagonal entries to that power. The determinant of a diagonal matrix is simply the product of all diagonal entries. Such computations generalize easily to .

The geometric transformation represented by a diagonalizable matrix is an inhomogeneous dilation (or anisotropic scaling). That is, it can scale the space by a different amount in different directions. The direction of each eigenvector is scaled by a factor given by the corresponding eigenvalue.

A square matrix that is not diagonalizable is called defective. It can happen that a matrix with real entries is defective over the real numbers, meaning that is impossible for any invertible and diagonal with real entries, but it is possible with complex entries, so that is diagonalizable over the complex numbers. For example, this is the case for a generic rotation matrix.

Many results for diagonalizable matrices hold only over an algebraically closed field (such as the complex numbers). In this case, diagonalizable matrices are dense in the space of all matrices, which means any defective matrix can be deformed into a diagonalizable matrix by a small perturbation; and the Jordan–Chevalley decomposition states that any matrix is uniquely the sum of a diagonalizable matrix and a nilpotent matrix. Over an algebraically closed field, diagonalizable matrices are equivalent to semi-simple matrices.

Definition

[edit]

A square matrix with entries in a field is called diagonalizable or nondefective if there exists an invertible matrix (i.e. an element of the general linear group GLn(F)), , such that is a diagonal matrix.

Characterization

[edit]

The fundamental fact about diagonalizable maps and matrices is expressed by the following:

  • An matrix over a field is diagonalizable if and only if the sum of the dimensions of its eigenspaces is equal to , which is the case if and only if there exists a basis of consisting of eigenvectors of . If such a basis has been found, one can form the matrix having these basis vectors as columns, and will be a diagonal matrix whose diagonal entries are the eigenvalues of . The matrix is known as a modal matrix for .
  • A linear map is diagonalizable if and only if the sum of the dimensions of its eigenspaces is equal to , which is the case if and only if there exists a basis of consisting of eigenvectors of . With respect to such a basis, will be represented by a diagonal matrix. The diagonal entries of this matrix are the eigenvalues of .

The following sufficient (but not necessary) condition is often useful.

  • An matrix is diagonalizable over the field if it has distinct eigenvalues in , i.e. if its characteristic polynomial has distinct roots in ; however, the converse may be false. Consider which has eigenvalues 1, 2, 2 (not all distinct) and is diagonalizable with diagonal form (similar to ) and change of basis matrix : The converse fails when has an eigenspace of dimension higher than 1. In this example, the eigenspace of associated with the eigenvalue 2 has dimension 2.
  • A linear map with is diagonalizable if it has distinct eigenvalues, i.e. if its characteristic polynomial has distinct roots in .

Let be a matrix over . If is diagonalizable, then so is any power of it. Conversely, if is invertible, is algebraically closed, and is diagonalizable for some that is not an integer multiple of the characteristic of , then is diagonalizable. Proof: If is diagonalizable, then is annihilated by some polynomial , which has no multiple root (since ) and is divided by the minimal polynomial of .

Over the complex numbers , almost every matrix is diagonalizable. More precisely: the set of complex matrices that are not diagonalizable over , considered as a subset of , has Lebesgue measure zero. One can also say that the diagonalizable matrices form a dense subset with respect to the Zariski topology: the non-diagonalizable matrices lie inside the vanishing set of the discriminant of the characteristic polynomial, which is a hypersurface. From that follows also density in the usual (strong) topology given by a norm. The same is not true over .

The Jordan–Chevalley decomposition expresses an operator as the sum of its semisimple (i.e., diagonalizable) part and its nilpotent part. Hence, a matrix is diagonalizable if and only if its nilpotent part is zero. Put in another way, a matrix is diagonalizable if each block in its Jordan form has no nilpotent part; i.e., each "block" is a one-by-one matrix.

Diagonalization

[edit]

Consider the two following arbitrary bases and . Suppose that there exists a linear transformation represented by a matrix which is written with respect to basis E. Suppose also that there exists the following eigen-equation:

The alpha eigenvectors are written also with respect to the E basis. Since the set F is both a set of eigenvectors for matrix A and it spans some arbitrary vector space, then we say that there exists a matrix which is a diagonal matrix that is similar to . In other words, is a diagonalizable matrix if the matrix is written in the basis F. We perform the change of basis calculation using the transition matrix , which changes basis from E to F as follows:

,

where is the transition matrix from E-basis to F-basis. The inverse can then be equated to a new transition matrix which changes basis from F to E instead and so we have the following relationship :

Both and transition matrices are invertible. Thus we can manipulate the matrices in the following fashion:The matrix will be denoted as , which is still in the E-basis. Similarly, the diagonal matrix is in the F-basis.

The diagonalization of a symmetric matrix can be interpreted as a rotation of the axes to align them with the eigenvectors.

If a matrix can be diagonalized, that is,

then:

The transition matrix S has the E-basis vectors as columns written in the basis F. Inversely, the inverse transition matrix P has F-basis vectors written in the basis of E so that we can represent P in block matrix form in the following manner:

as a result we can write:

In block matrix form, we can consider the A-matrix to be a matrix of 1x1 dimensions whilst P is a 1xn dimensional matrix. The D-matrix can be written in full form with all the diagonal elements as an nxn dimensional matrix:

Performing the above matrix multiplication we end up with the following result:Taking each component of the block matrix individually on both sides, we end up with the following:

So the column vectors of are right eigenvectors of , and the corresponding diagonal entry is the corresponding eigenvalue. The invertibility of also suggests that the eigenvectors are linearly independent and form a basis of . This is the necessary and sufficient condition for diagonalizability and the canonical approach of diagonalization. The row vectors of are the left eigenvectors of .

When a complex matrix is a Hermitian matrix (or more generally a normal matrix), eigenvectors of can be chosen to form an orthonormal basis of , and can be chosen to be a unitary matrix. If in addition, is a real symmetric matrix, then its eigenvectors can be chosen to be an orthonormal basis of and can be chosen to be an orthogonal matrix.

For most practical work matrices are diagonalized numerically using computer software. Many algorithms exist to accomplish this.

Simultaneous diagonalization

[edit]

A set of matrices is said to be simultaneously diagonalizable if there exists a single invertible matrix such that is a diagonal matrix for every in the set. The following theorem characterizes simultaneously diagonalizable matrices: A set of diagonalizable matrices commutes if and only if the set is simultaneously diagonalizable.[1]: p. 64 

The set of all diagonalizable matrices (over ) with is not simultaneously diagonalizable. For instance, the matrices

are diagonalizable but not simultaneously diagonalizable because they do not commute.

A set consists of commuting normal matrices if and only if it is simultaneously diagonalizable by a unitary matrix; that is, there exists a unitary matrix such that is diagonal for every in the set.

In the language of Lie theory, a set of simultaneously diagonalizable matrices generates a toral Lie algebra.

Examples

[edit]

Diagonalizable matrices

[edit]
  • Involutions are diagonalizable over the reals (and indeed any field of characteristic not 2), with ±1 on the diagonal.
  • Finite order endomorphisms are diagonalizable over (or any algebraically closed field where the characteristic of the field does not divide the order of the endomorphism) with roots of unity on the diagonal. This follows since the minimal polynomial is separable, because the roots of unity are distinct.
  • Projections are diagonalizable, with 0s and 1s on the diagonal.
  • Real symmetric matrices are diagonalizable by orthogonal matrices; i.e., given a real symmetric matrix , is diagonal for some orthogonal matrix . More generally, matrices are diagonalizable by unitary matrices if and only if they are normal. In the case of the real symmetric matrix, we see that , so clearly holds. Examples of normal matrices are real symmetric (or skew-symmetric) matrices (e.g. covariance matrices) and Hermitian matrices (or skew-Hermitian matrices). See spectral theorems for generalizations to infinite-dimensional vector spaces.

Matrices that are not diagonalizable

[edit]

In general, a rotation matrix is not diagonalizable over the reals, but all rotation matrices are diagonalizable over the complex field. Even if a matrix is not diagonalizable, it is always possible to "do the best one can", and find a matrix with the same properties consisting of eigenvalues on the leading diagonal, and either ones or zeroes on the superdiagonal – known as Jordan normal form.

Some matrices are not diagonalizable over any field, most notably nonzero nilpotent matrices. This happens more generally if the algebraic and geometric multiplicities of an eigenvalue do not coincide. For instance, consider

This matrix is not diagonalizable: there is no matrix such that is a diagonal matrix. Indeed, has one eigenvalue (namely zero) and this eigenvalue has algebraic multiplicity 2 and geometric multiplicity 1.

Some real matrices are not diagonalizable over the reals. Consider for instance the matrix

The matrix does not have any real eigenvalues, so there is no real matrix such that is a diagonal matrix. However, we can diagonalize if we allow complex numbers. Indeed, if we take

then is diagonal. It is easy to find that is the rotation matrix which rotates counterclockwise by angle

Note that the above examples show that the sum of diagonalizable matrices need not be diagonalizable.

How to diagonalize a matrix

[edit]

Diagonalizing a matrix is the same process as finding its eigenvalues and eigenvectors, in the case that the eigenvectors form a basis. For example, consider the matrix

The roots of the characteristic polynomial are the eigenvalues . Solving the linear system gives the eigenvectors and , while gives ; that is, for . These vectors form a basis of , so we can assemble them as the column vectors of a change-of-basis matrix to get: We may see this equation in terms of transformations: takes the standard basis to the eigenbasis, , so we have: so that has the standard basis as its eigenvectors, which is the defining property of .

Note that there is no preferred order of the eigenvectors in ; changing the order of the eigenvectors in just changes the order of the eigenvalues in the diagonalized form of .[2]

Application to matrix functions

[edit]

Diagonalization can be used to efficiently compute the powers of a matrix :

and the latter is easy to calculate since it only involves the powers of a diagonal matrix. For example, for the matrix with eigenvalues in the example above we compute:

This approach can be generalized to matrix exponential and other matrix functions that can be defined as power series. For example, defining , we have:

This is particularly useful in finding closed form expressions for terms of linear recursive sequences, such as the Fibonacci numbers.

Particular application

[edit]

For example, consider the following matrix:

Calculating the various powers of reveals a surprising pattern:

The above phenomenon can be explained by diagonalizing . To accomplish this, we need a basis of consisting of eigenvectors of . One such eigenvector basis is given by

where ei denotes the standard basis of Rn. The reverse change of basis is given by

Straightforward calculations show that

Thus, a and b are the eigenvalues corresponding to u and v, respectively. By linearity of matrix multiplication, we have that

Switching back to the standard basis, we have

The preceding relations, expressed in matrix form, are

thereby explaining the above phenomenon.

Quantum mechanical application

[edit]

In quantum mechanical and quantum chemical computations matrix diagonalization is one of the most frequently applied numerical processes. The basic reason is that the time-independent Schrödinger equation is an eigenvalue equation, albeit in most of the physical situations on an infinite dimensional Hilbert space.

A very common approximation is to truncate (or project) the Hilbert space to finite dimension, after which the Schrödinger equation can be formulated as an eigenvalue problem of a real symmetric, or complex Hermitian matrix. Formally this approximation is founded on the variational principle, valid for Hamiltonians that are bounded from below.

First-order perturbation theory also leads to matrix eigenvalue problem for degenerate states.

Operator theory

[edit]

Matrices can be generalized to linear operators. A diagonal matrix can be generalized to diagonal operators on Hilbert spaces.

Let be a Hilbert space. An operator is a diagonal operator iff there exists an orthonormal basis of , such that for some .

For any , define the p-Schatten norm as follows. Let be an operator, then , where is the trace. The p-Schatten class is the set of all operators with finite p-Schatten norm.

Weyl[3], von Neumann[4], and Kuroda[5], showed the following:

For any , any self-adjoint operator on a Hilbert space , and any , there exists a diagonal operator , such that .

In other words, any self-adjoint operator is an infinitesimal perturbation from a diagonal operator, where "infinitesimal" is in the sense of p-Schatten norm. In particular, since the Hilbert–Schmidt operator class is the 2-Schatten class, this means that any self-adjoint operator is diagonalizable after a perturbation by an infinitesimal Hilbert–Schmidt operator. In fact, the above result could be further generalized:

For any norm ideal that is not the trace class, with norm , any self-adjoint operator on a Hilbert space , and any , there exists a diagonal operator , such that .

The result is false for (the trace class). This is a simple corollary of the Kato[6]–Rosenblum[7][8]: Theorem XI.8  theorem, which states that if is self-adjoint, and is trace class, then have the same absolutely continuous part of the spectrum. The result is sharp, however, in the sense that if has no absolutely continuous part, then it can be diagonalized after perturbation by an infinitesimal trace class operator.[9]

For simultaneous diagonalization, it's known that, given a finite list of self-adjoint operators that commute with each other, for any , there exists a sequence of diagonal operators , such that , where is the n-Schatten norm. Note that [10]

See also

[edit]

References

[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
In linear algebra, a diagonalizable matrix is a square matrix AA that is similar to a , meaning there exists an PP such that P1AP=DP^{-1} A P = D, where DD is a whose entries are the eigenvalues of AA. This similarity transformation diagonalizes AA, effectively representing it in a basis of its eigenvectors, with the diagonal entries of DD corresponding to the scaling factors along those basis directions. A matrix ARn×nA \in \mathbb{R}^{n \times n} (or over the complex numbers) is diagonalizable it has a full set of nn linearly independent eigenvectors, which occurs when the geometric multiplicity equals the algebraic multiplicity for each eigenvalue. Matrices with distinct eigenvalues are always diagonalizable, as each eigenvalue has algebraic multiplicity one and thus a full eigenspace . Not all matrices are diagonalizable; for example, certain blocks with repeated eigenvalues and deficient eigenspaces are not. Diagonalization is crucial for simplifying matrix computations, such as raising a matrix to a power Ak=PDkP1A^k = P D^k P^{-1}, where DkD^k is easily computed by raising each diagonal entry to the kk-th power. It also facilitates the computation of matrix exponentials eA=PeDP1e^A = P e^D P^{-1}, which are essential in solving systems of linear differential equations and modeling continuous-time dynamical systems. In applications like Markov chains, , and , diagonalizable matrices enable efficient spectral decomposition and stability analysis.

Fundamentals

Definition

In linear algebra, a square matrix AA of size n×nn \times n over a field FF is diagonalizable if there exists an PP (also of size n×nn \times n) and a DD such that A=PDP1,A = P D P^{-1}, where the diagonal entries of DD are the eigenvalues of AA. This relation expresses AA as similar to a via a given by the columns of PP. The similarity transformation P1AP=DP^{-1} A P = D preserves key spectral properties of AA, including its eigenvalues, , trace, and , as these are invariant under similarity. For matrices with real entries, the field FF is typically extended to the of complex numbers to guarantee the existence of all eigenvalues, even if they are non-real. The origins of the diagonalization concept trace back to Joseph-Louis Lagrange's 18th-century investigations of quadratic forms, where he employed linear transformations to reduce them to diagonal form, and were formalized within modern linear algebra by David Hilbert's work on around 1900.

Characterization

A square matrix AA over a field FF is diagonalizable if and only if there exists a basis of the underlying consisting of nn linearly independent eigenvectors of AA, where nn is the of the . This condition ensures that AA can be represented by a in some basis, as the eigenvectors form the columns of the PP in the similarity transformation A=PDP1A = P D P^{-1}, where DD is diagonal. An equivalent characterization involves the multiplicities of the eigenvalues of AA. For each eigenvalue λ\lambda of AA, the geometric multiplicity, defined as dim(ker(AλI))\dim(\ker(A - \lambda I)), must equal the algebraic multiplicity, which is the multiplicity of λ\lambda as a root of the det(AλI)\det(A - \lambda I). This equality holds across all eigenvalues if and only if the sum of the geometric multiplicities is nn, guaranteeing a full basis of eigenvectors. Another criterion uses the minimal polynomial of AA, the monic polynomial of least degree that annihilates AA. The matrix AA is diagonalizable over FF if and only if its minimal polynomial factors into distinct linear factors over FF, meaning it has no repeated roots. This condition implies that the minimal polynomial splits completely into linear terms without multiplicity greater than one. Over an algebraically closed field such as C\mathbb{C}, the characteristic polynomial of any matrix always splits into linear factors by the fundamental theorem of algebra. In this setting, diagonalizability reduces to the minimal polynomial having distinct linear factors or, equivalently, the algebraic and geometric multiplicities matching for each eigenvalue, as the splitting is automatic. For matrices over general fields FF, diagonalizability requires both that the characteristic polynomial splits into linear factors over FF and that the geometric multiplicity equals the algebraic multiplicity for each root.

Diagonalization Techniques

Diagonalization Procedure

To diagonalize a square matrix ARn×nA \in \mathbb{R}^{n \times n} (or over C\mathbb{C}), the procedure involves computing its to determine if a basis of nn linearly independent eigenvectors exists, enabling the A=PDP1A = PDP^{-1} where DD is diagonal and PP is invertible./07%3A_Spectral_Theory/7.02%3A_Diagonalization) The first step is to find the eigenvalues by solving the characteristic equation det(AλI)=0\det(A - \lambda I) = 0, where II is the ; the roots λ1,,λk\lambda_1, \dots, \lambda_k (with possible multiplicities) are the eigenvalues of AA. For each distinct eigenvalue λi\lambda_i, compute the corresponding eigenspace by solving the eigenvector equation (AλiI)v=0(A - \lambda_i I)v = 0 to find a basis for the null space; the dimension of this eigenspace is the geometric multiplicity of λi\lambda_i./07%3A_Spectral_Theory/7.02%3A_Diagonalization) The matrix AA is diagonalizable the geometric multiplicity equals the algebraic multiplicity (from the ) for every eigenvalue, ensuring the eigenspaces collectively span Rn\mathbb{R}^n (or Cn\mathbb{C}^n) with nn linearly independent eigenvectors. If nn linearly independent eigenvectors v1,,vnv_1, \dots, v_n are obtained, form the matrix PP with these as columns and the D=diag(λ1,,λn)D = \operatorname{diag}(\lambda_1, \dots, \lambda_n); the diagonalization is then A=PDP1A = PDP^{-1}, which can be verified by direct computation./07%3A_Spectral_Theory/7.02%3A_Diagonalization) If the geometric multiplicity is less than the algebraic multiplicity for any eigenvalue, the matrix is defective and not diagonalizable over the field; in such cases, the Jordan canonical form provides an alternative block-diagonal representation using generalized eigenvectors, though it requires additional computational steps beyond standard eigendecomposition. For large matrices, numerical implementations in software such as MATLAB's eig function or Python's SciPy linalg.eig are essential, as they employ algorithms like the QR method for eigenvalue computation. However, these are sensitive to floating-point errors, particularly for matrices with clustered or nearly degenerate eigenvalues, where small perturbations can lead to inaccurate eigenvectors or failure to detect linear independence.

Simultaneous Diagonalization

A set of matrices {A1,,Ak}\{A_1, \dots, A_k\} is said to be simultaneously diagonalizable if there exists a single PP such that P1AiPP^{-1} A_i P is a for each i=1,,ki = 1, \dots, k. This extends the of diagonalization from individual matrices to families, requiring a common basis of eigenvectors. A fundamental result states that if a set of diagonalizable matrices commute pairwise (i.e., [Ai,Aj]=AiAjAjAi=0[A_i, A_j] = A_i A_j - A_j A_i = 0 for all i,ji, j), then they are simultaneously diagonalizable; conversely, if they are simultaneously diagonalizable, they . More generally, simultaneous diagonalizability holds if the matrices share a common basis of eigenvectors. For two commuting diagonalizable matrices AA and BB, there exists an invertible PP such that P1AP=D,P1BP=E,P^{-1} A P = D, \quad P^{-1} B P = E, where DD and EE are diagonal. A proof sketch proceeds by induction on the . For the base case, assume AA has a distinct eigenvalue λ\lambda with eigenspace VλV_\lambda. Since BB commutes with AA, BB preserves VλV_\lambda, so VλV_\lambda decomposes into generalized eigenspaces of BB. By diagonalizability of BB, these are eigenspaces, allowing simultaneous diagonalization on VλV_\lambda. The result extends to the full space by induction. One key application arises in the of quadratic forms, where a set of symmetric matrices can be simultaneously diagonalized via congruence (i.e., PTAiPP^T A_i P diagonal for nonsingular PP) if they commute, facilitating the reduction of multiple quadratic forms to . For non-commuting matrices, simultaneous diagonalization is not generally possible unless they share common eigenspaces, as commutativity ensures the preservation of eigenspaces under conjugation.

Examples

Diagonalizable Matrices

Diagonal matrices provide the simplest example of diagonalizable matrices. A diagonal matrix DD is already in diagonal form, so it is trivially diagonalizable with the identity matrix P=IP = I as the change-of-basis matrix, satisfying D=PDP1D = P D P^{-1}./07:_Spectral_Theory/7.02:_Diagonalization) Real symmetric matrices are always diagonalizable over the real numbers, as guaranteed by the spectral theorem. This theorem states that every real symmetric matrix has real eigenvalues and can be diagonalized using an orthogonal matrix PP, where PT=P1P^T = P^{-1}, yielding A=PDPTA = P D P^T with DD diagonal containing the eigenvalues. For instance, the symmetric matrix (2112)\begin{pmatrix} 2 & 1 \\ 1 & 2 \end{pmatrix} has eigenvalues 3 and 1, with corresponding orthonormal eigenvectors 12(11)\frac{1}{\sqrt{2}} \begin{pmatrix} 1 \\ 1 \end{pmatrix}
Add your contribution
Related Hubs
User Avatar
No comments yet.