Recent from talks
Contribute something
Nothing was collected or created yet.
Triangular matrix
View on WikipediaIn mathematics, a triangular matrix is a special kind of square matrix. A square matrix is called lower triangular if all the entries above the main diagonal are zero. Similarly, a square matrix is called upper triangular if all the entries below the main diagonal are zero.
Because matrix equations with triangular matrices are easier to solve, they are very important in numerical analysis. By the LU decomposition algorithm, an invertible matrix may be written as the product of a lower triangular matrix L and an upper triangular matrix U if and only if all its leading principal minors are non-zero.
Description
[edit]A matrix of the form
is called a lower triangular matrix or left triangular matrix, and analogously a matrix of the form
is called an upper triangular matrix or right triangular matrix. A lower or left triangular matrix is commonly denoted with the variable L, and an upper or right triangular matrix is commonly denoted with the variable U or R.
A matrix that is both upper and lower triangular is diagonal. Matrices that are similar to triangular matrices are called triangularisable.
A non-square (or sometimes any) matrix with zeros above (below) the diagonal is called a lower (upper) trapezoidal matrix. The non-zero entries form the shape of a trapezoid.
Examples
[edit]The matrix
is the lower triangular for the non symmetric matrix:
and
is the upper triangular for the non symmetric matrix:
Forward and back substitution
[edit]A matrix equation in the form or is very easy to solve by an iterative process called forward substitution for lower triangular matrices and analogously back substitution for upper triangular matrices. The process is so called because for lower triangular matrices, one first computes , then substitutes that forward into the next equation to solve for , and repeats through to . In an upper triangular matrix, one works backwards, first computing , then substituting that back into the previous equation to solve for , and repeating through .
Notice that this does not require inverting the matrix.
Forward substitution
[edit]The matrix equation Lx = b can be written as a system of linear equations
Observe that the first equation () only involves , and thus one can solve for directly. The second equation only involves and , and thus can be solved once one substitutes in the already solved value for . Continuing in this way, the -th equation only involves , and one can solve for using the previously solved values for . The resulting formulas are:
A matrix equation with an upper triangular matrix U can be solved in an analogous way, only working backwards.
Applications
[edit]Forward substitution is used in financial bootstrapping to construct a yield curve.
Properties
[edit]The transpose of an upper triangular matrix is a lower triangular matrix and vice versa.
A matrix which is both symmetric and triangular is diagonal. In a similar vein, a matrix which is both normal (meaning A*A = AA*, where A* is the conjugate transpose) and triangular is also diagonal. This can be seen by looking at the diagonal entries of A*A and AA*.
The determinant and permanent of a triangular matrix equal the product of the diagonal entries, as can be checked by direct computation.
In fact more is true: the eigenvalues of a triangular matrix are exactly its diagonal entries. Moreover, each eigenvalue occurs exactly k times on the diagonal, where k is its algebraic multiplicity, that is, its multiplicity as a root of the characteristic polynomial of A. In other words, the characteristic polynomial of a triangular n×n matrix A is exactly
- ,
that is, the unique degree n polynomial whose roots are the diagonal entries of A (with multiplicities). To see this, observe that is also triangular and hence its determinant is the product of its diagonal entries .[1]
Special forms
[edit]Unitriangular matrix
[edit]If the entries on the main diagonal of a (lower or upper) triangular matrix are all 1, the matrix is called (lower or upper) unitriangular.
Other names used for these matrices are unit (lower or upper) triangular, or very rarely normed (lower or upper) triangular. However, a unit triangular matrix is not the same as the unit matrix, and a normed triangular matrix has nothing to do with the notion of matrix norm.
All finite unitriangular matrices are unipotent.
Strictly triangular matrix
[edit]If all of the entries on the main diagonal of a (lower or upper) triangular matrix are also 0, the matrix is called strictly (lower or upper) triangular.
All finite strictly triangular matrices are nilpotent of index at most n as a consequence of the Cayley-Hamilton theorem.
Atomic triangular matrix
[edit]An atomic (lower or upper) triangular matrix is a special form of unitriangular matrix, where all of the off-diagonal elements are zero, except for the entries in a single column. Such a matrix is also called a Frobenius matrix, a Gauss matrix, or a Gauss transformation matrix.
Block triangular matrix
[edit]A block triangular matrix is a block matrix (partitioned matrix) that is a triangular matrix.
Upper block triangular
[edit]A matrix is upper block triangular if
- ,
where for all .[2]
Lower block triangular
[edit]A matrix is lower block triangular if
- ,
where for all .[2]
Triangularisability
[edit]A matrix that is similar to a triangular matrix is referred to as triangularizable. Abstractly, this is equivalent to stabilizing a flag: upper triangular matrices are precisely those that preserve the standard flag, which is given by the standard ordered basis and the resulting flag All flags are conjugate (as the general linear group acts transitively on bases), so any matrix that stabilises a flag is similar to one that stabilizes the standard flag.
Any complex square matrix is triangularizable.[1] In fact, a matrix A over a field containing all of the eigenvalues of A (for example, any matrix over an algebraically closed field) is similar to a triangular matrix. This can be proven by using induction on the fact that A has an eigenvector, by taking the quotient space by the eigenvector and inducting to show that A stabilizes a flag, and is thus triangularizable with respect to a basis for that flag.
A more precise statement is given by the Jordan normal form theorem, which states that in this situation, A is similar to an upper triangular matrix of a very particular form. The simpler triangularization result is often sufficient however, and in any case used in proving the Jordan normal form theorem.[1][3]
In the case of complex matrices, it is possible to say more about triangularization, namely, that any square matrix A has a Schur decomposition. This means that A is unitarily equivalent (i.e. similar, using a unitary matrix as change of basis) to an upper triangular matrix; this follows by taking an Hermitian basis for the flag.
Simultaneous triangularisability
[edit]A set of matrices are said to be simultaneously triangularisable if there is a basis under which they are all upper triangular; equivalently, if they are upper triangularizable by a single similarity matrix P. Such a set of matrices is more easily understood by considering the algebra of matrices it generates, namely all polynomials in the denoted Simultaneous triangularizability means that this algebra is conjugate into the Lie subalgebra of upper triangular matrices, and is equivalent to this algebra being a Lie subalgebra of a Borel subalgebra.
The basic result is that (over an algebraically closed field), the commuting matrices or more generally are simultaneously triangularizable. This can be proven by first showing that commuting matrices have a common eigenvector, and then inducting on dimension as before. This was proven by Frobenius, starting in 1878 for a commuting pair, as discussed at commuting matrices. As for a single matrix, over the complex numbers these can be triangularized by unitary matrices.
The fact that commuting matrices have a common eigenvector can be interpreted as a result of Hilbert's Nullstellensatz: commuting matrices form a commutative algebra over which can be interpreted as a variety in k-dimensional affine space, and the existence of a (common) eigenvalue (and hence a common eigenvector) corresponds to this variety having a point (being non-empty), which is the content of the (weak) Nullstellensatz.[citation needed] In algebraic terms, these operators correspond to an algebra representation of the polynomial algebra in k variables.
This is generalized by Lie's theorem, which shows that any representation of a solvable Lie algebra is simultaneously upper triangularizable, the case of commuting matrices being the abelian Lie algebra case, abelian being a fortiori solvable.
More generally and precisely, a set of matrices is simultaneously triangularisable if and only if the matrix is nilpotent for all polynomials p in k non-commuting variables, where is the commutator; for commuting the commutator vanishes so this holds. This was proven by Drazin, Dungey, and Gruenberg in 1951;[4] a brief proof is given by Prasolov in 1994.[5] One direction is clear: if the matrices are simultaneously triangularisable, then is strictly upper triangularizable (hence nilpotent), which is preserved by multiplication by any or combination thereof – it will still have 0s on the diagonal in the triangularizing basis.
Algebras of triangular matrices
[edit]
Upper triangularity is preserved by many operations:
- The sum of two upper triangular matrices is upper triangular.
- The product of two upper triangular matrices is upper triangular.
- The inverse of an upper triangular matrix, if it exists, is upper triangular.
- The product of an upper triangular matrix and a scalar is upper triangular.
Together these facts mean that the upper triangular matrices form a subalgebra of the associative algebra of square matrices for a given size. Additionally, this also shows that the upper triangular matrices can be viewed as a Lie subalgebra of the Lie algebra of square matrices of a fixed size, where the Lie bracket [a, b] given by the commutator ab − ba. The Lie algebra of all upper triangular matrices is a solvable Lie algebra. It is often referred to as a Borel subalgebra of the Lie algebra of all square matrices.
All these results hold if upper triangular is replaced by lower triangular throughout; in particular the lower triangular matrices also form a Lie algebra. However, operations mixing upper and lower triangular matrices do not in general produce triangular matrices. For instance, the sum of an upper and a lower triangular matrix can be any matrix; the product of a lower triangular with an upper triangular matrix is not necessarily triangular either.
The set of unitriangular matrices forms a Lie group.
The set of strictly upper (or lower) triangular matrices forms a nilpotent Lie algebra, denoted This algebra is the derived Lie algebra of , the Lie algebra of all upper triangular matrices; in symbols, In addition, is the Lie algebra of the Lie group of unitriangular matrices.
In fact, by Engel's theorem, any finite-dimensional nilpotent Lie algebra is conjugate to a subalgebra of the strictly upper triangular matrices, that is to say, a finite-dimensional nilpotent Lie algebra is simultaneously strictly upper triangularizable.
Algebras of upper triangular matrices have a natural generalization in functional analysis which yields nest algebras on Hilbert spaces.
Borel subgroups and Borel subalgebras
[edit]The set of invertible triangular matrices of a given kind (lower or upper) forms a group, indeed a Lie group, which is a subgroup of the general linear group of all invertible matrices. A triangular matrix is invertible precisely when its diagonal entries are invertible (non-zero).
Over the real numbers, this group is disconnected, having components accordingly as each diagonal entry is positive or negative. The identity component is invertible triangular matrices with positive entries on the diagonal, and the group of all invertible triangular matrices is a semidirect product of this group and the group of diagonal matrices with on the diagonal, corresponding to the components.
The Lie algebra of the Lie group of invertible upper triangular matrices is the set of all upper triangular matrices, not necessarily invertible, and is a solvable Lie algebra. These are, respectively, the standard Borel subgroup B of the Lie group GLn and the standard Borel subalgebra of the Lie algebra gln.
The upper triangular matrices are precisely those that stabilize the standard flag. The invertible ones among them form a subgroup of the general linear group, whose conjugate subgroups are those defined as the stabilizer of some (other) complete flag. These subgroups are Borel subgroups. The group of invertible lower triangular matrices is such a subgroup, since it is the stabilizer of the standard flag associated to the standard basis in reverse order.
The stabilizer of a partial flag obtained by forgetting some parts of the standard flag can be described as a set of block upper triangular matrices (but its elements are not all triangular matrices). The conjugates of such a group are the subgroups defined as the stabilizer of some partial flag. These subgroups are called parabolic subgroups.
Examples
[edit]The group of 2×2 upper unitriangular matrices is isomorphic to the additive group of the field of scalars; in the case of complex numbers it corresponds to a group formed of parabolic Möbius transformations; the 3×3 upper unitriangular matrices form the Heisenberg group.
See also
[edit]References
[edit]- ^ a b c Axler, Sheldon Jay (1997). Linear Algebra Done Right (2nd ed.). New York: Springer. pp. 86–87, 169. ISBN 0-387-22595-1. OCLC 54850562.
- ^ a b Bernstein, Dennis S. (2009). Matrix mathematics: theory, facts, and formulas (2 ed.). Princeton, NJ: Princeton University Press. p. 168. ISBN 978-0-691-14039-1.
- ^ Herstein, I. N. (1975). Topics in Algebra (2nd ed.). New York: Wiley. pp. 285–290. ISBN 0-471-01090-1. OCLC 3307396.
- ^ Drazin, M. P.; Dungey, J. W.; Gruenberg, K. W. (1951). "Some Theorems on Commutative Matrices". Journal of the London Mathematical Society. 26 (3): 221–228. doi:10.1112/jlms/s1-26.3.221.
- ^ Prasolov, V. V. (1994). Problems and Theorems in Linear Algebra. Simeon Ivanov. Providence, R.I.: American Mathematical Society. pp. 178–179. ISBN 9780821802366. OCLC 30076024.
Triangular matrix
View on GrokipediaDefinition and Fundamentals
Upper and Lower Triangular Matrices
A triangular matrix is a square matrix in which either all entries above the main diagonal are zero or all entries below the main diagonal are zero. An upper triangular matrix has zeros strictly below the main diagonal, while a lower triangular matrix has zeros strictly above it. Diagonal matrices, with all off-diagonal entries zero, qualify as both upper and lower triangular.[9] Formally, an matrix is upper triangular if for all , meaning entries below the diagonal vanish. Similarly, an matrix is lower triangular if for all , with entries above the diagonal being zero.[1] The concept traces back to Gaussian elimination, a method that Carl Friedrich Gauss applied in 1809 to solve linear systems in celestial mechanics, though the technique has earlier origins; it transforms the coefficient matrix into upper triangular form through row operations. This process reveals a pattern of non-zero entries forming a triangle above the diagonal, from which the modern terminology derives.[10] Consider a basic 2×2 upper triangular matrix: where are elements from the underlying field, such as real numbers. A 2×2 lower triangular example is: For a 3×3 upper triangular matrix, the form is: with zeros below the diagonal.[9] Diagonal matrices represent a special case of triangular matrices, where both upper and lower off-diagonal entries are zero, simplifying many algebraic operations while retaining the triangular structure.[1]Notation and Conventions
In mathematical literature, upper triangular matrices are commonly denoted by the letter , often rendered in boldface as to distinguish them as matrices, while lower triangular matrices are denoted by or . These notations facilitate clear representation in contexts such as factorizations and eigenvalue computations. The entries of an triangular matrix are indexed by row and column , where . The main diagonal consists of the entries where . The superdiagonal comprises the entries where (directly above the main diagonal), and the subdiagonal comprises those where (directly below the main diagonal).[11][12] This indexing convention aligns with standard matrix notation and aids in identifying the zero patterns characteristic of triangular forms. The zero pattern of a triangular matrix is typically visualized using explicit matrix diagrams that highlight the positions of zeros. For an upper triangular matrix, all entries below the main diagonal are zero ( for ), with potentially nonzero real or complex entries on the diagonal and above it: Analogously, a lower triangular matrix has zeros above the main diagonal ( for ), with nonzero entries on and below the diagonal.[13] These visualizations emphasize the structured sparsity compared to full matrices, where entries can be nonzero throughout. Triangular matrices are not symmetric in general, as symmetry requires for all , which would force all off-diagonal entries to zero if the matrix is triangular, reducing it to a diagonal matrix.[9] This contrasts with full matrices, which may have nonzero entries in symmetric positions without such restrictions.[13] Standard conventions assume triangular matrices are square of size over the field of real numbers or complex numbers , ensuring compatibility with common algebraic operations and spectral analysis.[13]Properties
Algebraic Properties
A triangular matrix possesses several key algebraic properties that distinguish it from general matrices. The transpose of an upper triangular matrix is lower triangular, and conversely, the transpose of a lower triangular matrix is upper triangular. This follows from the definition, as the entries , which swaps the zero patterns above and below the diagonal.[14] The determinant of a triangular matrix , whether upper or lower, is the product of its diagonal entries: . This result holds because cofactor expansion along a row or column of zeros (the off-diagonal parts) reduces the determinant recursively to the product of the diagonals, as proven by induction on the matrix size.[15][14] The trace of a triangular matrix, defined as the sum of its diagonal entries , remains invariant under similarity transformations. For any invertible matrix , , a property that holds for all square matrices but is particularly relevant for triangular forms where the trace equals the sum of the eigenvalues.[16][14] A triangular matrix is invertible if and only if all its diagonal entries are nonzero. In this case, the inverse is also triangular of the same type (upper or lower), preserving the zero pattern off the diagonal.[17][14] The rank of a triangular matrix equals the number of its nonzero diagonal entries. This follows from the fact that the diagonal entries determine the linear independence of the rows or columns, with zero diagonals corresponding to linearly dependent parts.[18][14]Spectral Properties
A triangular matrix possesses particularly straightforward spectral properties due to its structure. For an triangular matrix , whether upper or lower, the eigenvalues are precisely the diagonal entries for , counted with their algebraic multiplicities matching the multiplicities of these entries on the diagonal.[19] This follows from the fact that the characteristic equation simplifies dramatically for triangular matrices, as the determinant of is the product of the diagonal terms . The characteristic polynomial of is thus explicitly given by which directly encodes the eigenvalues and their multiplicities without requiring further computation.[20] A direct consequence is that the trace of , defined as the sum of its diagonal entries , equals the sum of the eigenvalues (with multiplicity), a property that holds generally but is immediately verifiable for triangular matrices.[21] Regarding the minimal polynomial, it always divides the characteristic polynomial, as per the Cayley-Hamilton theorem, and thus takes the form of a product of factors raised to powers at most the algebraic multiplicities. If all diagonal entries are distinct, the minimal polynomial equals the characteristic polynomial, since the distinct eigenvalues imply that the matrix is diagonalizable over the complex numbers.[22] In relation to the Jordan canonical form, a triangular matrix with distinct diagonal entries is similar to a diagonal matrix (hence its Jordan form is diagonal, a special case of upper triangular), but in general, the Jordan form of a triangular matrix is upper triangular with the same diagonal eigenvalues, though the original matrix need not exhibit the specific block structure of Jordan blocks unless the superdiagonal entries align accordingly.[23]Operations and Computations
Multiplication and Powers
The product of two upper triangular matrices is upper triangular, and the product of two lower triangular matrices is lower triangular. However, the product of an upper triangular matrix and a lower triangular matrix is not necessarily triangular.[24] The -entry of the product of two upper triangular matrices and is zero if , since the relevant summands vanish due to the zero entries below the diagonal in and . If , then [25] Any positive integer power of a triangular matrix is triangular of the same type (upper or lower), as it can be obtained by repeated multiplication, which preserves the structure. The diagonal entries of are the -th powers of the diagonal entries of , since off-diagonal contributions to the diagonal vanish in the expansion.[25] For a strictly triangular matrix (upper or lower, with zeros on the main diagonal), is nilpotent, meaning for an matrix . This follows from induction: each power shifts the nonzero entries further from the diagonal until all entries are zero after multiplications.[26] Triangular matrices do not commute under multiplication in general, though diagonal matrices (a special case) do.[22]Solving Linear Systems
One of the primary advantages of triangular matrices in numerical linear algebra is their role in efficiently solving linear systems of the form , where is triangular and is a known vector. Unlike general dense matrices, which require operations via methods like Gaussian elimination, triangular systems can be solved in time using specialized substitution algorithms that exploit the zero structure above or below the diagonal.[27] These methods, known as forward and backward substitution, iteratively compute the components of the solution vector without needing matrix inversion or factorization of the original system.[28] For a lower triangular system , where is an lower triangular matrix with nonzero diagonal entries, forward substitution proceeds from the first equation to the last. The first component is . For each subsequent , This formula isolates after substituting previously computed values, ensuring each equation involves only known terms.[27] A step-by-step pseudocode implementation is as follows:function forward_substitution(L, b):
n = length(b)
x = zeros(n)
for i = 1 to n:
sum = 0
for j = 1 to i-1:
sum += L[i,j] * x[j]
x[i] = (b[i] - sum) / L[i,i]
return x
function forward_substitution(L, b):
n = length(b)
x = zeros(n)
for i = 1 to n:
sum = 0
for j = 1 to i-1:
sum += L[i,j] * x[j]
x[i] = (b[i] - sum) / L[i,i]
return x
function backward_substitution(U, b):
n = [length](/page/Length)(b)
x = zeros(n)
for i = n downto 1:
sum = 0
for j = i+1 to n:
sum += U[i,j] * x[j]
x[i] = (b[i] - sum) / U[i,i]
return x
function backward_substitution(U, b):
n = [length](/page/Length)(b)
x = zeros(n)
for i = n downto 1:
sum = 0
for j = i+1 to n:
sum += U[i,j] * x[j]
x[i] = (b[i] - sum) / U[i,i]
return x
