Recent from talks
Nothing was collected or created yet.
Generalized inverse
View on WikipediaIn mathematics, and in particular, algebra, a generalized inverse (or, g-inverse) of an element x is an element y that has some properties of an inverse element but not necessarily all of them. The purpose of constructing a generalized inverse of a matrix is to obtain a matrix that can serve as an inverse in some sense for a wider class of matrices than invertible matrices. Generalized inverses can be defined in any mathematical structure that involves associative multiplication, that is, in a semigroup. This article describes generalized inverses of a matrix .
A matrix is a generalized inverse of a matrix if [1][2][3] A generalized inverse exists for an arbitrary matrix, and when a matrix has a regular inverse, this inverse is its unique generalized inverse.[1]
Motivation
[edit]Consider the linear system
where is an matrix and the column space of . If and is nonsingular then will be the solution of the system. Note that, if is nonsingular, then
Now suppose is rectangular (), or square and singular. Then we need a right candidate of order such that for all
That is, is a solution of the linear system . Equivalently, we need a matrix of order such that
Hence we can define the generalized inverse as follows: Given an matrix , an matrix is said to be a generalized inverse of if [1][2][3] The matrix has been termed a regular inverse of by some authors.[5]
Types
[edit]Important types of generalized inverse include:
- One-sided inverse (right inverse or left inverse)
- Right inverse: If the matrix has dimensions and , then there exists an matrix called the right inverse of such that , where is the identity matrix.
- Left inverse: If the matrix has dimensions and , then there exists an matrix called the left inverse of such that , where is the identity matrix.[6]
- Bott–Duffin inverse
- Drazin inverse
- Moore–Penrose inverse
Some generalized inverses are defined and classified based on the Penrose conditions:
where denotes conjugate transpose. If satisfies the first condition, then it is a generalized inverse of . If it satisfies the first two conditions, then it is a reflexive generalized inverse of . If it satisfies all four conditions, then it is the pseudoinverse of , which is denoted by and also known as the Moore–Penrose inverse, after the pioneering works by E. H. Moore and Roger Penrose.[2][7][8][9][10][11] It is convenient to define an -inverse of as an inverse that satisfies the subset of the Penrose conditions listed above. Relations, such as , can be established between these different classes of -inverses.[1]
When is non-singular, any generalized inverse and is therefore unique. For a singular , some generalised inverses, such as the Drazin inverse and the Moore–Penrose inverse, are unique, while others are not necessarily uniquely defined.
Examples
[edit]Reflexive generalized inverse
[edit]Let
Since , is singular and has no regular inverse. However, and satisfy Penrose conditions (1) and (2), but not (3) or (4). Hence, is a reflexive generalized inverse of .
One-sided inverse
[edit]Let
Since is not square, has no regular inverse. However, is a right inverse of . The matrix has no left inverse.
Inverse of other semigroups (or rings)
[edit]The element b is a generalized inverse of an element a if and only if , in any semigroup (or ring, since the multiplication function in any ring is a semigroup).
The generalized inverses of the element 3 in the ring are 3, 7, and 11, since in the ring :
The generalized inverses of the element 4 in the ring are 1, 4, 7, and 10, since in the ring :
If an element a in a semigroup (or ring) has an inverse, the inverse must be the only generalized inverse of this element, like the elements 1, 5, 7, and 11 in the ring .
In the ring any element is a generalized inverse of 0; however 2 has no generalized inverse, since there is no b in such that .
Construction
[edit]The following characterizations are easy to verify:
- A right inverse of a non-square matrix is given by , provided has full row rank.[6]
- A left inverse of a non-square matrix is given by , provided has full column rank.[6]
- If is a rank factorization, then is a g-inverse of , where is a right inverse of and is left inverse of .
- If for any non-singular matrices and , then is a generalized inverse of for arbitrary and .
- Let be of rank . Without loss of generality, letwhere is the non-singular submatrix of . Then,is a generalized inverse of if and only if .
Uses
[edit]Any generalized inverse can be used to determine whether a system of linear equations has any solutions, and if so to give all of them. If any solutions exist for the n × m linear system
- ,
with vector of unknowns and vector of constants, all solutions are given by
- ,
parametric on the arbitrary vector , where is any generalized inverse of . Solutions exist if and only if is a solution, that is, if and only if . If A has full column rank, the bracketed expression in this equation is the zero matrix and so the solution is unique.[12]
Generalized inverses of matrices
[edit]The generalized inverses of matrices can be characterized as follows. Let , and
be its singular-value decomposition. Then for any generalized inverse , there exist[1] matrices , , and such that
Conversely, any choice of , , and for matrix of this form is a generalized inverse of .[1] The -inverses are exactly those for which , the -inverses are exactly those for which , and the -inverses are exactly those for which . In particular, the pseudoinverse is given by :
Transformation consistency properties
[edit]In practical applications it is necessary to identify the class of matrix transformations that must be preserved by a generalized inverse. For example, the Moore–Penrose inverse, satisfies the following definition of consistency with respect to transformations involving unitary matrices U and V:
- .
The Drazin inverse, satisfies the following definition of consistency with respect to similarity transformations involving a nonsingular matrix S:
- .
The unit-consistent (UC) inverse,[13] satisfies the following definition of consistency with respect to transformations involving nonsingular diagonal matrices D and E:
- .
The fact that the Moore–Penrose inverse provides consistency with respect to rotations (which are orthonormal transformations) explains its widespread use in physics and other applications in which Euclidean distances must be preserved. The UC inverse, by contrast, is applicable when system behavior is expected to be invariant with respect to the choice of units on different state variables, e.g., miles versus kilometers.
See also
[edit]Citations
[edit]- ^ a b c d e f Ben-Israel & Greville 2003, pp. 2, 7
- ^ a b c Nakamura 1991, pp. 41–42
- ^ a b Rao & Mitra 1971, pp. vii, 20
- ^ Rao & Mitra 1971, p. 24
- ^ Rao & Mitra 1971, pp. 19–20
- ^ a b c Rao & Mitra 1971, p. 19
- ^ Rao & Mitra 1971, pp. 20, 28, 50–51
- ^ Ben-Israel & Greville 2003, p. 7
- ^ Campbell & Meyer 1991, p. 10
- ^ James 1978, p. 114
- ^ Nakamura 1991, p. 42
- ^ James 1978, pp. 109–110
- ^ Uhlmann 2018
Sources
[edit]Textbook
[edit]- Ben-Israel, Adi; Greville, Thomas Nall Eden (2003). Generalized Inverses: Theory and Applications (2nd ed.). New York, NY: Springer. doi:10.1007/b97366. ISBN 978-0-387-00293-4.
- Campbell, Stephen L.; Meyer, Carl D. (1991). Generalized Inverses of Linear Transformations. Dover. ISBN 978-0-486-66693-8.
- Horn, Roger Alan; Johnson, Charles Royal (1985). Matrix Analysis. Cambridge University Press. ISBN 978-0-521-38632-6.
- Nakamura, Yoshihiko (1991). Advanced Robotics: Redundancy and Optimization. Addison-Wesley. ISBN 978-0201151985.
- Rao, C. Radhakrishna; Mitra, Sujit Kumar (1971). Generalized Inverse of Matrices and its Applications. New York: John Wiley & Sons. pp. 240. ISBN 978-0-471-70821-6.
Publication
[edit]- James, M. (June 1978). "The generalised inverse". The Mathematical Gazette. 62 (420): 109–114. doi:10.2307/3617665. JSTOR 3617665.
- Uhlmann, Jeffrey K. (2018). "A Generalized Matrix Inverse that is Consistent with Respect to Diagonal Transformations". SIAM Journal on Matrix Analysis and Applications. 239 (2): 781–800. doi:10.1137/17M113890X.
- Zheng, Bing; Bapat, Ravindra (2004). "Generalized inverse A(2)T,S and a rank equation". Applied Mathematics and Computation. 155 (2): 407–415. doi:10.1016/S0096-3003(03)00786-0.
Generalized inverse
View on GrokipediaFundamentals
Definition and Motivation
In algebraic structures such as semigroups or rings, a generalized inverse of an element is an element (often denoted ) that satisfies the equation . This condition represents a minimal form of partial invertibility, where acts as an inverse for restricted to the image of , without requiring full invertibility or the existence of a two-sided inverse. Weaker variants include one-sided generalized inverses, such as those satisfying only (a left-ish inverse) or (a right-ish inverse), which capture asymmetric notions of partial reversal in non-commutative settings. In the context of linear algebra, the concept applies to matrices over fields like the real or complex numbers, where is an arbitrary matrix and (or ) is an matrix satisfying . This extends the classical matrix inverse, which exists only for square nonsingular matrices (where ), to handle singular square matrices or rectangular ones, where the determinant is undefined or the dimensions preclude a two-sided inverse. The equation ensures that is a projection onto the column space of , providing a way to "undo" the action of where possible. The primary motivation for generalized inverses arises in solving linear systems , where may be singular or rectangular, rendering standard inversion impossible. Here, if the system is consistent (i.e., lies in the column space of ), a generalized inverse yields a particular solution , while the full general solution is given by for arbitrary (or ), capturing the null space contributions. To derive this, note that substituting into gives , confirming consistency preservation; the term projects onto the kernel of , parameterizing all solutions. This framework addresses the failure of regular inverses in underdetermined or overdetermined systems, enabling systematic treatment of ill-posed problems. As a foundational concept, the generalized inverse establishes the limitations of classical invertibility and sets the stage for exploring specialized types, such as the Moore-Penrose inverse, which refines the basic definition with additional symmetry and orthogonality properties.Historical Background
The concept of the generalized inverse emerged in the early 20th century as a means to extend the notion of matrix inversion beyond nonsingular square matrices, particularly to address reciprocal systems in linear equations. In 1920, E.H. Moore introduced the idea in his abstract "On the Reciprocal of the General Algebraic Matrix," where he described a generalized reciprocal for arbitrary algebraic matrices, laying foundational groundwork for handling singular and rectangular cases. This work, later elaborated in a 1935 publication, motivated solutions to inconsistent linear systems by generalizing the inverse to encompass projections onto relevant subspaces. Mid-20th-century advancements formalized and expanded these ideas into broader algebraic structures. Roger Penrose's 1955 paper "A Generalized Inverse for Matrices" provided a rigorous definition through four axiomatic conditions, establishing the unique pseudoinverse now known as the Moore-Penrose inverse, which satisfies symmetry, idempotence, and orthogonality properties for real or complex matrices.[6] An explicit formula for the Moore-Penrose inverse via full-rank factorization of matrices was first pointed out by C. C. MacDuffee in private communications, bridging linear algebra with ring theory.[7] Parallel developments in semigroup theory introduced partial inverses; the algebraic framework of inverse semigroups, which model partial symmetries through unique idempotent inverses, was pioneered in the 1950s by Gordon B. Preston and Viktor V. Wagner, extending generalized inversion to non-invertible transformations.[8] In the late 20th century, specialized types of generalized inverses proliferated to address singular operators and ring elements. Michael P. Drazin introduced the Drazin inverse in 1958 for elements in associative rings, defined via a power condition that captures the invertible part of nilpotent perturbations, proving useful for differential equations and Markov chains. The group inverse, applicable to index-1 elements where the kernel and range align appropriately, was developed in semigroup and matrix contexts in the 1970s, emphasizing reflexive properties in partial algebraic structures. Extensions into the 21st century have refined these concepts with new characterizations and broader applicability. The core inverse, introduced by Oskar M. Baksalary and Götz Trenkler in 2010 as an alternative to the group inverse for index-1 matrices, combines outer inverse properties with range conditions to preserve core-EP structures.[9] Recent work includes a 2023 geometric characterization of the Moore-Penrose inverse using polar decompositions of operator perturbations in Hilbert spaces, enhancing perturbation theory.[10] Ongoing refinements, such as the W-weighted m-weak core inverse proposed in 2024, continue to extend these inverses to rectangular matrices without introducing major paradigm shifts.[11]Types
One-Sided and Reflexive Inverses
In the context of generalized inverses within semigroups and matrix algebras, one-sided inverses provide a minimal extension of classical left and right inverses to non-invertible elements. A right one-sided inverse of an element satisfies the equation , which captures a partial left-inversion property without requiring full invertibility. This condition holds for a strong right inverse when , applicable to surjective linear maps or matrices where the number of rows is at most the number of columns with full row rank. Similarly, a left one-sided inverse satisfies , representing a partial right-inversion, and aligns with the strong left inverse for injective maps or matrices with and full column rank. These definitions arise naturally in semigroup theory, where they characterize elements within Green's - and -classes relative to the variety of inverses .[6][1][12] A reflexive generalized inverse combines both one-sided conditions, satisfying and simultaneously, thereby acting as a two-sided partial inverse that preserves under composition with from either side. This makes a von Neumann regular inverse in semigroup terms, ensuring is regular (i.e., for some ). Unlike one-sided inverses, which apply to rectangular matrices or asymmetric semigroup elements (e.g., enabling solutions in over- or under-determined systems), reflexive inverses exhibit square-like behavior, requiring compatible dimensions or class structures where left and right properties align. The Moore-Penrose inverse represents a special reflexive type augmented with symmetry conditions for uniqueness.[1][12][13] Key properties of these inverses include their non-uniqueness—multiple may satisfy the equations for a given —and a close relation to idempotents: for a right one-sided inverse, is idempotent since , projecting onto the image of ; analogously, is idempotent for a left one-sided inverse. In finite semigroups, reflexive generalized inverses exist for every element, as finiteness implies regularity (every admits with and ), though one-sided versions may exist more broadly via class decompositions like for left inverses, where denotes idempotents in the left principal ideal. These structures facilitate applications in solving inconsistent equations or analyzing partial orderings in algebraic settings without full invertibility.[12][13]Moore-Penrose Inverse
The Moore-Penrose inverse, also known as the pseudoinverse, of a matrix is a unique matrix that generalizes the concept of the inverse for non-square or singular matrices, satisfying a specific set of four conditions introduced by Roger Penrose. These conditions ensure that provides a canonical way to solve linear systems in Hilbert spaces, particularly for complex matrices. The notion traces back to E. H. Moore's earlier work on the "general reciprocal" of matrices, which laid foundational ideas for handling divisors of zero in algebraic structures.[6] The four Penrose conditions defining are:Drazin and Group Inverses
The index of a square matrix , denoted , is defined as the smallest nonnegative integer such that , or equivalently, .[15] This index measures the "singularity depth" of and is finite if and only if the ascent (dimension of the generalized kernel growth) stabilizes.[16] The Drazin inverse of , denoted , is a generalized inverse that exists if and only if is finite. It is the unique matrix satisfying the conditions where .[15] The first equation ensures that "inverts" on the range of , while the latter two impose idempotence and commutativity. The Drazin inverse commutes with and is idempotent on the core subspace, with being the spectral idempotent projecting onto the range of .[17] When , is invertible and , which is a reflexive generalized inverse. A special case arises when , in which the Drazin inverse is called the group inverse, denoted .[15] It satisfies These equations characterize uniquely when it exists, and it arises naturally in the study of power-regular elements in semigroups where the index is at most 1. For square matrices over the complex numbers, the Drazin inverse relates closely to the Jordan canonical form of . Specifically, if where is the Jordan form, then , with obtained by replacing each Jordan block for a nonzero eigenvalue with the corresponding block of (adjusted for the nilpotent part via the finite index), and setting blocks for eigenvalue 0 to zero except for the core structure aligned with the index.[17] This construction inverts the semisimple part while annihilating the nilpotent component beyond the index. An equivalent characterization of the Drazin inverse involves the core polynomial equation which highlights that acts as an identity on the image of . This equation, along with commutativity and idempotence, ensures uniqueness and ties the inverse to the minimal polynomial of restricted to the non-nilpotent part.Other Types
The core inverse, applicable to square matrices of index at most one, is defined as the unique matrix satisfying the equations , , and .[9] This inverse can be explicitly expressed for such matrices as , where denotes the group inverse. It serves as an intermediate between the group inverse and the Moore-Penrose inverse, particularly useful for matrices where the index condition holds, and extends the Drazin inverse for higher indices in a specialized manner. The Bott-Duffin inverse, originally developed for analyzing electrical networks, provides a generalized {1,3}-inverse for square matrices that minimizes a specific norm in representations involving projections. For positive operators, it arises in contexts where an operator is represented as for a suitable projection , yielding the inverse as the minimizer over such decompositions.[18] This construction ensures invertibility in constrained subspaces and has niche applications in optimization problems requiring bounded representations.[19] Recent extensions include the extended core inverse, introduced in 2024, which for square complex matrices combines the sum and difference of the Moore-Penrose inverse, core-EP inverse, and MPCEP inverse to form a unique inner inverse satisfying specific matrix equations.[20] This variant reduces to the standard core inverse for index-one matrices and addresses limitations in prior extensions by preserving inner inverse properties.[20] In 2025, the generalized right core inverse was defined in Banach *-algebras as an extension of the pseudo right core inverse, characterized via right core decompositions and quasi-nilpotent parts, with polar-like properties that facilitate algebraic manipulations in non-commutative settings.[21]Constructions
For Matrices
One practical method for constructing a generalized inverse of a finite-dimensional matrix with rank relies on its rank factorization , where has full column rank and has full row rank.[22] To obtain a reflexive generalized inverse (satisfying both and ), compute the Moore-Penrose inverses and explicitly using their full-rank properties: and . Then set . This satisfies , confirming the {1}-inverse property; the reflexive property follows similarly from .[22] The Moore-Penrose inverse , a specific reflexive generalized inverse satisfying all four Penrose conditions, can be constructed via the singular value decomposition (SVD) of . Compute the SVD , where and are unitary matrices, is diagonal with nonnegative singular values on the main diagonal (and zeros elsewhere), and . Form as the matrix with diagonal entries for and zeros otherwise. Then . This construction satisfies the Penrose conditions.[23] The following steps outline the computation of using SVD in practice:function A_plus = moore_penrose(A, m, n)
[U, Sigma_diag, V] = svd(A); // Compute SVD: A = U * diag(Sigma_diag) * V^H
// Sigma_diag is m x n diagonal matrix with singular values on diagonal
r = rank(Sigma_diag); // Number of nonzero singular values
Sigma_plus = zeros(n, m);
for i = 1 to r
Sigma_plus(i, i) = 1 / Sigma_diag(i, i);
end
A_plus = V * Sigma_plus * U'; // For real matrices, H = transpose
end
function A_plus = moore_penrose(A, m, n)
[U, Sigma_diag, V] = svd(A); // Compute SVD: A = U * diag(Sigma_diag) * V^H
// Sigma_diag is m x n diagonal matrix with singular values on diagonal
r = rank(Sigma_diag); // Number of nonzero singular values
Sigma_plus = zeros(n, m);
for i = 1 to r
Sigma_plus(i, i) = 1 / Sigma_diag(i, i);
end
A_plus = V * Sigma_plus * U'; // For real matrices, H = transpose
end
