Recent from talks
Nothing was collected or created yet.
Hilbert matrix
View on WikipediaIn linear algebra, a Hilbert matrix, introduced by Hilbert (1894), is a square matrix with entries being the unit fractions
For example, this is the 5 × 5 Hilbert matrix:
The entries can also be defined by the integral
that is, as a Gramian matrix for powers of x. It arises in the least squares approximation of arbitrary functions by polynomials.
The Hilbert matrices are canonical examples of ill-conditioned matrices, being notoriously difficult to use in numerical computation. For example, the 2-norm condition number of the matrix above is about 4.8×105.
Historical note
[edit]Hilbert (1894) introduced the Hilbert matrix to study the following question in approximation theory: "Assume that I = [a, b], is a real interval. Is it then possible to find a non-zero polynomial P with integer coefficients, such that the integral
is smaller than any given bound ε > 0, taken arbitrarily small?" To answer this question, Hilbert derives an exact formula for the determinant of the Hilbert matrices and investigates their asymptotics. He concludes that the answer to his question is positive if the length b − a of the interval is smaller than 4.
Properties
[edit]The Hilbert matrix is symmetric and positive definite. The Hilbert matrix is also totally positive (meaning that the determinant of every submatrix is positive).
The Hilbert matrix is an example of a Hankel matrix. It is also a specific example of a Cauchy matrix.
The determinant can be expressed in closed form, as a special case of the Cauchy determinant. The determinant of the n × n Hilbert matrix is
where
Hilbert already mentioned the curious fact that the determinant of the Hilbert matrix is the reciprocal of an integer (see sequence OEIS: A005249 in the OEIS), which also follows from the identity
Using Stirling's approximation of the factorial, one can establish the following asymptotic result:
where an converges to the constant as , where A is the Glaisher–Kinkelin constant.
The inverse of the Hilbert matrix can be expressed in closed form using binomial coefficients; its entries are
where n is the order of the matrix.[1] It follows that the entries of the inverse matrix are all integers, and that the signs form a checkerboard pattern, being positive on the principal diagonal. For example,
The condition number of the n × n Hilbert matrix grows as .
Applications
[edit]The method of moments applied to polynomial distributions results in a Hankel matrix, which in the special case of approximating a probability distribution on the interval [0, 1] results in a Hilbert matrix. This matrix needs to be inverted to obtain the weight parameters of the polynomial distribution approximation.[2]
References
[edit]- ^ Choi, Man-Duen (1983). "Tricks or Treats with the Hilbert Matrix". The American Mathematical Monthly. 90 (5): 301–312. doi:10.2307/2975779. JSTOR 2975779.
- ^ Munkhammar, Joakim; Mattsson, Lars; Rydén, Jesper (2017). "Polynomial probability distribution estimation using the method of moments". PLOS ONE. 12 (4) e0174573. Bibcode:2017PLoSO..1274573M. doi:10.1371/journal.pone.0174573. PMC 5386244. PMID 28394949.
Further reading
[edit]- Hilbert, David (1894), "Ein Beitrag zur Theorie des Legendre'schen Polynoms", Acta Mathematica, 18: 155–159, doi:10.1007/BF02418278, ISSN 0001-5962, JFM 25.0817.02. Reprinted in Hilbert, David. "article 21". Collected papers. Vol. II.
- Beckermann, Bernhard (2000). "The condition number of real Vandermonde, Krylov and positive definite Hankel matrices". Numerische Mathematik. 85 (4): 553–577. CiteSeerX 10.1.1.23.5979. doi:10.1007/PL00005392. S2CID 17777214.
- Choi, M.-D. (1983). "Tricks or Treats with the Hilbert Matrix". American Mathematical Monthly. 90 (5): 301–312. doi:10.2307/2975779. JSTOR 2975779.
- Todd, John (1954). "The Condition of the Finite Segments of the Hilbert Matrix". In Taussky, O. (ed.). Contributions to the Solution of Systems of Linear Equations and the Determination of Eigenvalues. National Bureau of Standards, Applied Mathematics Series. Vol. 39. pp. 109–116. ISBN 0-598-85383-9. OCLC 540840412.
{{cite book}}: ISBN / Date incompatibility (help) - Wilf, H.S. (1970). Finite Sections of Some Classical Inequalities. Springer. doi:10.1007/978-3-642-86712-5. ISBN 978-3-540-04809-1.
Hilbert matrix
View on GrokipediaDefinition and Construction
Matrix Entries
The Hilbert matrix is defined by its entries for .[5] This uses the standard 1-based indexing convention.[5] The Hilbert matrix arises as a special case of a Cauchy matrix.[6] For illustration, the Hilbert matrix is with trace .[5] The Hilbert matrix is with trace .[5] These examples highlight the decreasing pattern in entries as grows, consistent with the reciprocal form.[5] The matrix is symmetric by construction.[7]Integral Representation
The entries of the finite-dimensional Hilbert matrix admit an integral representation given by for , where the equality follows from direct evaluation of the definite integral.[8] This formulation establishes the direct equivalence between the continuous integral expression and the discrete algebraic definition of the matrix entries. This integral representation arises naturally as the Gram matrix associated with the standard inner product applied to the monomial basis functions .[8] Specifically, the -th entry is the inner product , capturing the overlap between these basis elements in the space of square-integrable functions on . The space equipped with this inner product forms a Hilbert space, and the resulting Gram matrix inherits positive definiteness from the linear independence of the monomials in this space.[8] Intuitively, this positive definiteness reflects the fact that the monomials, while not orthogonal, span a finite-dimensional subspace without linear dependencies, ensuring that the quadratic form for any nonzero coefficient vector . This construction generalizes to broader contexts by incorporating weights or altering the integration domain; for instance, weighted versions take the form for a positive weight function , yielding matrices that retain similar Gram matrix properties in weighted spaces.[9] Such generalizations appear in studies of operators on Bergman or Hardy spaces, where the weight influences boundedness and spectral behavior.[10]Mathematical Properties
Algebraic Structure
The Hilbert matrix of order is symmetric, as its entries satisfy for all , since the denominators are invariant under index interchange. It belongs to the class of Hankel matrices, characterized by constant values along each anti-diagonal, because each entry depends solely on the sum . Additionally, the Hilbert matrix is a special case of a Cauchy matrix, obtained by setting the parameters and for in the general form . The Hilbert matrix is positive definite, meaning that for any nonzero real vector , the quadratic form . A proof of this property leverages the integral representation , which follows from direct computation of the integral. Substituting into the quadratic form yields The integrand is nonnegative for all and strictly positive on a set of positive measure unless , ensuring the integral is positive. As a symmetric positive definite matrix, the Hilbert matrix admits a unique Cholesky decomposition with a positive definite lower triangular factor . The Hilbert matrix is totally positive, meaning that every square submatrix has a positive determinant. This property arises from its structure as a Gram matrix of the monomial basis with respect to the inner product on the interval , where the positive definite kernel ensures all minors are positive. Total positivity implies variation-diminishing properties, such as the fact that the sign changes in the solution to are at most as many as in , which has applications in oscillation theory for linear systems. The Hankel structure of the Hilbert matrix connects it to moment matrices in approximation theory, as its entries form the Hankel matrix associated with the moments of the uniform distribution on , with . This representation underscores its role in problems involving orthogonal polynomials and quadrature, where the positive definiteness guarantees unique representations.Spectral Characteristics
The Hilbert matrix is symmetric positive definite, ensuring that its eigenvalues are real and positive. The largest eigenvalue approaches as increases, reflecting the spectral norm of the infinite-dimensional Hilbert operator. The eigenvalues exhibit a distribution where the larger ones cluster near , while the smaller ones decay rapidly toward zero. The smallest eigenvalue decays exponentially with , with the asymptotic behavior for a constant involving , leading to superpolynomial ill-conditioning. More precise approximations indicate for intermediate indices , providing bounds on the spectral decay. This rapid clustering of smaller eigenvalues near zero underscores the matrix's tendency toward singularity as grows. Since is symmetric, its singular values coincide with the eigenvalues . Consequently, the singular values also feature one large value near and the rest decaying geometrically to exponentially small magnitudes, amplifying sensitivity in low-rank approximations. The 2-norm condition number , exhibiting exponential growth dominated by the term in the logarithm. This logarithmic rate implies that even modest increases in render numerically unstable for standard floating-point arithmetic, as perturbations amplify by factors exceeding machine precision. For instance, when , .Explicit Formulas
The determinant of the Hilbert matrix admits a closed-form expression as where denotes the product of the first factorials (with by convention as the empty product).[5] This formula arises as a special case of the Cauchy determinant for matrices of the form , with and .[11] For small dimensions, explicit values include , , , and , illustrating the rapid decay that underscores the matrix's ill-conditioning.[5] This determinant carries a geometric interpretation as the square of the volume of the parallelepiped spanned by the monomials in the Hilbert space equipped with the inner product , since is the Gram matrix of these functions.[5] Combinatorially, the factorial products in count the number of permutations of multisets, reflecting the underlying structure of the matrix entries derived from beta integral representations. The inverse of also possesses an explicit entrywise formula: where the indices range from $1n(-1)^{i+j}H_n{0,1,\dots,n-1}$ in the theory of orthogonal polynomials associated with the monomial basis.[11]Historical Development
Origins with David Hilbert
David Hilbert first introduced the matrix now known as the Hilbert matrix in his 1894 paper "Ein Beitrag zur Theorie des Legendreschen Polynoms," published in Acta Mathematica. In this seminal work, Hilbert explored the expansion of arbitrary functions in series of Legendre polynomials, connecting the matrix to problems in integral equations and the least squares approximation of continuous functions by polynomials on bounded intervals. The matrix emerged naturally as the coefficient matrix in the normal equations for minimizing the L² error in such approximations, where the entries represent inner products of monomials under the uniform measure.[12] Hilbert employed the Hilbert matrix to prove the existence of non-zero polynomials with integer coefficients whose L² norm on the interval [0,1] is less than 1/4. This result stems from analyzing the quadratic form associated with the matrix, which quantifies the L² norm of polynomials. By examining the structure of the matrix, Hilbert demonstrated that its positive definiteness ensures the approximation problem is well-posed, with the non-zero determinant guaranteeing the invertibility of the system and thus the uniqueness of the least squares solution.[12] A key insight from Hilbert's analysis involves the moment problem, where the Hilbert matrix serves as the moment matrix for the moments of the uniform distribution on [0,1]. He showed that for any interval of length less than 4, non-zero polynomials with integer coefficients exist whose L² norm on that interval can be made arbitrarily small relative to their degree, facilitating effective approximations of continuous functions. This threshold of 4 arises from the scaling properties of the Gram matrix determinant, which decreases sufficiently fast to imply the presence of lattice points (corresponding to integer coefficients) in small L² balls when the interval is short enough. The invertibility ensured by the non-zero determinant further supports the stability of these approximation procedures in the context of the moment problem.[12]Subsequent Mathematical Advances
In the early 20th century, the Hilbert matrix was recognized as a special case of a Hankel matrix, where entries depend solely on the sum of indices, a property inherent to its construction that facilitated studies in operator theory.[13] It was also identified as a symmetric Cauchy matrix, with entries expressible in the form where and , enabling connections to broader classes of structured matrices.[13] By the mid-20th century, significant progress emerged in understanding the Hilbert matrix's total positivity, a property where all minors are nonnegative (and strictly positive for the infinite case), as detailed in Gantmacher and Krein's foundational work on oscillation matrices. This led to oscillation theorems characterizing the sign changes in eigenvectors, with Karlin extending these results to totally positive kernels and matrices, confirming the Hilbert matrix's strict total positivity and its implications for eigenvalue interlacing. In 1983, Choi provided an elegant exposition of the explicit inverse of the finite Hilbert matrix, whose entries are integers given by , and refined bounds on its determinant, building on earlier formulas. These results highlighted the matrix's algebraic richness without delving into numerical challenges. Early literature on the Hilbert matrix emphasized theoretical and algebraic properties, with limited awareness of its severe ill-conditioning—evident in exponentially growing condition numbers —until the rise of numerical computing in the 1950s and 1960s, when it became a benchmark for testing algorithms.[14] Recent advances from 2021 to 2024 have focused on the spectrum of the Hilbert operator on analytic function spaces. In Hardy spaces (1 < p < ∞), Silbermann (2021) described the spectrum as a curved contour in the complex plane, parametrized by for . For weighted Bergman spaces , Dmitrovic and Karapetrović (2023) established the operator norm as under specific p-ranges, later verified and extended by Dai (2024) for various α. Montes-Rodríguez and Virtanen (2024) determined the continuous spectrum on as [0, π] with spectral measure for x ∈ [0, π], leveraging the Mehler-Fock transform—a integral transform involving Legendre functions used in quantum mechanics for conical potentials—to derive these properties.[15]Applications and Uses
In Approximation Theory
The Hilbert matrix plays a central role in the method of moments for polynomial least-squares approximation on the unit interval [0,1]. In this framework, one seeks a polynomial that approximates a given continuous function by minimizing the error . The normal equations for the coefficients take the form , where represents the -th moment of , and is the Hilbert matrix with entries . This setup arises because the monomials form the basis, and the Hilbert matrix captures their Gram matrix under the inner product.[16][17] In the theory of orthogonal polynomials on [0,1] with respect to the Lebesgue measure, the Hilbert matrix facilitates the computation of expansion coefficients for function approximations. The matrix serves as the Gram matrix for the monomial basis, and its inversion yields the transformation coefficients that express the monic orthogonal polynomials (shifted Legendre polynomials) in terms of powers of . These coefficients enable the projection of onto the orthogonal basis, where the expansion is and , avoiding the direct solution of ill-conditioned systems for higher-degree approximations. The explicit inverse of the Hilbert matrix, as derived in foundational studies, supports this coefficient computation analytically for moderate .[18][19] Hilbert's original 1894 work leverages the Hilbert matrix in analyzing the L^2 norms of non-constant polynomials on symmetric intervals . He demonstrated that if the interval length , the infimum of over such non-zero polynomials is zero, implying arbitrarily good approximations to zero in the L^2 sense (and thus density of integer polynomials in ). This result, established through determinants and asymptotic properties of finite Hilbert matrices, highlights the matrix's role in quantifying how integer constraints enable dense approximation in .[20] Due to the Hilbert matrix's extreme ill-conditioning—its condition number grows as approximately for dimension —practical applications in these approximation problems are limited for large . Exact solutions to or related inversions demand symbolic computation or arbitrary-precision arithmetic, as floating-point methods introduce errors that dominate the approximation accuracy beyond . This constraint underscores the need for alternative bases, like orthogonal polynomials, to mitigate numerical instability while preserving theoretical insights.[21][22]In Numerical Analysis
The Hilbert matrix is widely used as a benchmark problem in numerical analysis to test the stability and accuracy of algorithms for solving linear systems, owing to its extreme ill-conditioning that amplifies rounding errors in finite-precision computations. For example, Gaussian elimination applied to Hilbert systems of order greater than 10 typically fails to produce reliable solutions, as the process propagates floating-point errors, resulting in relative residuals that can exceed machine epsilon by orders of magnitude.[23][14] This ill-conditioning is quantified by the matrix's 2-norm condition number, which grows exponentially as , reaching about for ; such a large value means that perturbations on the order of machine precision (around in double precision) can lead to solutions accurate to only a few digits.[23][24] In practice, software libraries incorporate the Hilbert matrix for validation of numerical routines: MATLAB'shilb function generates it to assess direct methods like LU decomposition, where computed factors deviate substantially from exact ones for moderate , while SciPy's scipy.linalg.hilbert in Python serves similar purposes for iterative methods such as GMRES, highlighting convergence issues under ill-conditioning.[14][25]
To address these numerical challenges, techniques such as diagonal scaling to equilibrate the matrix, preconditioning via rank-revealing decompositions or approximate inverses (e.g., using the Hilbert matrix's integer inverse for ), and high-precision arithmetic have been developed to achieve inverse-equivalent accuracy independent of the condition number.[26] In recent machine learning contexts, the Hilbert matrix analogs ill-conditioned kernel Gram matrices arising in reproducing kernel Hilbert spaces, aiding the study of regularization in support vector machines and Gaussian processes.[27]