Recent from talks
Contribute something
Nothing was collected or created yet.
Hilbert transform
View on WikipediaIn mathematics and signal processing, the Hilbert transform is a specific singular integral that takes a function, u(t) of a real variable and produces another function of a real variable H(u)(t). The Hilbert transform is given by the Cauchy principal value of the convolution with the function (see § Definition). The Hilbert transform has a particularly simple representation in the frequency domain: It imparts a phase shift of ±90° (π/2 radians) to every frequency component of a function, the sign of the shift depending on the sign of the frequency (see § Relationship with the Fourier transform). The Hilbert transform is important in signal processing, where it is a component of the analytic representation of a real-valued signal u(t). The Hilbert transform was first introduced by David Hilbert in this setting, to solve a special case of the Riemann–Hilbert problem for analytic functions.
Definition
[edit]The Hilbert transform of u can be thought of as the convolution of u(t) with the function h(t) = 1/πt, known as the Cauchy kernel. Because 1/t is not integrable across t = 0, the integral defining the convolution does not always converge. Instead, the Hilbert transform is defined using the Cauchy principal value (denoted here by p.v.). Explicitly, the Hilbert transform of a function (or signal) u(t) is given by
provided this integral exists as a principal value. This is precisely the convolution of u with the tempered distribution p.v. 1/πt.[1] Alternatively, by changing variables, the principal-value integral can be written explicitly[2] as
When the Hilbert transform is applied twice in succession to a function u, the result is
provided the integrals defining both iterations converge in a suitable sense. In particular, the inverse transform is . This fact can most easily be seen by considering the effect of the Hilbert transform on the Fourier transform of u(t) (see § Relationship with the Fourier transform below).
For an analytic function in the upper half-plane, the Hilbert transform describes the relationship between the real part and the imaginary part of the boundary values. That is, if f(z) is analytic in the upper half complex plane {z : Im{z} > 0}, and u(t) = Re{f (t + 0·i)}, then Im{f(t + 0·i)} = H(u)(t) up to an additive constant, provided this Hilbert transform exists.
Notation
[edit]In signal processing the Hilbert transform of u(t) is commonly denoted by .[3] However, in mathematics, this notation is already extensively used to denote the Fourier transform of u(t).[4] Occasionally, the Hilbert transform may be denoted by . Furthermore, many sources define the Hilbert transform as the negative of the one defined here.[5]
History
[edit]The Hilbert transform arose in Hilbert's 1905 work on a problem Riemann posed concerning analytic functions,[6][7] which has come to be known as the Riemann–Hilbert problem. Hilbert's work was mainly concerned with the Hilbert transform for functions defined on the circle.[8][9] Some of his earlier work related to the Discrete Hilbert Transform dates back to lectures he gave in Göttingen. The results were later published by Hermann Weyl in his dissertation.[10] Schur improved Hilbert's results about the discrete Hilbert transform and extended them to the integral case.[11] These results were restricted to the spaces L2 and ℓ2. In 1928, Marcel Riesz proved that the Hilbert transform can be defined for u in (Lp space) for 1 < p < ∞, that the Hilbert transform is a bounded operator on for 1 < p < ∞, and that similar results hold for the Hilbert transform on the circle as well as the discrete Hilbert transform.[12] The Hilbert transform was a motivating example for Antoni Zygmund and Alberto Calderón during their study of singular integrals.[13] Their investigations have played a fundamental role in modern harmonic analysis. Various generalizations of the Hilbert transform, such as the bilinear and trilinear Hilbert transforms are still active areas of research today.
Relationship with the Fourier transform
[edit]The Hilbert transform is a multiplier operator.[14] The multiplier of H is σH(ω) = −i sgn(ω), where sgn is the signum function. Therefore:
where denotes the Fourier transform. Since sgn(x) = sgn(2πx), it follows that this result applies to the three common definitions of .
By Euler's formula,
Therefore, H(u)(t) has the effect of shifting the phase of the negative frequency components of u(t) by +90° (π⁄2 radians) and the phase of the positive frequency components by −90°, and i·H(u)(t) has the effect of restoring the positive frequency components while shifting the negative frequency ones an additional +90°, resulting in their negation (i.e., a multiplication by −1).
When the Hilbert transform is applied twice, the phase of the negative and positive frequency components of u(t) are respectively shifted by +180° and −180°, which are equivalent amounts. The signal is negated; i.e., H(H(u)) = −u, because
Table of selected Hilbert transforms
[edit]In the following table, the frequency parameter is real.
| Signal |
Hilbert transform[fn 1] |
|---|---|
| [fn 2] |
|
| [fn 2] |
|
|
| |
|
| |
(see Dawson function) | |
| Sinc function |
|
| Dirac delta function |
|
| Characteristic function |
Notes
- ^ Some authors (e.g., Bracewell) use our −H as their definition of the forward transform. A consequence is that the right column of this table would be negated.
- ^ a b The Hilbert transform of the sin and cos functions can be defined by taking the principal value of the integral at infinity. This definition agrees with the result of defining the Hilbert transform distributionally.
An extensive table of Hilbert transforms is available.[15] Note that the Hilbert transform of a constant is zero.
Domain of definition
[edit]It is by no means obvious that the Hilbert transform is well-defined at all, as the improper integral defining it must converge in a suitable sense. However, the Hilbert transform is well-defined for a broad class of functions, namely those in for 1 < p < ∞.
More precisely, if u is in for 1 < p < ∞, then the limit defining the improper integral
exists for almost every t. The limit function is also in and is in fact the limit in the mean of the improper integral as well. That is,
as ε → 0 in the Lp norm, as well as pointwise almost everywhere, by the Titchmarsh theorem.[16]
In the case p = 1, the Hilbert transform still converges pointwise almost everywhere, but may itself fail to be integrable, even locally.[17] In particular, convergence in the mean does not in general happen in this case. The Hilbert transform of an L1 function does converge, however, in L1-weak, and the Hilbert transform is a bounded operator from L1 to L1,w.[18] (In particular, since the Hilbert transform is also a multiplier operator on L2, Marcinkiewicz interpolation and a duality argument furnishes an alternative proof that H is bounded on Lp.)
Properties
[edit]Boundedness
[edit]If 1 < p < ∞, then the Hilbert transform on is a bounded linear operator, meaning that there exists a constant Cp such that
for all .[19]
The best constant is given by[20]
An easy way to find the best for being a power of 2 is through the so-called Cotlar's identity that for all real valued f. The same best constants hold for the periodic Hilbert transform.
The boundedness of the Hilbert transform implies the convergence of the symmetric partial sum operator
to f in .[21]
Anti-self adjointness
[edit]The Hilbert transform is an anti-self adjoint operator relative to the duality pairing between and the dual space , where p and q are Hölder conjugates and 1 < p, q < ∞. Symbolically,
for and .[22]
Inverse transform
[edit]The Hilbert transform is an anti-involution,[23] meaning that
provided each transform is well-defined. Since H preserves the space , this implies in particular that the Hilbert transform is invertible on , and that
Complex structure
[edit]Because H2 = −I ("I" is the identity operator) on the real Banach space of real-valued functions in , the Hilbert transform defines a linear complex structure on this Banach space. In particular, when p = 2, the Hilbert transform gives the Hilbert space of real-valued functions in the structure of a complex Hilbert space.
The (complex) eigenstates of the Hilbert transform admit representations as holomorphic functions in the upper and lower half-planes in the Hardy space H2 by the Paley–Wiener theorem.
Differentiation
[edit]Formally, the derivative of the Hilbert transform is the Hilbert transform of the derivative, i.e. these two linear operators commute:
Iterating this identity,
This is rigorously true as stated provided u and its first k derivatives belong to .[24] One can check this easily in the frequency domain, where differentiation becomes multiplication by ω.
Convolutions
[edit]The Hilbert transform can formally be realized as a convolution with the tempered distribution[25]
Thus formally,
However, a priori this may only be defined for u a distribution of compact support. It is possible to work somewhat rigorously with this since compactly supported functions (which are distributions a fortiori) are dense in Lp. Alternatively, one may use the fact that h(t) is the distributional derivative of the function log|t|/π; to wit
For most operational purposes the Hilbert transform can be treated as a convolution. For example, in a formal sense, the Hilbert transform of a convolution is the convolution of the Hilbert transform applied on only one of either of the factors:
This is rigorously true if u and v are compactly supported distributions since, in that case,
By passing to an appropriate limit, it is thus also true if u ∈ Lp and v ∈ Lq provided that
from a theorem due to Titchmarsh.[26]
Invariance
[edit]The Hilbert transform has the following invariance properties on .
- It commutes with translations. That is, it commutes with the operators Ta f(x) = f(x + a) for all a in
- It commutes with positive dilations. That is it commutes with the operators Mλ f (x) = f (λ x) for all λ > 0.
- It anticommutes with the reflection R f (x) = f (−x).
Up to a multiplicative constant, the Hilbert transform is the only bounded operator on L2 with these properties.[27]
In fact there is a wider set of operators that commute with the Hilbert transform. The group acts by unitary operators Ug on the space by the formula
This unitary representation is an example of a principal series representation of In this case it is reducible, splitting as the orthogonal sum of two invariant subspaces, Hardy space and its conjugate. These are the spaces of L2 boundary values of holomorphic functions on the upper and lower halfplanes. and its conjugate consist of exactly those L2 functions with Fourier transforms vanishing on the negative and positive parts of the real axis respectively. Since the Hilbert transform is equal to H = −i (2P − I), with P being the orthogonal projection from onto and I the identity operator, it follows that and its orthogonal complement are eigenspaces of H for the eigenvalues ±i. In other words, H commutes with the operators Ug. The restrictions of the operators Ug to and its conjugate give irreducible representations of – the so-called limit of discrete series representations.[28]
Extending the domain of definition
[edit]Hilbert transform of distributions
[edit]It is further possible to extend the Hilbert transform to certain spaces of distributions (Pandey 1996, Chapter 3). Since the Hilbert transform commutes with differentiation, and is a bounded operator on Lp, H restricts to give a continuous transform on the inverse limit of Sobolev spaces:
The Hilbert transform can then be defined on the dual space of , denoted , consisting of Lp distributions. This is accomplished by the duality pairing:
For , define:
It is possible to define the Hilbert transform on the space of tempered distributions as well by an approach due to Gel'fand and Shilov,[29] but considerably more care is needed because of the singularity in the integral.
Hilbert transform of bounded functions
[edit]The Hilbert transform can be defined for functions in as well, but it requires some modifications and caveats. Properly understood, the Hilbert transform maps to the Banach space of bounded mean oscillation (BMO) classes.
Interpreted naïvely, the Hilbert transform of a bounded function is clearly ill-defined. For instance, with u = sgn(x), the integral defining H(u) diverges almost everywhere to ±∞. To alleviate such difficulties, the Hilbert transform of an L∞ function is therefore defined by the following regularized form of the integral
where as above h(x) = 1/πx and
The modified transform H agrees with the original transform up to an additive constant on functions of compact support from a general result by Calderón and Zygmund.[30] Furthermore, the resulting integral converges pointwise almost everywhere, and with respect to the BMO norm, to a function of bounded mean oscillation.
A deep result of Fefferman's work[31] is that a function is of bounded mean oscillation if and only if it has the form f + H(g) for some .
Conjugate functions
[edit]The Hilbert transform can be understood in terms of a pair of functions f(x) and g(x) such that the function is the boundary value of a holomorphic function F(z) in the upper half-plane.[32] Under these circumstances, if f and g are sufficiently integrable, then one is the Hilbert transform of the other.
Suppose that Then, by the theory of the Poisson integral, f admits a unique harmonic extension into the upper half-plane, and this extension is given by
which is the convolution of f with the Poisson kernel
Furthermore, there is a unique harmonic function v defined in the upper half-plane such that F(z) = u(z) + i v(z) is holomorphic and
This harmonic function is obtained from f by taking a convolution with the conjugate Poisson kernel
Thus
Indeed, the real and imaginary parts of the Cauchy kernel are
so that F = u + i v is holomorphic by Cauchy's integral formula.
The function v obtained from u in this way is called the harmonic conjugate of u. The (non-tangential) boundary limit of v(x,y) as y → 0 is the Hilbert transform of f. Thus, succinctly,
Titchmarsh's theorem
[edit]Titchmarsh's theorem (named for E. C. Titchmarsh who included it in his 1937 work) makes precise the relationship between the boundary values of holomorphic functions in the upper half-plane and the Hilbert transform.[33] It gives necessary and sufficient conditions for a complex-valued square-integrable function F(x) on the real line to be the boundary value of a function in the Hardy space H2(U) of holomorphic functions in the upper half-plane U.
The theorem states that the following conditions for a complex-valued square-integrable function are equivalent:
- F(x) is the limit as z → x of a holomorphic function F(z) in the upper half-plane such that
- The real and imaginary parts of F(x) are Hilbert transforms of each other.
- The Fourier transform vanishes for x < 0.
A weaker result is true for functions of class Lp for p > 1.[34] Specifically, if F(z) is a holomorphic function such that
for all y, then there is a complex-valued function F(x) in such that F(x + i y) → F(x) in the Lp norm as y → 0 (as well as holding pointwise almost everywhere). Furthermore,
where f is a real-valued function in and g is the Hilbert transform (of class Lp) of f.
This is not true in the case p = 1. In fact, the Hilbert transform of an L1 function f need not converge in the mean to another L1 function. Nevertheless,[35] the Hilbert transform of f does converge almost everywhere to a finite function g such that
This result is directly analogous to one by Andrey Kolmogorov for Hardy functions in the disc.[36] Although usually called Titchmarsh's theorem, the result aggregates much work of others, including Hardy, Paley and Wiener (see Paley–Wiener theorem), as well as work by Riesz, Hille, and Tamarkin[37]
Riemann–Hilbert problem
[edit]One form of the Riemann–Hilbert problem seeks to identify pairs of functions F+ and F− such that F+ is holomorphic on the upper half-plane and F− is holomorphic on the lower half-plane, such that for x along the real axis,
where f(x) is some given real-valued function of . The left-hand side of this equation may be understood either as the difference of the limits of F± from the appropriate half-planes, or as a hyperfunction distribution. Two functions of this form are a solution of the Riemann–Hilbert problem.
Formally, if F± solve the Riemann–Hilbert problem
then the Hilbert transform of f(x) is given by[38]
Hilbert transform on the circle
[edit]For a periodic function f the circular Hilbert transform is defined:
The circular Hilbert transform is used in giving a characterization of Hardy space and in the study of the conjugate function in Fourier series. The kernel, is known as the Hilbert kernel since it was in this form the Hilbert transform was originally studied.[8]
The Hilbert kernel (for the circular Hilbert transform) can be obtained by making the Cauchy kernel 1⁄x periodic. More precisely, for x ≠ 0
Many results about the circular Hilbert transform may be derived from the corresponding results for the Hilbert transform from this correspondence.
Another more direct connection is provided by the Cayley transform C(x) = (x – i) / (x + i), which carries the real line onto the circle and the upper half plane onto the unit disk. It induces a unitary map
of L2(T) onto The operator U carries the Hardy space H2(T) onto the Hardy space .[39]
Hilbert transform in signal processing
[edit]Bedrosian's theorem
[edit]Bedrosian's theorem states that the Hilbert transform of the product of a low-pass and a high-pass signal with non-overlapping spectra is given by the product of the low-pass signal and the Hilbert transform of the high-pass signal, or
where fLP and fHP are the low- and high-pass signals respectively.[40] A category of communication signals to which this applies is called the narrowband signal model. A member of that category is amplitude modulation of a high-frequency sinusoidal "carrier":
where um(t) is the narrow bandwidth "message" waveform, such as voice or music. Then by Bedrosian's theorem:[41]
Analytic representation
[edit]A specific type of conjugate function is:
known as the analytic representation of The name reflects its mathematical tractability, due largely to Euler's formula. Applying Bedrosian's theorem to the narrowband model, the analytic representation is:[42]
| Eq.1 |
A Fourier transform property indicates that this complex heterodyne operation can shift all the negative frequency components of um(t) above 0 Hz. In that case, the imaginary part of the result is a Hilbert transform of the real part. This is an indirect way to produce Hilbert transforms.
Angle (phase/frequency) modulation
[edit]The form:[43]
is called angle modulation, which includes both phase modulation and frequency modulation. The instantaneous frequency is For sufficiently large ω, compared to :
and:
Single sideband modulation (SSB)
[edit]When um(t) in Eq.1 is also an analytic representation (of a message waveform), that is:
the result is single-sideband modulation:
whose transmitted component is:[44][45]
Causality
[edit]The function presents two causality-based challenges to practical implementation in a convolution (in addition to its undefined value at 0):
- Its duration is infinite (technically infinite support). Finite-length windowing reduces the effective frequency range of the transform; shorter windows result in greater losses at low and high frequencies. See also quadrature filter.
- It is a non-causal filter. So a delayed version, is required. The corresponding output is subsequently delayed by When creating the imaginary part of an analytic signal, the source (real part) must also be delayed by .
Discrete Hilbert transform
[edit]




For a discrete function, with discrete-time Fourier transform (DTFT), , and discrete Hilbert transform the DTFT of in the region −π < ω < π is given by:
The inverse DTFT, using the convolution theorem, is:[46][47]
where
which is an infinite impulse response (IIR).
Practical considerations[48]
Method 1: Direct convolution of streaming data with an FIR approximation of which we will designate by Examples of truncated are shown in figures 1 and 2. Fig 1 has an odd number of anti-symmetric coefficients and is called Type III.[49] This type inherently exhibits responses of zero magnitude at frequencies 0 and Nyquist, resulting in a bandpass filter shape.[50][51] A Type IV design (even number of anti-symmetric coefficients) is shown in Fig 2.[52][53] It has a highpass frequency response.[54] Type III is the usual choice.[55][56] for these reasons:
- A typical (i.e. properly filtered and sampled) sequence has no useful components at the Nyquist frequency.
- The Type IV impulse response requires a sample shift in the sequence. That causes the zero-valued coefficients to become non-zero, as seen in Figure 2. So a Type III design is potentially twice as efficient as Type IV.
- The group delay of a Type III design is an integer number of samples, which facilitates aligning with to create an analytic signal. The group delay of Type IV is halfway between two samples.
The abrupt truncation of creates a rippling (Gibbs effect) of the flat frequency response. That can be mitigated by use of a window function to taper to zero.[57]
Method 2: Piecewise convolution. It is well known that direct convolution is computationally much more intensive than methods like overlap-save that give access to the efficiencies of the Fast Fourier transform via the convolution theorem.[58] Specifically, the discrete Fourier transform (DFT) of a segment of is multiplied pointwise with a DFT of the sequence. An inverse DFT is done on the product, and the transient artifacts at the leading and trailing edges of the segment are discarded. Over-lapping input segments prevent gaps in the output stream. An equivalent time domain description is that segments of length (an arbitrary parameter) are convolved with the periodic function:
When the duration of non-zero values of is the output sequence includes samples of outputs are discarded from each block of and the input blocks are overlapped by that amount to prevent gaps.
Method 3: Same as method 2, except the DFT of is replaced by samples of the distribution (whose real and imaginary components are all just or ) That convolves with a periodic summation:[A]
for some arbitrary parameter, is not an FIR, so the edge effects extend throughout the entire transform. Deciding what to delete and the corresponding amount of overlap is an application-dependent design issue.
Fig 3 depicts the difference between methods 2 and 3. Only half of the antisymmetric impulse response is shown, and only the non-zero coefficients. The blue graph corresponds to method 2 where is truncated by a rectangular window function, rather than tapered. It is generated by a Matlab function, hilb(65). Its transient effects are exactly known and readily discarded. The frequency response, which is determined by the function argument, is the only application-dependent design issue.
The red graph is corresponding to method 3. It is the inverse DFT of the distribution. Specifically, it is the function that is convolved with a segment of by the MATLAB function, hilbert(u,512).[61] The real part of the output sequence is the original input sequence, so that the complex output is an analytic representation of
When the input is a segment of a pure cosine, the resulting convolution for two different values of is depicted in Fig 4 (red and blue plots). Edge effects prevent the result from being a pure sine function (green plot). Since is not an FIR sequence, the theoretical extent of the effects is the entire output sequence. But the differences from a sine function diminish with distance from the edges. Parameter is the output sequence length. If it exceeds the length of the input sequence, the input is modified by appending zero-valued elements. In most cases, that reduces the magnitude of the edge distortions. But their duration is dominated by the inherent rise and fall times of the impulse response.
Fig 5 is an example of piecewise convolution, using both methods 2 (in blue) and 3 (red dots). A sine function is created by computing the Discrete Hilbert transform of a cosine function, which was processed in four overlapping segments, and pieced back together. As the FIR result (blue) shows, the distortions apparent in the IIR result (red) are not caused by the difference between and (green and red in Fig 3). The fact that is tapered (windowed) is actually helpful in this context. The real problem is that it's not windowed enough. Effectively, whereas the overlap-save method needs
Number-theoretic Hilbert transform
[edit]The number theoretic Hilbert transform is an extension[62] of the discrete Hilbert transform to integers modulo an appropriate prime number. In this it follows the generalization of discrete Fourier transform to number theoretic transforms. The number theoretic Hilbert transform can be used to generate sets of orthogonal discrete sequences.[63]
See also
[edit]Notes
[edit]Page citations
[edit]- ^ Due to Schwartz 1950; see Pandey 1996, Chapter 3.
- ^ Zygmund 1968, §XVI.1.
- ^ E.g., Brandwood 2003, p. 87.
- ^ E.g., Stein & Weiss 1971.
- ^ E.g., Bracewell 2000, p. 359.
- ^ Kress 1989.
- ^ Bitsadze 2001.
- ^ a b Khvedelidze 2001.
- ^ Hilbert 1953.
- ^ Hardy, Littlewood & Pólya 1952, §9.1.
- ^ Hardy, Littlewood & Pólya 1952, §9.2.
- ^ Riesz 1928.
- ^ Calderón & Zygmund 1952.
- ^ Duoandikoetxea 2000, Chapter 3.
- ^ King 2009b.
- ^ Titchmarsh 1948, Chapter 5.
- ^ Titchmarsh 1948, §5.14.
- ^ Stein & Weiss 1971, Lemma V.2.8.
- ^ This theorem is due to Riesz 1928, VII; see also Titchmarsh 1948, Theorem 101.
- ^ This result is due to Pichorides 1972; see also Grafakos 2004, Remark 4.1.8.
- ^ See for example Duoandikoetxea 2000, p. 59.
- ^ Titchmarsh 1948, Theorem 102.
- ^ Titchmarsh 1948, p. 120.
- ^ Pandey 1996, §3.3.
- ^ Duistermaat & Kolk 2010, p. 211.
- ^ Titchmarsh 1948, Theorem 104.
- ^ Stein 1970, §III.1.
- ^ See Bargmann 1947, Lang 1985, and Sugiura 1990.
- ^ Gel'fand & Shilov 1968.
- ^ Calderón & Zygmund 1952; see Fefferman 1971.
- ^ Fefferman 1971; Fefferman & Stein 1972
- ^ Titchmarsh 1948, Chapter V.
- ^ Titchmarsh 1948, Theorem 95.
- ^ Titchmarsh 1948, Theorem 103.
- ^ Titchmarsh 1948, Theorem 105.
- ^ Duren 1970, Theorem 4.2.
- ^ see King 2009a, § 4.22.
- ^ Pandey 1996, Chapter 2.
- ^ Rosenblum & Rovnyak 1997, p. 92.
- ^ Schreier & Scharf 2010, 14.
- ^ Bedrosian 1962.
- ^ Osgood, p. 320
- ^ Osgood, p. 320
- ^ Franks 1969, p. 88
- ^ Tretter 1995, p. 80 (7.9)
- ^ Carrick, Jaeger & harris 2011, p. 2
- ^ Rabiner & Gold 1975, p. 71 (Eq 2.195)
- ^ Oppenheim, Schafer & Buck 1999, p. 794-795
- ^ Isukapalli, p. 14
- ^ Isukapalli, p. 18
- ^ Rabiner & Gold 1975, p. 172 (Fig 3.74)
- ^ Isukapalli, p. 15
- ^ Rabiner & Gold 1975, p. 173 (Fig 3.75)
- ^ Isukapalli, p. 18
- ^ Carrick, Jaeger & harris 2011, p. 3
- ^ Rabiner & Gold 1975, p. 175
- ^ Carrick, Jaeger & harris 2011, p. 3
- ^ Rabiner & Gold 1975, p. 59 (2.163)
- ^ Johansson, p. 24
- ^ Johansson, p. 25
- ^ MathWorks. "hilbert – Discrete-time analytic signal using Hilbert transform". MATLAB Signal Processing Toolbox Documentation. Retrieved 2021-05-06.
- ^ Kak 1970.
- ^ Kak 2014.
References
[edit]- Bargmann, V. (1947). "Irreducible unitary representations of the Lorentz group". Ann. of Math. 48 (3): 568–640. doi:10.2307/1969129. JSTOR 1969129.
- Bedrosian, E. (December 1962). A product theorem for Hilbert transforms (PDF) (Report). Rand Corporation. RM-3439-PR.
- Bitsadze, A. V. (2001) [1994], "Boundary value problems of analytic function theory", Encyclopedia of Mathematics, EMS Press
- Bracewell, R. (2000). The Fourier Transform and Its Applications (3rd ed.). McGraw–Hill. ISBN 0-07-116043-4.
- Brandwood, David (2003). Fourier Transforms in Radar and Signal Processing. Boston: Artech House. ISBN 9781580531740.
- Calderón, A. P.; Zygmund, A. (1952). "On the existence of certain singular integrals". Acta Mathematica. 88 (1): 85–139. doi:10.1007/BF02392130.
- Carrick, Matt; Jaeger, Doug; harris, fred (2011). Design And Application Of A Hilbert Transformer In A Digital Receiver (PDF). Chantilly, VA: Proceedings of the SDR 11 Technical Conference and Product Exposition, Wireless Innovation Forum. Retrieved 2024-06-05.
- Duoandikoetxea, J. (2000). Fourier Analysis. American Mathematical Society. ISBN 0-8218-2172-5.
- Duistermaat, J. J.; Kolk, J. A. C. (2010). Distributions. Birkhäuser. doi:10.1007/978-0-8176-4675-2. ISBN 978-0-8176-4672-1.
- Duren, P. (1970). Theory of H^p Spaces. New York, NY: Academic Press.
- Fefferman, C. (1971). "Characterizations of bounded mean oscillation". Bulletin of the American Mathematical Society. 77 (4): 587–588. doi:10.1090/S0002-9904-1971-12763-5. MR 0280994.
- Fefferman, C.; Stein, E. M. (1972). "H^p spaces of several variables". Acta Mathematica. 129: 137–193. doi:10.1007/BF02392215. MR 0447953.
- Franks, L.E. (September 1969). Thomas Kailath (ed.). Signal Theory. Information theory. Englewood Cliffs, NJ: Prentice Hall. ISBN 0138100772.
- Gel'fand, I. M.; Shilov, G. E. (1968). Generalized Functions. Vol. 2. Academic Press. pp. 153–154. ISBN 0-12-279502-4.
- Grafakos, Loukas (2004). Classical and Modern Fourier Analysis. Pearson Education. pp. 253–257. ISBN 0-13-035399-X.
- Hardy, G. H.; Littlewood, J. E.; Pólya, G. (1952). Inequalities. Cambridge, UK: Cambridge University Press. ISBN 0-521-35880-9.
{{cite book}}: ISBN / Date incompatibility (help) - Hilbert, David (1953) [1912]. Grundzüge einer allgemeinen Theorie der linearen Integralgleichungen [Framework for a General Theory of Linear Integral Equations] (in German). Leipzig & Berlin, DE (1912); New York, NY (1953): B.G. Teubner (1912); Chelsea Pub. Co. (1953). ISBN 978-3-322-00681-3. OCLC 988251080. Retrieved 2020-12-18 – via archive.org.
{{cite book}}: ISBN / Date incompatibility (help)CS1 maint: location (link) - Isukapalli, Yogananda. "Types of linear phase FIR filters" (PDF). p. 18. Retrieved 2024-06-08.
- Johansson, Mathias. "The Hilbert transform, Masters Thesis" (PDF). Archived from the original (PDF) on 2012-02-05.; also http://www.fuchs-braun.com/media/d9140c7b3d5004fbffff8007fffffff0.pdf Archived 2021-05-01 at the Wayback Machine
- Kak, Subhash (1970). "The discrete Hilbert transform". Proc. IEEE. 58 (4): 585–586. doi:10.1109/PROC.1970.7696.
- Kak, Subhash (2014). "Number theoretic Hilbert transform". Circuits, Systems and Signal Processing. 33 (8): 2539–2548. arXiv:1308.1688. doi:10.1007/s00034-014-9759-8. S2CID 21226699.
- Khvedelidze, B. V. (2001) [1994], "Hilbert transform", Encyclopedia of Mathematics, EMS Press
- King, Frederick W. (2009a). Hilbert Transforms. Vol. 1. Cambridge, UK: Cambridge University Press.
- King, Frederick W. (2009b). Hilbert Transforms. Vol. 2. Cambridge, UK: Cambridge University Press. p. 453. ISBN 978-0-521-51720-1.
- Kress, Rainer (1989). Linear Integral Equations. New York, NY: Springer-Verlag. p. 91. ISBN 3-540-50616-0.
- Lang, Serge (1985). SL(2,). Graduate Texts in Mathematics. Vol. 105. New York, NY: Springer-Verlag. ISBN 0-387-96198-4.
- Oppenheim, Alan V.; Schafer, Ronald W.; Buck, John R. (1999). "11.4.1". Discrete-time signal processing (2nd ed.). Upper Saddle River, N.J.: Prentice Hall. ISBN 0-13-754920-2.
The exactness of the phase of type III and IV FIR systems is a compelling motivation for their use in approximating Hilbert transformers.
- Osgood, Brad, The Fourier Transform and its Applications (PDF), Stanford University, retrieved 2021-04-30
- Pandey, J. N. (1996). The Hilbert transform of Schwartz distributions and applications. Wiley-Interscience. ISBN 0-471-03373-1.
- Pichorides, S. (1972). "On the best value of the constants in the theorems of Riesz, Zygmund, and Kolmogorov". Studia Mathematica. 44 (2): 165–179. doi:10.4064/sm-44-2-165-179.
- Rabiner, Lawrence R.; Gold, Bernard (1975). Theory and application of digital signal processing. Englewood Cliffs, N.J.: Prentice-Hall. ISBN 0-13-914101-4.
- Riesz, Marcel (1928). "Sur les fonctions conjuguées". Mathematische Zeitschrift (in French). 27 (1): 218–244. doi:10.1007/BF01171098. S2CID 123261514.
- Rosenblum, Marvin; Rovnyak, James (1997). Hardy classes and operator theory. Dover. ISBN 0-486-69536-0.
- Schwartz, Laurent (1950). Théorie des distributions. Paris, FR: Hermann.
- Schreier, P.; Scharf, L. (2010). Statistical signal processing of complex-valued data: The theory of improper and noncircular signals. Cambridge, UK: Cambridge University Press.
- Smith, J. O. (2007). "Analytic Signals and Hilbert Transform Filters, in Mathematics of the Discrete Fourier Transform (DFT) with Audio Applications" (2nd ed.). Retrieved 2021-04-29.; also https://www.dsprelated.com/freebooks/mdft/Analytic_Signals_Hilbert_Transform.html
- Stein, Elias (1970). Singular integrals and differentiability properties of functions. Princeton University Press. ISBN 0-691-08079-8.
- Stein, Elias; Weiss, Guido (1971). Introduction to Fourier Analysis on Euclidean Spaces. Princeton University Press. ISBN 0-691-08078-X.
- Sugiura, Mitsuo (1990). Unitary Representations and Harmonic Analysis: An Introduction. North-Holland Mathematical Library. Vol. 44 (2nd ed.). Elsevier. ISBN 0444885935.
- Titchmarsh, E. (1986) [1948]. Introduction to the theory of Fourier integrals (2nd ed.). Oxford, UK: Clarendon Press. ISBN 978-0-8284-0324-5.
- Tretter, Steven A. (1995). R.W.Lucky (ed.). Communication System Design Using DSP Algorithms. New York: Springer. ISBN 0306450321.
- Zygmund, Antoni (1988) [1968]. Trigonometric Series (2nd ed.). Cambridge, UK: Cambridge University Press. ISBN 978-0-521-35885-9.
Further reading
[edit]- Benedetto, John J. (1996). Harmonic Analysis and its Applications. Boca Raton, FL: CRC Press. ISBN 0849378796.
- Carlson; Crilly & Rutledge (2002). Communication Systems (4th ed.). McGraw-Hill. ISBN 0-07-011127-8.
- Gold, B.; Oppenheim, A. V.; Rader, C. M. (1969). "Theory and Implementation of the Discrete Hilbert Transform". Proceedings of the 1969 Polytechnic Institute of Brooklyn Symposium. New York.
- Grafakos, Loukas (1994). "An elementary proof of the square summability of the discrete Hilbert transform". American Mathematical Monthly. 101 (5). Mathematical Association of America: 456–458. doi:10.2307/2974910. JSTOR 2974910.
- Titchmarsh, E. (1926). "Reciprocal formulae involving series and integrals". Mathematische Zeitschrift. 25 (1): 321–347. doi:10.1007/BF01283842. S2CID 186237099.
External links
[edit]- Derivation of the boundedness of the Hilbert transform
- Mathworld Hilbert transform — Contains a table of transforms
- Weisstein, Eric W. "Titchmarsh theorem". MathWorld.
- "GS256 Lecture 3: Hilbert Transformation" (PDF). Archived from the original (PDF) on 2012-02-27. an entry level introduction to Hilbert transformation.
Hilbert transform
View on GrokipediaMathematical Foundations
Definition
The Hilbert transform is a linear integral operator that acts on a real-valued function defined on the real line, producing another function interpreted as the imaginary part of the corresponding analytic signal. It is formally defined by the Cauchy principal value integral where . denotes the Cauchy principal value, ensuring the integral converges in a symmetric limiting sense around the singularity at .[1] This operator is central in signal processing and harmonic analysis, transforming the input function while preserving its energy in the sense for suitable functions.[4] In the frequency domain, the Hilbert transform corresponds to a phase shift of (or radians) for positive frequencies and (or radians) for negative frequencies, effectively converting sines to cosines and vice versa without altering amplitudes.[5] This property makes it a quadrature filter, useful for extracting envelope and instantaneous phase information from real signals. For a real-valued function , the analytic signal is then given by , where provides the imaginary component, suppressing negative frequency contributions to yield a complex representation concentrated in the positive frequency domain.[6] A simple illustration of the transform's behavior is its action on constant functions: the Hilbert transform of any constant is zero, as the principal value integral of over symmetric limits vanishes due to the odd integrand.[1] This result underscores the transform's insensitivity to DC components, aligning with its role in emphasizing oscillatory content.Notation and Conventions
The Hilbert transform operator is commonly denoted by , so that the transform of a function is written as , , or .[1][7] This notation emphasizes the operator's role as a linear map on suitable function spaces.[8] A distinction is drawn between the two-sided Hilbert transform, which operates symmetrically over the entire real line, and one-sided variants associated with analytic extensions to the upper or lower half-plane; the latter project onto positive or negative frequency components, respectively, and are often expressed using combinations like for the upper half-plane.[1][7] Sign conventions for the Hilbert transform vary across fields, with signal processing typically adopting a positive orientation that induces a counterclockwise phase shift of for positive frequencies (equivalent to multiplication by in the frequency domain).[1][7] In contrast, some mathematical treatments employ the opposite sign, yielding a clockwise phase shift and aligning with boundary values from the lower half-plane.[9] The prefactor in the principal value integral representation is standardly under the counterclockwise convention, though appears in literature favoring the clockwise alternative; this discrepancy arises from the choice of kernel sign and ensures consistency with the desired phase behavior.[1][7][9] In multivariable generalizations, such as the Riesz transforms that extend the Hilbert transform to higher dimensions, vector arguments are conventionally denoted in boldface (e.g., ) to distinguish them from scalar variables.[10]Initial Domain of Definition
The Hilbert transform is initially defined for functions in the Schwartz space , consisting of smooth functions that decay rapidly at infinity along with all their derivatives, ensuring the principal value integral converges absolutely.[2][11] For , the transform is given by where the Cauchy principal value addresses the singularity at in the kernel , taken as the limit [2][8] This formulation guarantees that remains well-defined and belongs to under these conditions.[2] The domain extends naturally to the Lebesgue space through the Plancherel theorem, which equates the norms of a function and its Fourier transform, allowing the Hilbert transform to be represented as multiplication by in the frequency domain.[8][2] This extension preserves the norm, making an isometry on and ensuring convergence in the sense for dense subsets like the Schwartz space.[8][11] Further, the Hilbert transform converges and maps to itself for , with the principal value integral providing pointwise almost everywhere convergence for functions in these spaces.[2][8] The boundedness on these spaces relies on the rapid decay and smoothness properties inherited from the Schwartz class, extended via density arguments.[2]Historical Context
Origins and Early Development
The Hilbert transform originated in David Hilbert's foundational work on linear integral equations, where he introduced it in 1905 to address the Dirichlet problem for the Laplace equation as part of solving the Riemann-Hilbert boundary value problem for analytic functions. In his paper "Grundzüge einer allgemeinen Theorie der linearen Integralgleichungen. Erste Mitteilung," published in the Nachrichten von der Gesellschaft der Wissenschaften zu Göttingen, Hilbert formulated the transform as a principal value integral to connect the real and imaginary parts of boundary values for analytic functions in the unit disk, enabling the solution of boundary value problems through integral representations.[12][13] This approach transformed the Dirichlet problem into a Fredholm-type integral equation, providing a method to recover the harmonic function inside the domain from its boundary data.[2] The concept drew early inspiration from late 19th-century developments in complex analysis, particularly Henri Poincaré's 1885 investigations into analytic functions of two variables and conformal mappings, which emphasized the role of conjugate harmonic functions and the Cauchy integral formula in representing solutions to boundary problems. Poincaré's work highlighted how the real and imaginary parts of analytic functions behave as conjugate harmonics, laying groundwork for integral operators that link boundary values to interior solutions, a theme Hilbert extended through his singular integral. These connections underscored the transform's utility in unifying complex analysis with integral methods for potential-theoretic problems. In potential theory, the Hilbert transform found initial applications in representing harmonic conjugates, allowing the construction of analytic functions from given real parts on the boundary, essential for solving elliptic partial differential equations (PDEs). Hilbert's framework addressed boundary value problems for such PDEs, including the representation of multi-variable functions as superpositions, aligning with the motivations of his 13th problem from the 1900 International Congress of Mathematicians, where he sought general methods for expressing solutions to higher-dimensional equations.[14] This integration of integral equations with PDE boundary conditions marked a pivotal step in early 20th-century mathematical physics, emphasizing the transform's role in rigorous solution theory without relying on series expansions.[13]Key Contributors and Milestones
In the early 20th century, building on David Hilbert's foundational 1905 work on boundary value problems for analytic functions, significant advancements in the theory of conjugate functions—closely related to the Hilbert transform—emerged through the efforts of key mathematicians. A pivotal contribution came from Edward Charles Titchmarsh in 1925, who established a theorem linking conjugate trigonometric integrals to Poisson integrals, providing deeper insights into the representation and properties of harmonic functions in the upper half-plane. This result, detailed in his paper on conjugate trigonometrical integrals, clarified the relationship between the Hilbert transform and boundary values of analytic functions, influencing subsequent studies in complex analysis. During the 1930s, Antoni Zygmund advanced the understanding of the Hilbert transform through his work on bounded analytic functions and their connections to Hardy spaces. In his seminal 1935 treatise Trigonometric Series, Zygmund explored the boundedness of conjugate functions in these spaces, laying groundwork for applications in Fourier analysis and establishing key inequalities that bounded the norms of Hilbert transforms on relevant function classes. The Riesz brothers, Frigyes and Marcel, made foundational contributions in the 1910s and 1920s to the study of maximal functions associated with analytic functions, which later connected the Hilbert transform to broader developments in harmonic analysis post-World War II. Marcel Riesz's 1928 proof of the L^p-boundedness (1 < p < ∞) of the conjugate function operator on the real line was particularly influential, providing a cornerstone for modern operator theory and maximal inequalities in harmonic analysis. A major milestone occurred in the 1950s with the recognition of the Hilbert transform as the prototypical singular integral operator within the Calderón-Zygmund theory. Alberto Calderón and Antoni Zygmund's collaborative work, including their 1952 paper on singular integrals, developed a general framework for the boundedness of such operators on L^p spaces, unifying the Hilbert transform with other Calderón-Zygmund kernels and enabling extensions to higher dimensions and more complex settings in analysis.Core Relationships and Computations
Connection to the Fourier Transform
The Fourier transform of a function is defined as with the convention extending naturally to via density of the Schwartz class and Plancherel's theorem, which establishes that the Fourier transform is a unitary operator on .[8] The Hilbert transform admits a simple representation in the frequency domain: for , where is the sign function defined by if , if , and . This multiplier property characterizes the Hilbert transform as a Fourier multiplier operator with symbol .[8] To derive this, recall that the Hilbert transform is a convolution , where in the principal value sense. By the convolution theorem for the Fourier transform, The Fourier transform of the kernel is computed in the sense of tempered distributions: , which follows from the distributional Fourier transform of the principal value distribution being , scaled by the factor . Plancherel's theorem then justifies the extension of this pointwise multiplier formula to all functions, as the operator norm of multiplication by is bounded by 1 on .[8][15] This frequency-domain representation implies that the Hilbert transform acts as a phase shifter: it multiplies the Fourier coefficients by for positive frequencies () and by for negative frequencies (), while preserving the magnitude spectrum since . Thus, rotates the phase of each frequency component by without altering amplitudes, a property central to its role in analytic signal construction.[8]Table of Selected Hilbert Transforms
The Hilbert transform provides explicit closed-form expressions for many common functions, facilitating both theoretical analysis and practical computations in signal processing and harmonic analysis. These transforms are typically derived using the Fourier domain representation, where the Fourier transform of the function is multiplied by before applying the inverse Fourier transform. The following table presents selected examples, focusing on the Heaviside step function, the Gaussian, and a representative power-related function such as the Cauchy distribution density. Each entry includes the input function , its Hilbert transform , a brief note on verification via the Fourier method, and a citation to a primary source.| Input function | Hilbert transform | Verification note | Citation |
|---|---|---|---|
| Heaviside step: | Obtained as the inverse Fourier transform of (principal value), yielding the logarithmic form. | https://doi.org/10.1017/CBO9780511721458 (King, 2009, Vol. 1, App. A) | |
| Gaussian: | , where (Dawson's function, related to the imaginary error function via ) | Obtained as the inverse Fourier transform of , resulting in the Dawson expression equivalent to error function forms. | https://doi.org/10.1016/S0096-3003(08)00578-X (Abdullah et al., 2009) |
| Cauchy density: | Obtained as the inverse Fourier transform of $-i \operatorname{sgn}(\omega) \cdot \pi e^{- | \omega |
