In mathematics, the convolution theorem states that under suitable conditions the Fourier transform of a convolution of two functions (or signals) is the product of their Fourier transforms. More generally, convolution in one domain (e.g., time domain) equals point-wise multiplication in the other domain (e.g., frequency domain). Other versions of the convolution theorem are applicable to various Fourier-related transforms.
where denotes the Fourier transform operator. The transform may be normalized in other ways, in which case constant scaling factors (typically or ) will appear in the convolution theorem below. The convolution of and is defined by:
In this context the asterisk denotes convolution, instead of standard multiplication. The tensor product symbol is sometimes used instead.
By a derivation similar to Eq.1, there is an analogous theorem for sequences, such as samples of two continuous functions, where now denotes the discrete-time Fourier transform (DTFT) operator. Consider two sequences and with transforms and :
and as defined above, are periodic, with a period of 1. Consider -periodic sequences and :
and
These functions occur as the result of sampling and at intervals of and performing an inverse discrete Fourier transform (DFT) on samples (see § Sampling the DTFT). The discrete convolution:
is also -periodic, and is called a periodic convolution. Redefining the operator as the -length DFT, the corresponding theorem is:[5][4]: p. 548
Eq.4a
And therefore:
Eq.4b
Under the right conditions, it is possible for this -length sequence to contain a distortion-free segment of a convolution. But when the non-zero portion of the or sequence is equal or longer than some distortion is inevitable. Such is the case when the sequence is obtained by directly sampling the DTFT of the infinitely long § Discrete Hilbert transformimpulse response.[A]
For and sequences whose non-zero duration is less than or equal to a final simplification is:
As a partial reciprocal, it has been shown [6]
that any linear transform that turns convolution into a product is the DFT (up to a permutation of coefficients).
Derivations of Eq.4
A time-domain derivation proceeds as follows:
A frequency-domain derivation follows from § Periodic data, which indicates that the DTFTs can be written as:
The product with is thereby reduced to a discrete-frequency function:
where the equivalence of and follows from § Sampling the DTFT. Therefore, the equivalence of (5a) and (5b) requires:
The convolution theorem extends to tempered distributions.
Here, is an arbitrary tempered distribution:
But must be "rapidly decreasing" towards and in order to guarantee the existence of both, convolution and multiplication product. Equivalently, if is a smooth "slowly growing" ordinary function, it guarantees the existence of both, multiplication and convolution product.[7][8][9]
In particular, every compactly supported tempered distribution, such as the Dirac delta, is "rapidly decreasing". Equivalently, bandlimited functions, such as the function that is constantly are smooth "slowly growing" ordinary functions. If, for example, is the Dirac comb both equations yield the Poisson summation formula and if, furthermore, is the Dirac delta then is constantly one and these equations yield the Dirac comb identity.
^
McGillem, Clare D.; Cooper, George R. (1984). Continuous and Discrete Signal and System Analysis (2 ed.). Holt, Rinehart and Winston. p. 118 (3–102). ISBN0-03-061703-0.
^ ab
Weisstein, Eric W. "Convolution Theorem". From MathWorld--A Wolfram Web Resource. Retrieved 8 February 2021.
Katznelson, Yitzhak (1976), An introduction to Harmonic Analysis, Dover, ISBN0-486-63331-4
Li, Bing; Babu, G. Jogesh (2019), "Convolution Theorem and Asymptotic Efficiency", A Graduate Course on Statistical Inference, New York: Springer, pp. 295–327, ISBN978-1-4939-9759-6
Crutchfield, Steve (October 9, 2010), "The Joy of Convolution", Johns Hopkins University, retrieved November 19, 2010
The convolution theorem is a fundamental principle in Fourier analysis that states the Fourier transform of the convolution of two functions equals the pointwise product of their individual Fourier transforms, and conversely, the inverse Fourier transform of the product of two Fourier transforms equals the convolution of the original functions.[1] This relationship, often expressed mathematically as F{f∗g}=F{f}⋅F{g} where f∗g(t)=∫−∞∞f(τ)g(t−τ)dτ, enables efficient computation of convolutions in the frequency domain, particularly using fast Fourier transform algorithms.[2]The theorem holds for functions in various domains, including time and spatial signals, and is commutative, meaning f∗g=g∗f.[2] It extends to multidimensional cases and discrete signals, where the discrete Fourier transform (DFT) replaces the continuous version, facilitating numerical implementations.[3] Originally derived in the context of integral transforms, the theorem's proof relies on the linearity and properties of the Fourier transform, such as the transform of a shifted function.[1]In applications, the convolution theorem is pivotal in signal processing for tasks like filtering, where convolving a signal with an impulse response in the time domain corresponds to multiplying their spectra in the frequency domain, simplifying deconvolution and system analysis.[2] It also underpins image processing techniques, such as blurring or template matching, by converting computationally intensive spatial convolutions into efficient frequency-domain multiplications.[3] More broadly, in machine learning, the theorem supports convolutional neural networks (CNNs) by enabling fast computation of large-kernel convolutions via the fast Fourier transform (FFT), reducing complexity from O(N2M2) to O(N2logN) for kernel size M and input size N.[3] These uses highlight the theorem's role in transforming complex integral operations into algebraic ones, advancing fields from physics to computer vision.[1]
Core statements
Continuous aperiodic case
In the continuous aperiodic case, the convolution of two functions f,g∈L1(R) (the space of absolutely integrable functions on the real line) is defined as
(f∗g)(t)=∫−∞∞f(τ)g(t−τ)dτ.
This operation measures the overlap between f and a reflected, shifted version of g, and it extends to L2(R) (square-integrable functions) via density arguments and Plancherel's theorem.[4][5]The convolution theorem states that the Fourier transform of the convolution equals the pointwise product of the individual Fourier transforms:
F{f∗g}(ω)=F{f}(ω)⋅F{g}(ω),
where the Fourier transform is defined as
F{f}(ω)=∫−∞∞f(t)e−2πiωtdt.
This holds for functions in L1(R) by direct computation using Fubini's theorem for interchanging integrals, and for L2(R) by continuity of the Fourier transform as a unitary operator on that space.[4][1]The inverse form of the theorem asserts the duality: the inverse Fourier transform of the product of the transforms yields the convolution,
f∗g=F−1{F{f}⋅F{g}}.
Here, the inverse transform is
F−1{f^}(t)=∫−∞∞f^(ω)e2πiωtdω,
ensuring the operation is reversible under suitable integrability conditions, such as when F{f} and F{g} are in L1(R). This duality simplifies computations in signal processing and partial differential equations by converting integration in the time domain to multiplication in the frequency domain.[4][1]A illustrative example is the convolution of two Gaussian functions, f(t)=e−πt2 and g(t)=e−πt2, each of whose Fourier transform is itself, F{f}(ω)=e−πω2 and similarly for g. By the theorem, F{f∗g}(ω)=e−2πω2, so the inverse transform yields f∗g(t)=21e−2πt2, another Gaussian with variance scaled by the sum of the originals. More generally, for f(t)=e−at2 and g(t)=e−bt2 with a,b>0, the result is (f∗g)(t)=a+bπe−a+babt2, demonstrating how the product of exponential transforms produces a convolved Gaussian.[2][4]The convolution theorem for the Fourier transform was first linked to the transform by Émile Borel in 1899 through work on divergent series and was fully established by Percy Daniell in 1920 via generalizations of Stieltjes-Volterra products.[6]
Discrete aperiodic case
In the discrete aperiodic case, the convolution of two sequences f and g, defined over the integers Z, is given by
(f∗g)[n]=k=−∞∑∞f[k]g[n−k]
for all n∈Z, where the sequences belong to the space ℓ1(Z) (absolutely summable) or ℓ2(Z) (square-summable).[7] This operation models the output of a linear time-invariant system in discrete-time signal processing when one sequence acts as the input and the other as the impulse response.[8]The discrete-time Fourier transform (DTFT) provides the frequency-domain representation for such aperiodic sequences, defined as
F{f}(ω)=F(ω)=n=−∞∑∞f[n]e−jωn,
where ω∈[−π,π] is the normalized angular frequency.[9] Similarly for g, yielding G(ω).The convolution theorem states that the DTFT of the convolved sequence is the pointwise product of the individual DTFTs:
F{f∗g}(ω)=F(ω)⋅G(ω).
[7][8] This property simplifies computations by transforming time-domain convolution into frequency-domain multiplication, which is particularly efficient for long sequences in digital signal processing applications.The inverse relation follows from the inverse DTFT, which reconstructs the time-domain sequence via
f∗g=F−1{F⋅G},
where the inverse transform is
(f∗g)[n]=2π1∫−ππF(ω)G(ω)ejωndω.
[9] This bidirectional equivalence holds under the specified sequence spaces, enabling seamless switching between domains.For convergence, the DTFT of sequences in ℓ1(Z) is guaranteed to exist and be continuous for all ω, as the absolute summability ensures the infinite sum converges uniformly.[10] For ℓ2(Z) sequences, the DTFT exists in the mean-square sense but may exhibit discontinuities, though the convolution theorem still applies when both inputs are in these spaces.[10]A representative example illustrates the theorem: consider the convolution of two identical rectangular sequences, each of lengthM+1, defined as x[n]=1 for 0≤n≤M and 0 otherwise. The time-domain convolution yields a triangular sequence of length2M+1, peaking at M+1 and linearly tapering to the sides. In the frequency domain, the DTFT of each rectangular sequence is the Dirichlet kernelX(ω)=sin(ω/2)sin(ω(M+1)/2)e−jωM/2, so the DTFT of the convolution is ∣X(ω)∣2, demonstrating how multiplication produces the frequency response of the triangular output.[11][7]
Periodic variants
Continuous periodic functions
For P-periodic continuous functions f and g, the periodic convolution is defined by
(f∗g)(t)=P1∫0Pf(τ)g(t−τ)dτ.
This operation produces another P-periodic function, capturing the "overlap" of f and g over one period while ensuring consistency with Fourier analysis on the circle. The normalization factor 1/P aligns the integral with the measure of the fundamental domain [0,P], preventing scaling issues that would arise in the frequency domain.[12]The Fourier series of a P-periodic function h is expressed as h(t)=∑n=−∞∞cn(h)e2πint/P, where the complex coefficients are
cn(h)=P1∫0Ph(t)e−2πint/Pdt.
The convolution theorem adapts to this setting by stating that the Fourier coefficients of the periodic convolution satisfy
cn(f∗g)=cn(f)cn(g)
for every integern. This pointwise multiplication in the frequency domain simplifies computations for periodic signals, such as filtering or system responses in engineering contexts. The theorem holds under mild conditions, such as square-integrability over the period, ensuring the integrals converge.[12][13]This periodic formulation relates to the aperiodic Fourier transform through the Dirac comb, or Shah function, \ShP(t)=∑k=−∞∞δ(t−kP). A P-periodic function can be viewed as the convolution of an aperiodic function (supported on [0,P]) with \ShP, yielding the infinite periodic extension. The Fourier transform of \ShP is (2π/P)\Sh2π/P(ω), another comb, so the transform of the periodic function becomes the aperiodic transform modulated by this comb. Applying the convolution theorem in the transform domain then recovers the coefficient multiplication after sampling at discrete frequencies 2πn/P.[14]A concrete example illustrates the theorem: the periodic convolution of two P-periodic square waves, each alternating between +1/2 and −1/2 with period P (a standard even square wave), produces a P-periodic triangular wave ranging from −1/2 to 1/2. The square wave has Fourier coefficients cn=−i/(nπ) for odd n and 0 otherwise, decaying as 1/∣n∣; their product yields cn(f∗g)∝1/n2 for odd n, matching the known series for the triangular wave and highlighting the smoothing effect of convolution.[15][16]The 1/P factor in the convolution definition is essential for the theorem's clean form, as it mirrors the normalization in the coefficient integral. Omitting it would introduce an extraneous 1/P in the product cn(f∗g)=(1/P)cn(f)cn(g), complicating reconstructions and applications like signal processing. This choice ensures the periodic case parallels the aperiodic transform limit as P→∞.[12]
Discrete circular convolution
In the discrete circular convolution, two finite-length sequences $ f[n] $ and $ g[n] $, each of length $ N $, are convolved using periodic extension with wrap-around, defined as
(f⊛g)[n]=k=0∑N−1f[k]g[(n−k)modN],0≤n<N.
This operation treats the sequences as periodic with period $ N $, incorporating aliasing effects from the modulo arithmetic that differ from linear convolution.[17]The convolution theorem for the discrete Fourier transform (DFT) states that the DFT of the circular convolution equals the pointwise product of the individual DFTs:
DFT(f⊛g)[k]=DFT(f)[k]⋅DFT(g)[k],0≤k<N,
where the DFT is given by
X[k]=n=0∑N−1x[n]e−2πikn/N.
This equivalence holds because the DFT basis functions are eigenfunctions of the circulant matrix representing circular convolution.[17][18]Conversely, the circular convolution can be recovered via the inverse DFT (IDFT):
f⊛g=IDFT{DFT(f)⋅DFT(g)},
with the IDFT defined as
x[n]=N1k=0∑N−1X[k]e2πikn/N.
This relation enables efficient computation using fast Fourier transform (FFT) algorithms, reducing complexity from $ O(N^2) $ for direct summation to $ O(N \log N) $.[17][18]A representative example involves the circular convolution of two 4-point finite impulse response sequences: $ h[n] = {-2, -1, 3, 1} $ and $ x[n] = {-1, 0, 2, 1} $. Their DFTs are $ H[k] = {1, -5 + 2i, 1, -5 - 2i} $ and $ X[k] = {2, -3 + i, 0, -3 - i} $, respectively. The product $ Y[k] = H[k] X[k] = {2, 13 - 11i, 0, 13 + 11i} $, and the IDFT yields $ y[n] = {7, 6, -6, -5} $, matching direct circular summation. This demonstrates utility in FFT-based filtering for periodic or block-processed signals.[19]To approximate linear convolution without aliasing using circular convolution, sequences are zero-padded to length at least $ N_1 + N_2 - 1 $, where $ N_1 $ and $ N_2 $ are original lengths, ensuring the wrap-around does not overlap valid terms.[17][20]
Proofs
Continuous case derivation
The derivation for the continuous aperiodic case assumes that f,g∈L1(R), the space of Lebesgue integrable functions over the real line, which guarantees the absolute convergence of the relevant integrals.[21] Under this assumption, the convolutionf∗g is well-defined and also belongs to L1(R), with ∥f∗g∥1≤∥f∥1∥g∥1.[21]Consider the Fourier transform of the convolution, defined as
[21] The integrand satisfies ∣f(τ)g(t−τ)e−2πiωt∣=∣f(τ)g(t−τ)∣, and its double integral over R2 equals ∥f∥1∥g∥1<∞, so Fubini's theorem permits interchanging the order of integration:[21]
[21] establishing the theorem.The result extends to functions in L2(R) via the Plancherel theorem, which shows that the Fourier transform is a unitary operator (isometry) on L2(R).[22] Specifically, the Schwartz functions (smooth with rapid decay) are dense in L2(R), the theorem holds on this dense subspace by the L1 case (as Schwartz functions are in L1∩L2), and continuity of the Fourier transform on L2 yields the extension.[22]An alternative derivation uses the Fourier inversion formula, which recovers g(t−τ) as the inverse transform of F{g}.[23] Substituting into the convolution integral and interchanging orders (justified by Tonelli's theorem for nonnegativity or Fubini for integrability) with a change of variables θ=t−τ leads to the inverse transform of the product F{f}⋅F{g}, confirming the theorem.[23]For functions not in L1(R) but in L2(R), the Fourier transforms are defined only in the L2 sense (not pointwise), and convolutions may require approximation by mollifiers; the theorem holds with equality in the L2 norm, addressing convergence via the Plancherel isometry.[22]
Discrete case derivation
The discrete convolution theorem for aperiodic signals states that the discrete-time Fourier transform (DTFT) of the convolution of two sequences equals the pointwise product of their individual DTFTs, under suitable convergence conditions.[7] Consider two discrete-time signals f[n] and g[n] with DTFTs F{f}(ω)=F(ω) and F{g}(ω)=G(ω), where the convolution is defined as
Assuming absolute convergence of the sequences, i.e., ∑∣f[k]∣<∞ and ∑∣g[k]∣<∞, the double summation allows reordering by Fubini's theorem for sums.[7] This yields
For periodic variants, the discrete Fourier transform (DFT) applies to finite-length sequences of length N, where convolution becomes circular. The circular convolution is
(f⊛g)[n]=k=0∑N−1f[k]g[(n−k)modN].
The DFT of this circular convolution equals the product of the individual DFTs, leveraging the orthogonality of the complex exponentials.[24] Specifically, the DFT basis functions satisfy
n=0∑N−1e2πi(k−l)n/N=NδklmodN,
which follows from the sum of a finite geometric series: for k=lmodN, the sum is 1−e2πi(k−l)/N1−e2πi(k−l)=0, and for k=lmodN, it is N. To derive the theorem, compute the DFT of f⊛g:
using the shift property and periodicity, resulting in F{f}[m]⋅F{g}[m].[24]The theorem extends to square-summable sequences (ℓ2) via density arguments and an analog of Parseval's theorem, which equates the energy in time and frequency domains: ∑n=−∞∞∣f[n]∣2=2π1∫−ππ∣F(ω)∣2dω.[25] Finite-support sequences (in ℓ1∩ℓ2) satisfy the theorem directly, and ℓ2 sequences approximate these densely, allowing the result by continuity of the DTFT on ℓ2.[25]
Generalizations
Inverse Fourier transform
The inverse Fourier transform of the pointwise product of two Fourier transforms equals the convolution of the original functions, providing the dual form of the convolution theorem. Specifically, if $ F = \mathcal{F}{f} $ and $ G = \mathcal{F}{g} $, then
F−1{F⋅G}=f∗g,
where the convolution is defined as $ (f * g)(t) = \int_{-\infty}^{\infty} f(\tau) g(t - \tau) , d\tau $, assuming $ f $ and $ g $ are sufficiently integrable functions such as those in the Schwartz class or $ L^1(\mathbb{R}) $.[1][4]This inverse form follows directly from the forward convolution theorem and the Fourier inversion theorem, which ensures that applying the forward and inverse transforms recovers the original function under appropriate conditions. By taking the inverse Fourier transform of both sides of the forward theorem $ \mathcal{F}{f * g} = F \cdot G $, and invoking the linearity and invertibility of the transform, the equivalence is established, highlighting the symmetry between the time and frequency domains.[1][26]In the discrete setting, the analogous result holds for finite sequences using the discrete Fourier transform (DFT) and its inverse (IDFT). For sequences $ x[n] $ and $ y[n] $ of length $ N $, the circular convolution is given by
IDFT{DFT{x}⋅DFT{y}}=x⊛y,
where $ (x \circledast y)[n] = \sum_{m=0}^{N-1} x[m] y[(n - m) \mod N] $, assuming zero-padding if necessary to avoid wrap-around effects in linear convolution approximations.[27][28]A practical illustration arises in signal smoothing, where multiplying the Fourier transform of a signal by a low-pass filter's transfer function (e.g., a rectangular window in frequency) attenuates high frequencies, and the inverse transform yields a convolved time-domain signal that smooths the original via the filter's impulse response, such as a sinc function.[4] This duality is particularly advantageous in filter design, as pointwise multiplication in the frequency domain is computationally simpler than direct time-domain convolution, enabling efficient implementation via fast Fourier transform algorithms.[27]
Tempered distributions
Tempered distributions are continuous linear functionals on the Schwartz spaceS(R), the space of smooth functions that decay rapidly at infinity along with all their derivatives.[29] These distributions, denoted S′(R), extend the notion of functions to include generalized objects like the Dirac delta, enabling the Fourier transform to be defined via ⟨f^,ϕ⟩=⟨f,ϕ^⟩ for ϕ∈S(R).[30]The convolution theorem generalizes to tempered distributions as follows: if f∈S′(R) is a tempered distribution and g∈Cc∞(R) is a smooth function with compact support, then the convolutionf∗g is a smooth function defined by (f∗g)(x)=⟨f,τxg~⟩, where τxg~(y)=g(x−y) and g~(y)=g(−y), and the Fourier transform satisfies F{f∗g}=F{f}⋅F{g} in the distributional sense.[29] This product of distributions is well-defined because F{g} is a Schwartz function, which multiplies smoothly with any tempered distribution.[30]Convolution between two tempered distributions is defined provided at least one has compact support, ensuring the result is again a tempered distribution; for instance, if f∈E′(R) (compactly supported distributions) and g∈S′(R), then f∗g∈S′(R).[31] A key property is that the Dirac delta distribution δ, defined by ⟨δ,ϕ⟩=ϕ(0), acts as the identity under convolution: δ∗f=f for any tempered distribution f.[32]As an illustrative example, since the Fourier transform of δ is the constant function 1 (with appropriate normalization), the theorem yields F{δ∗f}=F{δ}⋅F{f}=1⋅F{f}=F{f}, confirming consistency with δ∗f=f.[32]A proof sketch proceeds by leveraging the density of Schwartz functions in suitable topologies and the pairing definition: for a test function ϕ∈S(R), ⟨f∗g,ϕ⟩=⟨f,g~∗ϕ⟩, where g~∗ϕ is a Schwartz function; applying the Fourier transform to both sides and using the known theorem for Schwartz functions gives ⟨F{f∗g},ϕ^⟩=⟨F{f}⋅F{g},ϕ^⟩, establishing the equality in the distributional sense.[29]
Multidimensional extensions
The convolution theorem extends naturally to functions defined on Rd or Zd, generalizing the one-dimensional case where d=1. In the continuous setting, the multidimensional convolution of two integrable functions f and g is defined as
(f∗g)(x)=∫Rdf(y)g(x−y)dy,
where x,y∈Rd.[33]The multidimensional Fourier transform of a function f:Rd→C is given by
F{f}(ω)=∫Rdf(x)e−2πiω⋅xdx,
with ω∈Rd and ⋅ denoting the dot product; the inverse transform follows analogously with the positive exponent.[33] Under suitable integrability conditions (e.g., f,g∈L1(Rd)), the convolution theorem states that the Fourier transform of the convolution equals the pointwise product of the individual transforms:
This property holds componentwise in the frequency domain, facilitating efficient computation for higher-dimensional signals such as those in imaging or physics simulations.[34]In the discrete case, particularly for multidimensional arrays like digital images, the theorem applies via the multidimensional discrete Fourier transform (DFT). For two M×N arrays f(x,y) and h(x,y) in two dimensions (d=2), the 2D circular convolution is computed efficiently in the frequency domain: the 2D DFT of the convolution equals the product of the 2D DFTs, F(u,v)⋅H(u,v), followed by an inverse 2D DFT to recover the spatial result.[35] To avoid wraparound artifacts from periodicity, zero-padding is typically applied to extend the arrays to at least (2M−1)×(2N−1).[35] This is widely used in image processing for filtering operations, where the fast Fourier transform (FFT) implementation reduces complexity from O(M2N2) to O(MNlog(MN)).[35]A representative application is 2D Gaussian convolution for image blurring, where the kernel is a radially symmetric Gaussian G(x)=2πσ21e−2σ2∥x∥2. The 2D Fourier transform of this kernel is also Gaussian, F{G}(ω)=e−2π2σ2∥ω∥2, which is radial (depending only on ∥ω∥) and acts as a low-pass filter by attenuating high frequencies.[36] Convolving an image I with G thus multiplies F{I} by F{G} in the frequency domain, yielding a blurred image upon inversion; larger σ widens the Gaussian in space and narrows it in frequency, removing finer details.[36]For separable functions, where f(x)=∏k=1dfk(xk) over Cartesian coordinates, the multidimensional Fourier transform factors as F{f}(ω)=∏k=1dF{fk}(ωk), and the convolution theorem applies componentwise across dimensions.[33] This separability simplifies computations in applications like multidimensional filtering, reducing higher-dimensional operations to iterated one-dimensional transforms.[33]
Other integral transforms
The convolution theorem extends to other integral transforms, such as the Laplace transform and the Z-transform, which share the property that the transform of a convolution equals the product of the individual transforms, albeit with adaptations for their domains./09%3A_Transform_Techniques_in_Physics/9.09%3A_The_Convolution_Theorem)[37]For the Laplace transform, the theorem applies to causal functions, where the convolution is defined as the one-sided integral (f∗g)(t)=∫0tf(τ)g(t−τ)dτ for t≥0, and the transform satisfies L{f∗g}(s)=L{f}(s)⋅L{g}(s)./09%3A_Transform_Techniques_in_Physics/9.09%3A_The_Convolution_Theorem)[38] This formulation is particularly useful in analyzing linear time-invariant systems, such as where the output response is the convolution of an input with the system's impulse response, transforming to the product of their Laplace transforms (the transfer function).[39]The Z-transform, the discrete analog of the Laplace transform, exhibits a similar property for sequences: the Z-transform of the one-sided convolution (f∗g)[n]=∑k=0nf[k]g[n−k] is Z{f∗g}(z)=Z{f}(z)⋅Z{g}(z).[40][41]Unlike the bilateral Fourier transform, both the Laplace and Z-transforms are typically unilateral (one-sided), integrating or summing from zero onward, which suits them for stability analysis in causal systems and initial value problems but lacks the full inverse duality of the Fourier case.[42] These theorems were developed in the context of control theory following the Fourier transform's establishment, with the unilateral Laplace transform gaining prominence through applications in electrical engineering and system dynamics during the early 20th century.[43][44]
Applications
Signal processing
In signal processing, the convolution theorem plays a central role in the analysis and implementation of linear time-invariant (LTI) systems, where the output signal is obtained by convolving the input signal with the system's impulse response. This operation, expressed as $ y[n] = \sum_{k=-\infty}^{\infty} x[k] h[n - k] $ for discrete-time signals, corresponds in the frequency domain to multiplication of their discrete Fourier transforms: $ Y(\omega) = X(\omega) H(\omega) $. This duality enables filter design by specifying the desired frequency response $ H(\omega) $ and then deriving the impulse response $ h[n] $ via inverse transform, facilitating efficient implementation of filters like equalizers and noise reducers.[45]A key practical advantage arises from accelerating convolution computations using the fast Fourier transform (FFT), which reduces the complexity from $ O(N^2) $ for direct time-domain convolution of length-$ N $ signals to $ O(N \log N) $ by transforming to the frequency domain, multiplying the spectra, and applying the inverse FFT. This FFT-based approach, leveraging the convolution theorem, is essential for processing long signals in real-time applications, such as echo cancellation and spectral analysis.[46][47]For instance, a low-pass filter can be designed by multiplying the input's frequency spectrum by a rectangular window function that attenuates high frequencies, with the equivalent time-domain operation being convolution with a sinc function $ h[n] = \frac{\sin(\omega_c n)}{\pi n} $ (where $ \omega_c $ is the cutoff frequency), effectively smoothing the signal while preserving low-frequency components. This method is widely used in audio equalization to remove unwanted high-frequency noise.[48][49]To handle long signals and mitigate artifacts from the circular convolution inherent in finite DFTs, techniques like overlap-add and overlap-save are employed. In the overlap-add method, the input is segmented into overlapping blocks, each convolved via FFT, and the results are added after aligning overlaps, yielding linear convolution without edge effects. Similarly, overlap-save discards corrupted overlap portions post-convolution, ensuring efficient streaming processing for extended data streams.[46][50]In modern audio processing, the convolution theorem underpins real-time systems enhanced by GPU-accelerated FFT libraries, enabling low-latency convolution reverbs and spatial audio rendering since the early 2000s, with implementations like NVIDIA's cuSignal achieving up to 10x speedups over CPU methods for multichannel processing.[51][52]
Probability and statistics
In probability theory, the convolution of two probability density functions corresponds to the density of the sum of two independent continuous random variables. If X and Y are independent random variables with densities fX and fY, the density fZ of Z=X+Y is given by the convolution
fZ(z)=∫−∞∞fX(x)fY(z−x)dx.
[53] This operation arises naturally from the joint density under independence, fX,Y(x,y)=fX(x)fY(y), and marginalizing over the sum.[54]The convolution theorem simplifies analysis via characteristic functions, which are the Fourier transforms of the densities. For independent X and Y, the characteristic function of Z=X+Y is the product
ϕZ(t)=E[eitZ]=ϕX(t)ϕY(t),
where ϕX(t)=E[eitX].[55] This multiplicative property holds because the expectation factors under independence, providing a direct application of the theorem without computing the integral convolution explicitly.[56] The density fZ can then be recovered by the inverse Fourier transform of ϕZ.[54]A representative example is the Irwin–Hall distribution, which describes the sum of n independent uniform random variables on [0,1]. The density is the n-fold convolution of the uniform density fU(u)=1 for u∈[0,1]. The characteristic function is the product ϕ(t)=[iteit−1]n, equivalent to a product of sinc functions since iteit−1=eit/2⋅t/2sin(t/2).[57] /05%3A_Special_Distributions/5.25%3A_The_Irwin-Hall_Distribution) This product form highlights how the theorem captures the smoothing effect of repeated convolutions, leading to a piecewise polynomial density.[53]In statistics, the convolution theorem enables deconvolution to correct for measurement errors. Consider observations Yj=Xj+ϵj, where {ϵj} are i.i.d. errors independent of the signal {Xj} with known error densityfϵ. The densityfY=fϵ∗fX, so the characteristic functions satisfy ϕY(t)=ϕϵ(t)ϕX(t). Solving for the signal gives ϕX(t)=ϕY(t)/ϕϵ(t), with fX obtained via inverse Fourier transform.[58] This approach is widely used in density estimation under errors, though it requires regularization if ϕϵ(t) vanishes.[59]The theorem extends to multivariate settings via multivariate characteristic functions. For independent random vectors X and Y in Rd, the characteristic function of Z=X+Y is ϕZ(t)=ϕX(t)ϕY(t), where t∈Rd and ϕX(t)=E[eit⊤X].[54] This supports extensions to multinomial distributions, where the sum of independent multinomials follows a multinomial via convolution of probability mass functions, with the multivariate characteristic function product confirming the result.[60]