Hubbry Logo
logo
Laplace transform
Community hub

Laplace transform

logo
0 subscribers
Read side by side
from Wikipedia

In mathematics, the Laplace transform, named after Pierre-Simon Laplace (/ləˈplɑːs/), is an integral transform that converts a function of a real variable (usually , in the time domain) to a function of a complex variable (in the complex-valued frequency domain, also known as s-domain, or s-plane). The functions are often denoted by for the time-domain representation, and for the frequency-domain.

The transform is useful for converting differentiation and integration in the time domain into much easier multiplication and division in the Laplace domain (analogous to how logarithms are useful for simplifying multiplication and division into addition and subtraction). This gives the transform many applications in science and engineering, mostly as a tool for solving linear differential equations[1] and dynamical systems by simplifying ordinary differential equations and integral equations into algebraic polynomial equations, and by simplifying convolution into multiplication.[2][3]

For example, through the Laplace transform, the equation of the simple harmonic oscillator (Hooke's law) is converted into the algebraic equation which incorporates the initial conditions and , and can be solved for the unknown function Once solved, the inverse Laplace transform can be used to revert it back to the original domain. This is often aided by referencing tables such as that given below.

The Laplace transform is defined (for suitable functions ) by the integral where s is a complex number.

The Laplace transform is related to many other transforms. It is essentially the same as the Mellin transform, and is closely related to the Fourier transform. Unlike the Fourier transform, the Laplace transform is often an analytic function, meaning that it has a convergent power series, the coefficients of which represent the moments of the original function. Moreover, the techniques of complex analysis, and especially contour integrals, can be used for simplifying calculations.

History

[edit]
Pierre-Simon, marquis de Laplace

The Laplace transform is named after mathematician and astronomer Pierre-Simon, Marquis de Laplace, who used a similar transform in his work on probability theory.[4] Laplace wrote extensively about the use of generating functions (1814), and the integral form of the Laplace transform evolved naturally as a result.[5]

Laplace's use of generating functions was similar to what is now known as the z-transform, and he gave little attention to the continuous variable case which was discussed by Niels Henrik Abel.[6]

From 1744, Leonhard Euler investigated integrals of the form as solutions of differential equations, introducing in particular the gamma function.[7] Joseph-Louis Lagrange was an admirer of Euler and, in his work on integrating probability density functions, investigated expressions of the form which resembles a Laplace transform.[8][9]

These types of integrals seem first to have attracted Laplace's attention in 1782, where he was following in the spirit of Euler in using the integrals themselves as solutions of equations.[10] However, in 1785, Laplace took the critical step forward when, rather than simply looking for a solution in the form of an integral, he started to apply the transforms in the sense that was later to become popular. He used an integral of the form akin to a Mellin transform, to transform the whole of a difference equation, in order to look for solutions of the transformed equation. He then went on to apply the Laplace transform in the same way and started to derive some of its properties, beginning to appreciate its potential power.[11]

Laplace also recognised that Joseph Fourier's method of Fourier series for solving the diffusion equation could only apply to a limited region of space, because those solutions were periodic. In 1809, Laplace applied his transform to find solutions that diffused indefinitely in space.[12] In 1821, Cauchy developed an operational calculus for the Laplace transform that could be used to study linear differential equations in much the same way the transform is now used in basic engineering. This method was popularized, and perhaps rediscovered, by Oliver Heaviside around the turn of the century.[13]

Bernhard Riemann used the Laplace transform in his 1859 paper On the number of primes less than a given magnitude, in which he also developed the inversion theorem. Riemann used the Laplace transform to develop the functional equation of the Riemann zeta function, and his method is still used to relate the modular transformation law of the Jacobi theta function, which is simple to prove via Poisson summation, to the functional equation.[14]

Hjalmar Mellin was among the first to study the Laplace transform, rigorously in the Karl Weierstrass school of analysis, and apply it to the study of differential equations and special functions, at the turn of the 20th century.[15] At around the same time, Heaviside was busy with his operational calculus. Thomas Joannes Stieltjes considered a generalization of the Laplace transform connected to his work on moments. Other contributors in this time period included Mathias Lerch,[16] Oliver Heaviside, and Thomas Bromwich.[17]

In 1929, Vannevar Bush and Norbert Wiener published Operational Circuit Analysis as a text for engineering analysis of electrical circuits, applying both Fourier transforms and operational calculus, and in which they included one of the first predecessors of the modern table of Laplace transforms. In 1934, Raymond Paley and Norbert Wiener published the important work Fourier transforms in the complex domain, about what is now called the Laplace transform (see below). Also during the 30s, the Laplace transform was instrumental in G H Hardy and John Edensor Littlewood's study of tauberian theorems, and this application was later expounded on by Widder (1941), who developed other aspects of the theory such as a new method for inversion. Edward Charles Titchmarsh wrote the influential Introduction to the theory of the Fourier integral (1937).

The current widespread use of the transform (mainly in engineering) came about during and soon after World War II,[18] replacing the earlier Heaviside operational calculus. The advantages of the Laplace transform had been emphasized by Gustav Doetsch,[19] to whom the name Laplace transform is apparently due.

Formal definition

[edit]
for various complex frequencies in the s-domain which can be expressed as The axis contains pure cosines. Positive contains damped cosines. Negative contains exponentially growing cosines.

The Laplace transform of a function f(t), defined for all real numbers t ≥ 0, is the function F(s), which is a unilateral transform defined by[citation needed]

   (Eq. 1)

where s is a complex frequency-domain parameter with real numbers σ and ω.

An alternate notation for the Laplace transform is instead of F.[3] Thus in functional notation. This is often written, especially in engineering settings, as , with the understanding that the dummy variable does not appear in the function .

The meaning of the integral depends on types of functions of interest. A necessary condition for existence of the integral is that f must be locally integrable on [0, ∞). For locally integrable functions that decay at infinity or are of exponential type (), the integral can be understood to be a (proper) Lebesgue integral. However, for many applications it is necessary to regard it as a conditionally convergent improper integral at . Still more generally, the integral can be understood in a weak sense, and this is dealt with below.

One can define the Laplace transform of a finite Borel measure μ by the Lebesgue integral[20]

An important special case is where μ is a probability measure, for example, the Dirac delta function. In operational calculus, the Laplace transform of a measure is often treated as though the measure came from a probability density function f. In that case, to avoid potential confusion, one often writes where the lower limit of 0 is shorthand notation for

This limit emphasizes that any point mass located at 0 is entirely captured by the Laplace transform. Although with the Lebesgue integral, it is not necessary to take such a limit, it does appear more naturally in connection with the Laplace–Stieltjes transform.

Bilateral Laplace transform

[edit]

When one says "the Laplace transform" without qualification, the unilateral or one-sided transform is usually intended. The Laplace transform can be alternatively defined as the bilateral Laplace transform, or two-sided Laplace transform, by extending the limits of integration to be the entire real axis. If that is done, the common unilateral transform simply becomes a special case of the bilateral transform, where the definition of the function being transformed is multiplied by the Heaviside step function.

The bilateral Laplace transform F(s) is defined as follows:

   (Eq. 2)

An alternate notation for the bilateral Laplace transform is , instead of F.

Inverse Laplace transform

[edit]

Two integrable functions have the same Laplace transform only if they differ on a set of Lebesgue measure zero. This means that, on the range of the transform, there is an inverse transform. In fact, besides integrable functions, the Laplace transform is a one-to-one mapping from one function space into another in many other function spaces as well, although there is usually no easy characterization of the range.

Typical function spaces in which this is true include the spaces of bounded continuous functions, the space L(0, ∞), or more generally tempered distributions on (0, ∞). The Laplace transform is also defined and injective for suitable spaces of tempered distributions.

In these cases, the image of the Laplace transform lives in a space of analytic functions in the region of convergence. The inverse Laplace transform is given by the following complex integral, which is known by various names (the Bromwich integral, the Fourier–Mellin integral, and Mellin's inverse formula):

   (Eq. 3)

where γ is a real number so that the contour path of integration is in the region of convergence of F(s). In most applications, the contour can be closed, allowing the use of the residue theorem. An alternative formula for the inverse Laplace transform is given by Post's inversion formula. The limit here is interpreted in the weak-* topology.

In practice, it is typically more convenient to decompose a Laplace transform into known transforms of functions obtained from a table and construct the inverse by inspection.

Probability theory

[edit]

In pure and applied probability, the Laplace transform is defined as an expected value. If X is a random variable with probability density function f, then the Laplace transform of f is given by the expectation where is the expectation of random variable .

By convention, this is referred to as the Laplace transform of the random variable X itself. Here, replacing s by t gives the moment generating function of X. The Laplace transform has applications throughout probability theory, including first passage times of stochastic processes such as Markov chains, and renewal theory.

Of particular use is the ability to recover the cumulative distribution function of a continuous random variable X by means of the Laplace transform as follows:[21]

Algebraic construction

[edit]

The Laplace transform can be alternatively defined in a purely algebraic manner by applying a field of fractions construction to the convolution ring of functions on the positive half-line. The resulting space of abstract operators is exactly equivalent to Laplace space, but in this construction the forward and reverse transforms never need to be explicitly defined (avoiding the related difficulties with proving convergence).[22]

Region of convergence

[edit]

If f is a locally integrable function (or more generally a Borel measure locally of bounded variation), then the Laplace transform F(s) of f converges provided that the limit exists.

The Laplace transform converges absolutely if the integral exists as a proper Lebesgue integral. The Laplace transform is usually understood as conditionally convergent, meaning that it converges in the former but not in the latter sense.

The set of values for which F(s) converges absolutely is either of the form Re(s) > a or Re(s) ≥ a, where a is an extended real constant with −∞ ≤ a ≤ ∞ (a consequence of the dominated convergence theorem). The constant a is known as the abscissa of absolute convergence, and depends on the growth behavior of f(t).[23] Analogously, the two-sided transform converges absolutely in a strip of the form a < Re(s) < b, and possibly including the lines Re(s) = a or Re(s) = b.[24] The subset of values of s for which the Laplace transform converges absolutely is called the region of absolute convergence, or the domain of absolute convergence. In the two-sided case, it is sometimes called the strip of absolute convergence. The Laplace transform is analytic in the region of absolute convergence: this is a consequence of Fubini's theorem and Morera's theorem.

Similarly, the set of values for which F(s) converges (conditionally or absolutely) is known as the region of conditional convergence, or simply the region of convergence (ROC). If the Laplace transform converges (conditionally) at s = s0, then it automatically converges for all s with Re(s) > Re(s0). Therefore, the region of convergence is a half-plane of the form Re(s) > a, possibly including some points of the boundary line Re(s) = a.

In the region of convergence Re(s) > Re(s0), the Laplace transform of f can be expressed by integrating by parts as the integral

That is, F(s) can effectively be expressed, in the region of convergence, as the absolutely convergent Laplace transform of some other function. In particular, it is analytic.

There are several Paley–Wiener theorems concerning the relationship between the decay properties of f, and the properties of the Laplace transform within the region of convergence.

In engineering applications, a function corresponding to a linear time-invariant (LTI) system is stable if every bounded input produces a bounded output. This is equivalent to the absolute convergence of the Laplace transform of the impulse response function in the region Re(s) ≥ 0. As a result, LTI systems are stable, provided that the poles of the Laplace transform of the impulse response function have negative real part.

This ROC is used in knowing about the causality and stability of a system.

Properties and theorems

[edit]

The Laplace transform's key property is that it converts differentiation and integration in the time domain into multiplication and division by s in the Laplace domain. Thus, the Laplace variable s is also known as an operator variable in the Laplace domain: either the derivative operator or (for s−1) the integration operator.

Given the functions f(t) and g(t), and their respective Laplace transforms F(s) and G(s),

the following table is a list of properties of unilateral Laplace transform:[25]

Properties of the unilateral Laplace transform
Property Time domain s domain Comment
Linearity Can be proved using basic rules of integration.
Frequency-domain derivative F is the first derivative of F with respect to s.
Frequency-domain general derivative More general form, nth derivative of F(s).
Derivative f is assumed to be a differentiable function, and its derivative is assumed to be of exponential type. This can then be obtained by integration by parts
Second derivative f is assumed twice differentiable and the second derivative to be of exponential type. Follows by applying the Differentiation property to f′(t).
General derivative f is assumed to be n-times differentiable, with nth derivative of exponential type. Follows by mathematical induction.
Frequency-domain integration This is deduced using the nature of frequency differentiation and conditional convergence.
Time-domain integration u(t) is the Heaviside step function and (uf)(t) is the convolution of u(t) and f(t).
Frequency shifting
Time shifting

a > 0, u(t) is the Heaviside step function
Time scaling a > 0
Multiplication The integration is done along the vertical line Re(σ) = c that lies entirely within the region of convergence of F.[26]
Convolution
Circular convolution For periodic functions with period T.
Complex conjugation
Periodic function f(t) is a periodic function of period T so that f(t) = f(t + T), for all t ≥ 0. This is the result of the time shifting property and the geometric series.
Periodic summation

Initial value theorem
Final value theorem
, if all poles of are in the left half-plane.
The final value theorem is useful because it gives the long-term behaviour without having to perform partial fraction decompositions (or other difficult algebra). If F(s) has a pole in the right-hand plane or poles on the imaginary axis (e.g., if or ), then the behaviour of this formula is undefined.

Relation to power series

[edit]

The Laplace transform can be viewed as a continuous analogue of a power series.[27] If a(n) is a discrete function of a positive integer n, then the power series associated to a(n) is the series where x is a real variable (see Z-transform). Replacing summation over n with integration over t, a continuous version of the power series becomes where the discrete function a(n) is replaced by the continuous one f(t).

Changing the base of the power from x to e gives

For this to converge for, say, all bounded functions f, it is necessary to require that ln x < 0. Making the substitution s = ln x gives just the Laplace transform:

In other words, the Laplace transform is a continuous analog of a power series, in which the discrete parameter n is replaced by the continuous parameter t, and x is replaced by es.

Analogously to a power series, if , then the power series converges to an analytic function in , if , the Laplace transform converges to an analytic function in [28]

Relation to moments

[edit]

The quantities

are the moments of the function f. If the first n moments of f converge absolutely, then by repeated differentiation under the integral, This is of special significance in probability theory, where the moments of a random variable X are given by the expectation values . Then, the relation holds

Transform of a function's derivative

[edit]

It is often convenient to use the differentiation property of the Laplace transform to find the transform of a function's derivative. This can be derived from the basic expression for a Laplace transform as follows: yielding and in the bilateral case,

The general result where denotes the nth derivative of f, can then be established with an inductive argument.

Evaluating integrals over the positive real axis

[edit]

A useful property of the Laplace transform is the following: under suitable assumptions on the behaviour of in a right neighbourhood of and on the decay rate of in a left neighbourhood of . The above formula is a variation of integration by parts, with the operators and being replaced by and . Let us prove the equivalent formulation:

By plugging in the left-hand side turns into: but assuming Fubini's theorem holds, by reversing the order of integration we get the wanted right-hand side.

This method can be used to compute integrals that would otherwise be difficult to compute using elementary methods of real calculus. For example,

Relationship to other transforms

[edit]

Laplace–Stieltjes transform

[edit]

The (unilateral) Laplace–Stieltjes transform of a function g : ℝ → ℝ is defined by the Lebesgue–Stieltjes integral

The function g is assumed to be of bounded variation. If g is the antiderivative of f:

then the Laplace–Stieltjes transform of g and the Laplace transform of f coincide. In general, the Laplace–Stieltjes transform is the Laplace transform of the Stieltjes measure associated to g. So in practice, the only distinction between the two transforms is that the Laplace transform is thought of as operating on the density function of the measure, whereas the Laplace–Stieltjes transform is thought of as operating on its cumulative distribution function.[29]

Fourier transform

[edit]

Let be a complex-valued Lebesgue integrable function supported on , and let be its Laplace transform. Then, within the region of convergence, we have

which is the Fourier transform of the function .[30]

Indeed, the Fourier transform is a special case (under certain conditions) of the bilateral Laplace transform. The main difference is that the Fourier transform of a function is a complex function of a real variable (frequency ), the Laplace transform of a function is a complex function of a complex variable (damping factor and frequency ). The Laplace transform is usually restricted to transformation of functions of t with t ≥ 0. A consequence of this restriction is that the Laplace transform of a function is a holomorphic function of the variable s. Unlike the Fourier transform, the Laplace transform of a distribution is generally a well-behaved function. Techniques of complex variables can also be used to directly study Laplace transforms. As a holomorphic function, the Laplace transform has a power series representation. This power series expresses a function as a linear superposition of moments of the function. This perspective has applications in probability theory.

Formally, the Fourier transform is equivalent to evaluating the bilateral Laplace transform with imaginary argument s = [31][32] when the condition explained below is fulfilled,

This convention of the Fourier transform ( in Fourier transform § Other conventions) requires a factor of 1/2π on the inverse Fourier transform. This relationship between the Laplace and Fourier transforms is often used to determine the frequency spectrum of a signal or dynamical system.

The above relation is valid as stated if and only if the region of convergence (ROC) of F(s) contains the imaginary axis, σ = 0.

For example, the function f(t) = cos(ω0t) has a Laplace transform F(s) = s/(s2 + ω02) whose ROC is Re(s) > 0. As s = 0 is a pole of F(s), substituting s = in F(s) does not yield the Fourier transform of f(t)u(t), which contains terms proportional to the Dirac delta functions δ(ω ± ω0).

However, a relation of the form holds under much weaker conditions. For instance, this holds for the above example provided that the limit is understood as a weak limit of measures (see vague topology). General conditions relating the limit of the Laplace transform of a function on the boundary to the Fourier transform take the form of Paley–Wiener theorems.

Mellin transform

[edit]

The Mellin transform and its inverse are related to the two-sided Laplace transform by a simple change of variables.

If in the Mellin transform we set θ = et we get a two-sided Laplace transform.

Z-transform

[edit]

The unilateral or one-sided Z-transform is simply the Laplace transform of an ideally sampled signal with the substitution of where T = 1/fs is the sampling interval (in units of time e.g., seconds) and fs is the sampling rate (in samples per second or hertz).

Let be a sampling impulse train (also called a Dirac comb) and be the sampled representation of the continuous-time x(t)

The Laplace transform of the sampled signal xq(t) is

This is the precise definition of the unilateral Z-transform of the discrete function x[n]

with the substitution of zesT.

Comparing the last two equations, we find the relationship between the unilateral Z-transform and the Laplace transform of the sampled signal,

The similarity between the Z- and Laplace transforms is expanded upon in the theory of time scale calculus.

Borel transform

[edit]

The integral form of the Borel transform is a special case of the Laplace transform for f an entire function of exponential type, meaning that for some constants A and B. The generalized Borel transform allows a different weighting function to be used, rather than the exponential function, to transform functions not of exponential type. Nachbin's theorem gives necessary and sufficient conditions for the Borel transform to be well defined.

Fundamental relationships

[edit]

Since an ordinary Laplace transform can be written as a special case of a two-sided transform, and since the two-sided transform can be written as the sum of two one-sided transforms, the theory of the Laplace-, Fourier-, Mellin-, and Z-transforms are at bottom the same subject. However, a different point of view and different characteristic problems are associated with each of these four major integral transforms.

Table of selected Laplace transforms

[edit]

The following table provides Laplace transforms for many common functions of a single variable.[33][34] For definitions and explanations, see the Explanatory Notes at the end of the table.

Because the Laplace transform is a linear operator,

  • The Laplace transform of a sum is the sum of Laplace transforms of each term.
  • The Laplace transform of a multiple of a function is that multiple times the Laplace transformation of that function.

Using this linearity, and various trigonometric, hyperbolic, and complex number (etc.) properties and/or identities, some Laplace transforms can be obtained from others more quickly than by using the definition directly.

The unilateral Laplace transform takes as input a function whose time domain is the non-negative reals, which is why all of the time domain functions in the table below are multiples of the Heaviside step function, u(t).

The entries of the table that involve a time delay τ are required to be causal (meaning that τ > 0). A causal system is a system where the impulse response h(t) is zero for all time t prior to t = 0. In general, the region of convergence for causal systems is not the same as that of anticausal systems.

Selected Laplace transforms
Function Time domain
Laplace s-domain
Region of convergence Reference
unit impulse all s inspection
delayed impulse all s time shift of
unit impulse
unit step integrate unit impulse
delayed unit step time shift of
unit step
product of delayed function and delayed step u-substitution,
rectangular impulse
ramp integrate unit
impulse twice
nth power
(for integer n)

(n > −1)
integrate unit
step n times
qth power
(for complex q)

[35][36]
nth root Set q = 1/n above.
nth power with frequency shift Integrate unit step,
apply frequency shift
delayed nth power
with frequency shift
integrate unit step,
apply frequency shift,
apply time shift
exponential decay Frequency shift of
unit step
two-sided exponential decay
(only for bilateral transform)
Frequency shift of
unit step
exponential approach unit step minus
exponential decay
sine [37]
cosine [37]
hyperbolic sine [38]
hyperbolic cosine [38]
exponentially decaying
sine wave
[37]
exponentially decaying
cosine wave
[37]
natural logarithm [38]
Bessel function
of the first kind,
of order n

(n > −1)
[39]
Error function [39]
Explanatory notes:

s-domain equivalent circuits and impedances

[edit]

The Laplace transform is often used in circuit analysis, and simple conversions to the s-domain of circuit elements can be made. Circuit elements can be transformed into impedances, very similar to phasor impedances.

Here is a summary of equivalents:

s-domain equivalent circuits
s-domain equivalent circuits

Note that the resistor is exactly the same in the time domain and the s-domain. The sources are put in if there are initial conditions on the circuit elements. For example, if a capacitor has an initial voltage across it, or if the inductor has an initial current through it, the sources inserted in the s-domain account for that.

The equivalents for current and voltage sources are simply derived from the transformations in the table above.

Examples and applications

[edit]

The Laplace transform is used frequently in engineering and physics; the output of a linear time-invariant system can be calculated by convolving its unit impulse response with the input signal. Performing this calculation in Laplace space turns the convolution into a multiplication; the latter being easier to solve because of its algebraic form. For more information, see control theory. The Laplace transform is invertible on a large class of functions. Given a simple mathematical or functional description of an input or output to a system, the Laplace transform provides an alternative functional description that often simplifies the process of analyzing the behavior of the system, or in synthesizing a new system based on a set of specifications.[40]

The Laplace transform can also be used to solve differential equations and is used extensively in mechanical engineering and electrical engineering. The Laplace transform reduces a linear differential equation to an algebraic equation, which can then be solved by the formal rules of algebra. The original differential equation can then be solved by applying the inverse Laplace transform. English electrical engineer Oliver Heaviside first proposed a similar scheme, although without using the Laplace transform; and the resulting operational calculus is credited as the Heaviside calculus.

Evaluating improper integrals

[edit]

Let . Then (see the table above)

From which one gets:

In the limit , one gets provided that the interchange of limits can be justified. This is often possible as a consequence of the final value theorem. Even when the interchange cannot be justified the calculation can be suggestive. For example, with a ≠ 0 ≠ b, proceeding formally one has

Complex impedance of a capacitor

[edit]

In the theory of electrical circuits, the current flow in a capacitor is proportional to the capacitance and rate of change in the electrical potential (with equations as for the SI unit system). Symbolically, this is expressed by the differential equation where C is the capacitance of the capacitor, i = i(t) is the electric current through the capacitor as a function of time, and v = v(t) is the voltage across the terminals of the capacitor, also as a function of time.

Taking the Laplace transform of this equation, we obtain where and

Solving for V(s) we have

The definition of the complex impedance Z (in ohms) is the ratio of the complex voltage V divided by the complex current I while holding the initial state V0 at zero:

Using this definition and the previous equation, we find: which is the correct expression for the complex impedance of a capacitor. In addition, the Laplace transform has large applications in control theory.

Impulse response

[edit]

Consider a linear time-invariant system with transfer function

The impulse response is simply the inverse Laplace transform of this transfer function:

Partial fraction expansion

To evaluate this inverse transform, we begin by expanding H(s) using the method of partial fraction expansion,

The unknown constants P and R are the residues located at the corresponding poles of the transfer function. Each residue represents the relative contribution of that singularity to the transfer function's overall shape.

By the residue theorem, the inverse Laplace transform depends only upon the poles and their residues. To find the residue P, we multiply both sides of the equation by s + α to get

Then by letting s = −α, the contribution from R vanishes and all that is left is

Similarly, the residue R is given by

Note that and so the substitution of R and P into the expanded expression for H(s) gives

Finally, using the linearity property and the known transform for exponential decay (see Item #3 in the Table of Laplace Transforms, above), we can take the inverse Laplace transform of H(s) to obtain which is the impulse response of the system.

Convolution

The same result can be achieved using the convolution property as if the system is a series of filters with transfer functions 1/(s + α) and 1/(s + β). That is, the inverse of is

Phase delay

[edit]
Time function Laplace transform

Starting with the Laplace transform, we find the inverse by first rearranging terms in the fraction:

We are now able to take the inverse Laplace transform of our terms:

This is just the sine of the sum of the arguments, yielding:

We can apply similar logic to find that

Statistical mechanics

[edit]

In statistical mechanics, the Laplace transform of the density of states defines the partition function.[41] That is, the canonical partition function is given by and the inverse is given by

Spatial (not time) structure from astronomical spectrum

[edit]

The wide and general applicability of the Laplace transform and its inverse is illustrated by an application in astronomy which provides some information on the spatial distribution of matter of an astronomical source of radiofrequency thermal radiation too distant to resolve as more than a point, given its flux density spectrum, rather than relating the time domain with the spectrum (frequency domain).

Assuming certain properties of the object, e.g. spherical shape and constant temperature, calculations based on carrying out an inverse Laplace transformation on the spectrum of the object can produce the only possible model of the distribution of matter in it (density as a function of distance from the center) consistent with the spectrum.[42] When independent information on the structure of an object is available, the inverse Laplace transform method has been found to be in good agreement.

Birth and death processes

[edit]

Consider a random walk, with steps occurring with probabilities .[43] Suppose also that the time step is a Poisson process, with parameter . Then the probability of the walk being at the lattice point at time is

This leads to a system of integral equations (or equivalently a system of differential equations). However, because it is a system of convolution equations, the Laplace transform converts it into a system of linear equations for

namely:

which may now be solved by standard methods.

Tauberian theory

[edit]

The Laplace transform of the measure on is given by

It is intuitively clear that, for small , the exponentially decaying integrand will become more sensitive to the concentration of the measure on larger subsets of the domain. To make this more precise, introduce the distribution function:

Formally, we expect a limit of the following kind:

Tauberian theorems are theorems relating the asymptotics of the Laplace transform, as , to those of the distribution of as . They are thus of importance in asymptotic formulae of probability and statistics, where often the spectral side has asymptotics that are simpler to infer.[44]

Two Tauberian theorems of note are the Hardy–Littlewood Tauberian theorem and Wiener's Tauberian theorem. The Wiener theorem generalizes the Ikehara Tauberian theorem, which is the following statement:

Let A(x) be a non-negative, monotonic nondecreasing function of x, defined for 0 ≤ x < ∞. Suppose that

converges for ℜ(s) > 1 to the function ƒ(s) and that, for some non-negative number c,

has an extension as a continuous function for ℜ(s) ≥ 1. Then the limit as x goes to infinity of exA(x) is equal to c.

This statement can be applied in particular to the logarithmic derivative of Riemann zeta function, and thus provides an extremely short way to prove the prime number theorem.[45]

See also

[edit]

Notes

[edit]

References

[edit]

Further reading

[edit]
[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
The Laplace transform is an integral transform that maps a function of a real variable $ t $, typically representing time, to a function of a complex variable $ s $, defined mathematically as $ \mathcal{L}{f(t)}(s) = \int_0^\infty e^{-st} f(t) , dt $ for functions $ f(t) $ where the integral converges, often for $ \Re(s) > \sigma $ in the region of convergence.[1] This transformation simplifies the analysis of linear systems by converting differential equations into algebraic equations in the $ s $-domain, facilitating solutions that can be inverted back to the time domain.[2] It is particularly effective for initial value problems involving piecewise continuous or exponential-type functions, leveraging properties such as linearity and differentiation to handle derivatives directly.[1] The transform's development traces back to the 18th century, with early work on related integrals by Leonhard Euler and Joseph Louis Lagrange, but it is primarily credited to the French mathematician and astronomer Pierre-Simon Laplace (1749–1827), who formalized and extended it in 1809 within his treatise Théorie analytique des probabilités for solving indefinite integrals and probabilistic problems.[3] In the late 19th century, Oliver Heaviside independently rediscovered and popularized a version of the transform through his operational calculus (1880–1887), applying it to electrical circuit analysis without rigorous complex analysis, which spurred its practical adoption in engineering despite initial mathematical critiques.[3] The modern rigorous formulation, incorporating complex variables and convergence criteria, emerged in the early 20th century, building on contributions from mathematicians like Bromwich for the inverse transform via contour integration.[1] Key properties of the Laplace transform include linearity, which allows $ \mathcal{L}{af(t) + bg(t)} = a\mathcal{L}{f(t)} + b\mathcal{L}{g(t)} $, and the differentiation rule $ \mathcal{L}{f'(t)}(s) = sF(s) - f(0) $, enabling straightforward handling of higher-order derivatives in differential equations.[1] The inverse transform recovers $ f(t) $ from $ F(s) $, often using partial fraction decomposition or tables of standard transforms for common functions like exponentials, steps, and sinusoids.[2] In applications, it is indispensable in electrical engineering for analyzing circuits and signals, in control theory for stability assessment of feedback systems like aircraft dynamics, and in mechanical engineering for solving problems in beam deflection and vibration analysis.[2] Broader uses extend to physics for heat conduction and wave propagation, as well as signal processing, where it relates inputs to outputs in linear time-invariant systems.[3]

Formal Definition

Unilateral Laplace Transform

The unilateral Laplace transform of a function f(t)f(t) defined for t0t \geq 0 is given by the integral
L{f(t)}(s)=F(s)=0f(t)estdt, \mathcal{L}\{f(t)\}(s) = F(s) = \int_{0}^{\infty} f(t) e^{-st} \, dt,
where sCs \in \mathbb{C} is a complex variable, and the integral converges in a suitable region of the complex plane.[4][5] This one-sided transform assumes f(t)=0f(t) = 0 for t<0t < 0, focusing exclusively on the behavior of causal signals starting at t=0t = 0.[6] The notation L{f(t)}=F(s)\mathcal{L}\{f(t)\} = F(s) is commonly used to denote this transform, with F(s)F(s) representing the image of f(t)f(t) in the s-domain.[7] Unlike the bilateral version, the unilateral transform is preferred in engineering applications for analyzing systems with nonzero initial conditions, as it simplifies the transformation of derivatives to include terms like f(0)f(0) directly, facilitating solutions to initial value problems in linear differential equations.[7][5] The existence of the transform depends on a region of convergence in the s-plane where the integral is finite.[6] To illustrate, consider the constant function f(t)=1f(t) = 1 for t0t \geq 0 (the unit step function u(t)u(t)). Its unilateral Laplace transform is computed as
F(s)=0estdt=[ests]0=1s, F(s) = \int_{0}^{\infty} e^{-st} \, dt = \left[ -\frac{e^{-st}}{s} \right]_{0}^{\infty} = \frac{1}{s},
valid for Re(s)>0\operatorname{Re}(s) > 0.[8][6] For an exponential function f(t)=eatu(t)f(t) = e^{-at} u(t) with a>0a > 0, the transform yields
F(s)=0eatestdt=0e(s+a)tdt=1s+a, F(s) = \int_{0}^{\infty} e^{-at} e^{-st} \, dt = \int_{0}^{\infty} e^{-(s+a)t} \, dt = \frac{1}{s + a},
converging for Re(s)>a\operatorname{Re}(s) > -a.[8][6] These examples demonstrate how the unilateral transform converts time-domain signals into algebraic expressions in the s-domain, aiding in system analysis.

Bilateral Laplace Transform

The bilateral Laplace transform of a function f(t)f(t) defined over the entire real line is given by
F(s)=f(t)estdt, F(s) = \int_{-\infty}^{\infty} f(t) e^{-st} \, dt,
where s=σ+iωs = \sigma + i\omega is a complex variable.[9] The integral converges absolutely for values of ss within a vertical strip in the complex ss-plane, defined by α<Re(s)<β\alpha < \operatorname{Re}(s) < \beta, where the constants α\alpha and β\beta are determined by the exponential growth rates of f(t)|f(t)| as t+t \to +\infty and tt \to -\infty, respectively; outside this strip of convergence, the transform may diverge.[10][11] When σ=0\sigma = 0, so s=iωs = i\omega, the bilateral Laplace transform reduces to the Fourier transform F(iω)=f(t)eiωtdtF(i\omega) = \int_{-\infty}^{\infty} f(t) e^{-i\omega t} \, dt, assuming the imaginary axis lies within the strip of convergence.[12] In contrast to the unilateral Laplace transform, which integrates only over t0t \geq 0 and requires f(t)=0f(t) = 0 for t<0t < 0, the bilateral form accommodates signals nonzero over negative times, enabling analysis of non-causal signals; the unilateral transform is thus a special case of the bilateral when f(t)=0f(t) = 0 for t<0t < 0.[6] For instance, the bilateral transform of the non-causal signal defined by its transform X(s)=s+1(s+2)(s+3)(s1)X(s) = \frac{s+1}{(s+2)(s+3)(s-1)} yields, in the region Re(s)<3\operatorname{Re}(s) < -3, the time-domain expression x(t)=(13e2t+12e3t16et)u(t)x(t) = \left( -\frac{1}{3} e^{-2t} + \frac{1}{2} e^{-3t} - \frac{1}{6} e^{t} \right) u(-t), where the unit step u(t)u(-t) confines support to t<0t < 0.[13] For entire functions of exponential type, the bilateral Laplace transform admits an algebraic construction through term-by-term integration of the function's power series expansion, facilitating its representation as a holomorphic function in the complex plane.[14]

Region of Convergence

The region of convergence (ROC) of the Laplace transform of a time-domain function $ f(t) $ is defined as the set of complex values $ s = \sigma + j\omega $ for which the integral $ \left| \int_{-\infty}^{\infty} f(t) e^{-st} , dt \right| < \infty $, ensuring the transform exists and is finite.[15] This condition typically requires absolute integrability, where $ \int_{-\infty}^{\infty} |f(t) e^{-st}| , dt < \infty $, which guarantees that the transform is analytic within the ROC.[16] In the complex $ s $-plane, the ROC commonly appears as an open vertical strip $ {\sigma_1 < \Re(s) < \sigma_2} $, where the boundaries $ \sigma_1 $ and $ \sigma_2 $ (which may be $ \pm \infty $) are determined by the locations of the poles of the rational Laplace transform function.[15] Distinctions between absolute and conditional convergence play a key role in the implications of the ROC. Absolute convergence, as defined above, ensures uniform convergence on compact subsets of the ROC and allows for term-by-term differentiation and integration of the transform series expansion.[16] In contrast, conditional convergence occurs when the original integral converges but the absolute integral does not, which is rarer in Laplace transform applications and may lead to discontinuities or limited analytic properties outside the primary ROC.[15] The ROC is essential for determining the uniqueness of the time-domain function: if two functions have Laplace transforms that coincide on an open set within the intersection of their ROCs (with the intersection having a limit point), then the functions are identical almost everywhere.[](Oppenheim, A. V., & Willsky, A. S. (1997). Signals and Systems (2nd ed.). Prentice Hall.) Illustrative examples highlight the structure of the ROC. For a right-sided exponential function $ f(t) = e^{at} u(t) $ where $ u(t) $ is the unit step function and $ a $ is a complex constant, the ROC is the right half-plane $ \Re(s) > \Re(a) $, as the integral converges for sufficiently large positive real parts of $ s $ to dampen the growth of $ e^{at} $.[17] For a delayed exponential $ f(t) = e^{a(t - \tau)} u(t - \tau) $ with delay $ \tau > 0 $, the ROC remains the same strip $ \Re(s) > \Re(a) $, unaffected by the finite delay, though the transform itself acquires a multiplicative factor $ e^{-s\tau} $.[16] In cases of entire functions, such as those with compact support (e.g., finite-duration signals), the ROC encompasses the entire $ s $-plane, corresponding to an infinite radius of convergence for the power series expansion of the transform around infinity.[15] This connection underscores how the ROC aligns with the radius of convergence in the asymptotic power series representation $ F(s) = \sum_{n=0}^{\infty} \frac{(-1)^n m_n}{n! s^{n+1}} $ for large $ |s| $, where $ m_n $ are the moments of $ f(t) $.[](Oppenheim, A. V., & Willsky, A. S. (1997). Signals and Systems (2nd ed.). Prentice Hall.)

Inverse Laplace Transform

Bromwich Integral

The inverse Laplace transform can be expressed using the Bromwich integral, a complex contour integral that recovers the original time-domain function f(t)f(t) from its Laplace transform F(s)F(s):
f(t)=12πiγiγ+iF(s)estds, f(t) = \frac{1}{2\pi i} \int_{\gamma - i\infty}^{\gamma + i\infty} F(s) e^{st} \, ds,
where the integration path is a vertical line in the complex ss-plane with real part Re(s)=γ\operatorname{Re}(s) = \gamma, and γ\gamma lies within the region of convergence (ROC) of F(s)F(s).[18] This formulation, introduced by Thomas John I'Anson Bromwich, provides a rigorous theoretical basis for inversion through complex analysis.[19] The Bromwich contour is specifically a straight vertical line segment extending from γi\gamma - i\infty to γ+i\gamma + i\infty, positioned such that γ\gamma exceeds the real parts of all singularities (poles or branch points) of F(s)F(s), ensuring the contour lies to the right of these singularities in the ROC.[20] This placement guarantees the integral's convergence, as the exponential term este^{st} decays appropriately for t>0t > 0 when closing the contour in the left half-plane.[21] To evaluate the Bromwich integral practically, especially when F(s)F(s) is rational with isolated poles, the residue theorem from complex analysis is applied. For t>0t > 0, the contour is closed with a large semicircular arc in the left half-plane, enclosing all poles of F(s)F(s). The integral over the closed contour equals 2πi2\pi i times the sum of the residues of F(s)estF(s) e^{st} at those poles. The contribution from the semicircular arc vanishes as its radius tends to infinity, provided the conditions of Jordan's lemma are satisfied—namely, that F(s)|F(s)| decays sufficiently fast (e.g., F(s)M/sk|F(s)| \leq M / |s|^k for some M>0M > 0, k>0k > 0) in the left half-plane, ensuring the arc integral approaches zero.[21] Thus, the Bromwich integral simplifies to the sum of these residues:
f(t)=Res[F(s)est;sk], f(t) = \sum \operatorname{Res} \left[ F(s) e^{st}; s_k \right],
where the sum is over all poles sks_k of F(s)F(s) to the left of the contour.[20] This method is valid under the assumption that F(s)F(s) is analytic in the ROC except at isolated singularities, and the unilateral transform context implies f(t)=0f(t) = 0 for t<0t < 0. As a representative example, consider F(s)=1s+aF(s) = \frac{1}{s + a} with Re(a)>0\operatorname{Re}(a) > 0, so the ROC is Re(s)>Re(a)\operatorname{Re}(s) > -\operatorname{Re}(a). Choose γ>Re(a)\gamma > -\operatorname{Re}(a); the function has a simple pole at s=as = -a. The residue of F(s)estF(s) e^{st} at this pole is
Res[ests+a;s=a]=eat, \operatorname{Res} \left[ \frac{e^{st}}{s + a}; s = -a \right] = e^{-at},
yielding f(t)=eatu(t)f(t) = e^{-at} u(t), where u(t)u(t) is the unit step function.[20] This inversion demonstrates the direct computation via residues, confirming the forward transform consistency.[21]

Post's Inversion Formula

Post's inversion formula, named after Emil Post who introduced it in 1930, expresses the inverse Laplace transform as a limit involving higher-order derivatives of the transform function F(s)F(s). For a continuous function f(t)f(t) on [0,)[0, \infty) of exponential order (i.e., there exists bb such that supt>0f(t)/ebt<\sup_{t>0} |f(t)| / e^{bt} < \infty), the formula is given by
f(t)=limk(1)kk!(kt)k+1F(k)(kt),t>0, f(t) = \lim_{k \to \infty} \frac{(-1)^k}{k!} \left( \frac{k}{t} \right)^{k+1} F^{(k)} \left( \frac{k}{t} \right), \quad t > 0,
where F(k)F^{(k)} denotes the kk-th derivative of F(s)F(s) with respect to ss, and the argument k/t>bk/t > b to ensure it lies in the ROC.[22][23] The derivation relies on properties of the Laplace transform and the behavior of a sequence of functions that approximate a delta function at tt, using the fact that the kk-th derivative corresponds to $ \mathcal{L} { (-1)^k t^k f(t) } (s) = F^{(k)}(s) $. By constructing an approximating sequence and taking the limit, the formula recovers f(t)f(t).[22] This approach is useful for numerical inversion, especially with symbolic computation tools that can evaluate high-order derivatives, as it avoids identifying poles or using contour integration. In practice, the limit is approximated by computing terms for large finite kk, providing a sequence that converges to f(t)f(t).[24] For example, consider F(s)=1s2F(s) = \frac{1}{s^2}, whose inverse is f(t)=tf(t) = t for t0t \geq 0, with ROC Re(s)>0\operatorname{Re}(s) > 0. The derivatives are F(k)(s)=(1)k(k+1)!/sk+2F^{(k)}(s) = (-1)^k (k+1)! / s^{k+2}. Substituting into the formula gives
(1)kk!(kt)k+1(1)k(k+1)!(k/t)k+2=(k+1)kkk+1tk+1tk+2kk+2=k+1ktt \frac{(-1)^k}{k!} \left( \frac{k}{t} \right)^{k+1} (-1)^k \frac{(k+1)!}{(k/t)^{k+2}} = \frac{(k+1)}{k} \cdot \frac{k^{k+1}}{t^{k+1}} \cdot \frac{t^{k+2}}{k^{k+2}} = \frac{k+1}{k} \cdot t \to t
as kk \to \infty, recovering the ramp function f(t)=tf(t) = t.[22] Despite its theoretical elegance, the formula often exhibits slow convergence and numerical instability for large kk due to error amplification in higher derivatives, particularly near singularities or at the ROC boundary, requiring careful implementation for practical use.[25] The Post's inversion formula offers a derivative-based alternative to the Bromwich integral for theoretical and numerical computation of the inverse Laplace transform.

Properties and Theorems

Linearity and Shifting Theorems

The Laplace transform exhibits linearity, meaning that for any scalar constants aa and bb, and functions f(t)f(t) and g(t)g(t) whose individual Laplace transforms exist, L{af(t)+bg(t)}=aF(s)+bG(s)\mathcal{L}\{a f(t) + b g(t)\} = a F(s) + b G(s), where F(s)=L{f(t)}F(s) = \mathcal{L}\{f(t)\} and G(s)=L{g(t)}G(s) = \mathcal{L}\{g(t)\}.[26] This property follows directly from the linearity of the integral defining the transform:
L{af(t)+bg(t)}=0est[af(t)+bg(t)]dt=a0estf(t)dt+b0estg(t)dt=aF(s)+bG(s). \mathcal{L}\{a f(t) + b g(t)\} = \int_0^\infty e^{-st} [a f(t) + b g(t)] \, dt = a \int_0^\infty e^{-st} f(t) \, dt + b \int_0^\infty e^{-st} g(t) \, dt = a F(s) + b G(s).
[26] For example, if f(t)=ectf(t) = e^{ct} with L{ect}=1sc\mathcal{L}\{e^{ct}\} = \frac{1}{s - c} for Re(s)>c\operatorname{Re}(s) > c, then L{3ect}=31sc=3sc\mathcal{L}\{3 e^{ct}\} = 3 \cdot \frac{1}{s - c} = \frac{3}{s - c}.[26] The time-shifting theorem states that for a function f(t)f(t) with Laplace transform F(s)F(s), and delay τ>0\tau > 0, L{f(tτ)u(tτ)}=esτF(s)\mathcal{L}\{f(t - \tau) u(t - \tau)\} = e^{-s \tau} F(s), where u(t)u(t) is the unit step function, and the region of convergence (ROC) remains the same or expands.[27] To prove this, substitute into the integral definition and change variables σ=tτ\sigma = t - \tau:
L{f(tτ)u(tτ)}=τf(tτ)estdt=0f(σ)es(σ+τ)dσ=esτ0f(σ)esσdσ=esτF(s). \mathcal{L}\{f(t - \tau) u(t - \tau)\} = \int_\tau^\infty f(t - \tau) e^{-st} \, dt = \int_0^\infty f(\sigma) e^{-s(\sigma + \tau)} \, d\sigma = e^{-s \tau} \int_0^\infty f(\sigma) e^{-s \sigma} \, d\sigma = e^{-s \tau} F(s).
[27] An example is the unit step function u(t)u(t) with L{u(t)}=1s\mathcal{L}\{u(t)\} = \frac{1}{s} for Re(s)>0\operatorname{Re}(s) > 0; shifting gives L{u(tτ)}=esτ/s\mathcal{L}\{u(t - \tau)\} = e^{-s \tau} / s.[28] The frequency-shifting theorem, also known as the first shifting theorem, asserts that L{eatf(t)}=F(sa)\mathcal{L}\{e^{a t} f(t)\} = F(s - a), where the ROC shifts by aa (to the right if Re(a)>0\operatorname{Re}(a) > 0).[27] The proof uses the integral definition:
L{eatf(t)}=0eatf(t)estdt=0f(t)e(sa)tdt=F(sa). \mathcal{L}\{e^{a t} f(t)\} = \int_0^\infty e^{a t} f(t) e^{-s t} \, dt = \int_0^\infty f(t) e^{-(s - a) t} \, dt = F(s - a).
[27] For instance, applying this to the ramp function tu(t)t u(t) with L{tu(t)}=1s2\mathcal{L}\{t u(t)\} = \frac{1}{s^2} for Re(s)>0\operatorname{Re}(s) > 0 yields L{teatu(t)}=1(sa)2\mathcal{L}\{t e^{a t} u(t)\} = \frac{1}{(s - a)^2}.[27]

Differentiation and Integration in s-Domain

One of the key advantages of the Laplace transform in solving differential equations arises from its ability to convert time-domain differentiation into algebraic multiplication in the s-domain. For a function f(t)f(t) that is piecewise continuous and of exponential order, the Laplace transform of its first derivative is given by L{f(t)}=sF(s)f(0)\mathcal{L}\{f'(t)\} = s F(s) - f(0), where F(s)=L{f(t)}F(s) = \mathcal{L}\{f(t)\} and f(0)f(0) is the initial value at t=0t = 0.[29][27] This result is derived using integration by parts on the definition L{f(t)}=0f(t)estdt\mathcal{L}\{f'(t)\} = \int_0^\infty f'(t) e^{-st} \, dt. Setting u=estu = e^{-st} and dv=f(t)dtdv = f'(t) \, dt, so du=sestdtdu = -s e^{-st} \, dt and v=f(t)v = f(t), yields:
0f(t)estdt=[f(t)est]0+s0f(t)estdt. \int_0^\infty f'(t) e^{-st} \, dt = \left[ f(t) e^{-st} \right]_0^\infty + s \int_0^\infty f(t) e^{-st} \, dt.
The boundary term at infinity vanishes under the exponential order assumption for Re(s)>s0\operatorname{Re}(s) > s_0, leaving f(0)+sF(s)-f(0) + s F(s).[29][27] The property extends to higher-order derivatives. For the nnth derivative f(n)(t)f^{(n)}(t), assuming f(k)(t)f^{(k)}(t) for k=0,,n1k = 0, \dots, n-1 are continuous and of exponential order while f(n)(t)f^{(n)}(t) is piecewise continuous, the transform is:
L{f(n)(t)}=snF(s)k=0n1sn1kf(k)(0). \mathcal{L}\{f^{(n)}(t)\} = s^n F(s) - \sum_{k=0}^{n-1} s^{n-1-k} f^{(k)}(0).
This follows by repeated application of the first-derivative formula, incorporating initial conditions at each step.[29][27] In the s-domain, integration from 0 to tt corresponds to division by ss. Specifically, if g(t)=0tf(τ)dτg(t) = \int_0^t f(\tau) \, d\tau with f(t)f(t) piecewise continuous and of exponential order, then L{g(t)}=F(s)s\mathcal{L}\{g(t)\} = \frac{F(s)}{s}, assuming the integral starts from zero initial value.[29][27] The proof again uses integration by parts on L{g(t)}=L{f(t)}=F(s)\mathcal{L}\{g'(t)\} = \mathcal{L}\{f(t)\} = F(s), applying the differentiation property to obtain F(s)s+g(0)/s\frac{F(s)}{s} + g(0)/s, which simplifies under g(0)=0g(0) = 0.[29][27] These properties are illustrated in the context of a damped harmonic oscillator governed by x(t)+2x(t)+2x(t)=0x''(t) + 2x'(t) + 2x(t) = 0, with initial conditions x(0)=1x(0) = 1 and x(0)=1x'(0) = -1. Applying the Laplace transform and using the differentiation formulas yields:
s2X(s)s1(1)+2(sX(s)1)+2X(s)=0, s^2 X(s) - s \cdot 1 - (-1) + 2(s X(s) - 1) + 2 X(s) = 0,
which simplifies to X(s)=s+1s2+2s+2X(s) = \frac{s + 1}{s^2 + 2s + 2}. Completing the square in the denominator shows this corresponds to the transform of etcoste^{-t} \cos t, demonstrating how initial conditions via differentiation properties determine the damped oscillatory solution.[30]

Convolution Theorem

The convolution of two functions f(t)f(t) and g(t)g(t), assuming both are causal (zero for t<0t < 0), is defined for the unilateral Laplace transform as
(fg)(t)=0tf(τ)g(tτ)dτ. (f * g)(t) = \int_0^t f(\tau) g(t - \tau) \, d\tau.
/09%3A_Transform_Techniques_in_Physics/9.09%3A_The_Convolution_Theorem) The convolution theorem states that the unilateral Laplace transform of this convolution is the product of the individual transforms:
L{fg}(s)=F(s)G(s), \mathcal{L}\{f * g\}(s) = F(s) G(s),
where the region of convergence (ROC) of the product is at least the intersection of the ROCs of F(s)F(s) and G(s)G(s)./09%3A_Transform_Techniques_in_Physics/9.09%3A_The_Convolution_Theorem)[27] For the bilateral Laplace transform, the convolution is over the entire real line:
(fg)(t)=f(τ)g(tτ)dτ, (f * g)(t) = \int_{-\infty}^{\infty} f(\tau) g(t - \tau) \, d\tau,
and the theorem holds analogously, with L{fg}(s)=F(s)G(s)\mathcal{L}\{f * g\}(s) = F(s) G(s), provided the ROCs overlap sufficiently.[31] To prove the unilateral case, start with the definition:
L{fg}(s)=0est(0tf(τ)g(tτ)dτ)dt. \mathcal{L}\{f * g\}(s) = \int_0^\infty e^{-st} \left( \int_0^t f(\tau) g(t - \tau) \, d\tau \right) dt.
The region of integration is 0τt<0 \leq \tau \leq t < \infty. Applying Fubini's theorem to interchange the order of integration yields
0f(τ)(τestg(tτ)dt)dτ. \int_0^\infty f(\tau) \left( \int_\tau^\infty e^{-st} g(t - \tau) \, dt \right) d\tau.
Substitute u=tτu = t - \tau, so the inner integral becomes esτ0esug(u)du=esτG(s)e^{-s\tau} \int_0^\infty e^{-su} g(u) \, du = e^{-s\tau} G(s). Thus,
0f(τ)esτG(s)dτ=F(s)G(s). \int_0^\infty f(\tau) e^{-s\tau} G(s) \, d\tau = F(s) G(s).
The bilateral proof follows a similar interchange over R\mathbb{R}.[32]/09%3A_Transform_Techniques_in_Physics/9.09%3A_The_Convolution_Theorem) As an example, consider the convolution of f(t)=eatu(t)f(t) = e^{at} u(t) and g(t)=ebtu(t)g(t) = e^{bt} u(t) with aba \neq b, where u(t)u(t) is the unit step function. The convolution is
(fg)(t)=0teaτeb(tτ)dτ=ebte(ab)t1ab=eatebtab,t0. (f * g)(t) = \int_0^t e^{a\tau} e^{b(t - \tau)} \, d\tau = e^{bt} \frac{e^{(a-b)t} - 1}{a - b} = \frac{e^{at} - e^{bt}}{a - b}, \quad t \geq 0.
The Laplace transforms are F(s)=1/(sa)F(s) = 1/(s - a) and G(s)=1/(sb)G(s) = 1/(s - b) for Re(s)>max(a,b)\operatorname{Re}(s) > \max(a, b), and their product is 1/((sa)(sb))1/((s - a)(s - b)), whose inverse matches the convolution result, confirming the theorem. The ROC of the product is Re(s)>max(a,b)\operatorname{Re}(s) > \max(a, b), the intersection of the individual ROCs.[33][34]

Initial and Final Value Theorems

The initial value theorem provides a direct method to determine the initial value of a time-domain function f(t)f(t) from its Laplace transform F(s)F(s), assuming f(t)f(t) is causal and the limits exist. Specifically, for a function f(t)f(t) with Laplace transform F(s)F(s), the theorem states that
limt0+f(t)=limssF(s), \lim_{t \to 0^+} f(t) = \lim_{s \to \infty} s F(s),
provided that the limits on both sides exist and f(t)f(t) is piecewise continuous with at most a finite number of discontinuities in any finite interval.[27] This holds within the region of convergence (ROC) of F(s)F(s), which must extend to infinity in the right-half plane for the limit as ss \to \infty./08:_Pulse_Inputs_Dirac_Delta_Function_Impulse_Response_Initial_Value_Theorem_Convolution_Sum/8.06:_Derivation_of_the_Initial-Value_Theorem) The proof relies on the Laplace transform of the derivative: L{f(t)}(s)=sF(s)f(0+)\mathcal{L}\{f'(t)\}(s) = s F(s) - f(0^+). As ss \to \infty, if f(t)f'(t) is bounded near t=0+t=0^+ and the ROC allows it, L{f(t)}(s)0\mathcal{L}\{f'(t)\}(s) \to 0, yielding f(0+)=limssF(s)f(0^+) = \lim_{s \to \infty} s F(s).[27] An alternative proof uses the series expansion of f(t)f(t) around t=0t=0, where f(t)=n=0f(n)(0+)n!tnf(t) = \sum_{n=0}^\infty \frac{f^{(n)}(0^+)}{n!} t^n, leading to F(s)=n=0f(n)(0+)sn+1F(s) = \sum_{n=0}^\infty \frac{f^{(n)}(0^+)}{s^{n+1}}, so sF(s)=n=0f(n)(0+)sns F(s) = \sum_{n=0}^\infty \frac{f^{(n)}(0^+)}{s^{n}}, and the limit as ss \to \infty isolates the n=0n=0 term f(0+)f(0^+)./08:_Pulse_Inputs_Dirac_Delta_Function_Impulse_Response_Initial_Value_Theorem_Convolution_Sum/8.06:_Derivation_of_the_Initial-Value_Theorem) The final value theorem complements this by relating the steady-state behavior to the s-domain:
limtf(t)=lims0sF(s), \lim_{t \to \infty} f(t) = \lim_{s \to 0} s F(s),
valid if the limit exists and all poles of sF(s)s F(s) (or equivalently, of F(s)F(s)) lie in the open left-half plane Re(s)<0\operatorname{Re}(s) < 0, excluding possibly a simple pole at s=0s=0.[27] This ensures the integral defining F(s)F(s) converges at s=0s=0 and that f(t)f(t) approaches a constant without oscillation. The proof again uses the derivative property: lims0L{f(t)}(s)=lims0[sF(s)f(0+)]\lim_{s \to 0} \mathcal{L}\{f'(t)\}(s) = \lim_{s \to 0} [s F(s) - f(0^+)]. Under the pole condition, lims0L{f(t)}(s)=0\lim_{s \to 0} \mathcal{L}\{f'(t)\}(s) = 0 because f(t)0f'(t) \to 0 as tt \to \infty, so lims0sF(s)=limtf(t)\lim_{s \to 0} s F(s) = \lim_{t \to \infty} f(t)./15:_Input-Error_Operations/15.03:_Derivation_of_the_Final-Value_Theorem) If poles are on the imaginary axis or right-half plane, the theorem fails, as f(t)f(t) may diverge or oscillate. These theorems are illustrated by standard examples. For the unit step function f(t)=u(t)f(t) = u(t), where F(s)=1/sF(s) = 1/s, the initial value is limss(1/s)=1\lim_{s \to \infty} s \cdot (1/s) = 1, matching the jump at t=0+t=0^+, and the final value is lims0s(1/s)=1\lim_{s \to 0} s \cdot (1/s) = 1, confirming the steady-state level.[27] For an exponentially decaying function f(t)=eatu(t)f(t) = e^{-at} u(t) with a>0a > 0, F(s)=1/(s+a)F(s) = 1/(s+a), the initial value is limss/(s+a)=1\lim_{s \to \infty} s/(s+a) = 1, and the final value is lims0s/(s+a)=0\lim_{s \to 0} s/(s+a) = 0, reflecting the decay to zero, with poles at s=as = -a satisfying the left-half plane condition. The initial value theorem connects directly to the Maclaurin series expansion of f(t)f(t) around t=0t=0, where the coefficients f(n)(0+)/n!f^{(n)}(0^+)/n! represent scaled moments of the distribution near the origin. Higher-order extensions, such as limssn+1[F(s)f(0+)/sf(0+)/s2f(n1)(0+)/sn]=f(n)(0+)\lim_{s \to \infty} s^{n+1} [F(s) - f(0^+)/s - f'(0^+)/s^2 - \cdots - f^{(n-1)}(0^+)/s^n ] = f^{(n)}(0^+), link successive terms in the asymptotic expansion of F(s)F(s) for large ss to these Taylor coefficients, providing insight into initial transients./08:_Pulse_Inputs_Dirac_Delta_Function_Impulse_Response_Initial_Value_Theorem_Convolution_Sum/8.06:_Derivation_of_the_Initial-Value_Theorem)

Relations to Other Transforms

Fourier Transform

The Laplace transform serves as an analytic continuation of the Fourier transform, extending the analysis from the imaginary axis in the complex plane to a broader region defined by the region of convergence (ROC). Specifically, by substituting $ s = \sigma + i\omega $ into the Laplace transform and setting $ \sigma = 0 $, the expression reduces to the Fourier transform along the line $ s = i\omega $. This relationship highlights the Laplace transform's role in generalizing the Fourier transform to handle signals that may not converge under pure oscillatory exponentials, by incorporating a damping factor $ e^{-\sigma t} $ when $ \sigma > 0 $.[9][35] For the bilateral Laplace transform, defined as
F(s)=f(t)estdt, F(s) = \int_{-\infty}^{\infty} f(t) \, e^{-s t} \, dt,
substituting $ s = i\omega $ yields the standard Fourier transform
F(iω)=f(t)eiωtdt, F(i\omega) = \int_{-\infty}^{\infty} f(t) \, e^{-i \omega t} \, dt,
provided the ROC includes the imaginary axis $ \sigma = 0 $. This condition requires the signal $ f(t) $ to be of exponential order and absolutely integrable over $ (-\infty, \infty) $, ensuring the integral converges on the $ j\omega $-axis. In contrast, the unilateral (one-sided) Laplace transform,
F(s)=0f(t)estdt, F(s) = \int_{0}^{\infty} f(t) \, e^{-s t} \, dt,
assumes causality ($ f(t) = 0 $ for $ t < 0 $) and corresponds to the one-sided Fourier transform when evaluated at $ s = i\omega $, again contingent on the ROC encompassing the imaginary axis. The unilateral form is particularly useful for analyzing causal systems in engineering applications.[7][35] An illustrative example is the unilateral Laplace transform of a unit rectangular pulse $ f(t) = 1 $ for $ 0 < t < 1 $ and $ 0 $ otherwise, given by
F(s)=1ess,Re(s)>0. F(s) = \frac{1 - e^{-s}}{s}, \quad \operatorname{Re}(s) > 0.
Evaluating at $ s = i\omega $ produces the one-sided Fourier transform $ F(i\omega) = \frac{1 - e^{-i\omega}}{i\omega} $, which matches the expected sinc-like form modulated for the causal domain. The requirement $ \operatorname{Re}(s) > 0 $ introduces exponential damping $ e^{-\sigma t} $ in the integrand, facilitating convergence for pulses or similar finite-duration signals that are already Fourier-transformable but demonstrating how the Laplace framework extends to cases needing additional stability, such as growing exponentials.[36] When the ROC includes the imaginary axis, a fundamental theorem establishes that the inverse Laplace transform can be computed using Fourier inversion methods: the Bromwich integral along the line $ \operatorname{Re}(s) = 0 $ (i.e., the $ j\omega $-axis) coincides with the inverse Fourier transform, recovering $ f(t) $ via
f(t)=12πF(iω)eiωtdω. f(t) = \frac{1}{2\pi} \int_{-\infty}^{\infty} F(i\omega) \, e^{i \omega t} \, d\omega.
This equivalence underscores the Laplace transform's utility as a tool for frequency-domain analysis while providing analytic continuation for inversion in stable systems.[7][35]

Z-Transform

The Z-transform serves as the discrete-time counterpart to the continuous-time Laplace transform, facilitating the analysis of sampled signals derived from continuous systems.[37] It converts sequences of sampled values into a complex frequency-domain representation, enabling the solution of difference equations in a manner analogous to how the Laplace transform handles differential equations.[38] The Z-transform of a discrete-time signal $ f[n] = f(nT) $, where $ T $ is the sampling interval and $ u[n] $ denotes the unit step, is defined as
X(z)=n=0f(nT)zn, X(z) = \sum_{n=0}^{\infty} f(nT) z^{-n},
for $ |z| $ within the region of convergence.[38] This formulation mirrors the Laplace transform $ F(s) = \int_{0}^{\infty} f(t) e^{-st} , dt $ through the exponential mapping $ z = e^{sT} $, or inversely $ s = \frac{1}{T} \ln z $, which relates the continuous s-plane to the discrete z-plane.[37] For sampled continuous signals, the Z-transform of $ f(nT) $ provides a discrete approximation to the Laplace transform, with the approximation tightening as the sampling period $ T $ approaches zero.[37] An alternative mapping, the bilinear transform, substitutes $ s = \frac{2}{T} \frac{1 - z^{-1}}{1 + z^{-1}} $ into the Laplace-domain transfer function to obtain the Z-domain equivalent.[38] This transformation preserves the stability of pole-zero configurations by mapping the left half of the s-plane to the interior of the unit disk in the z-plane, maintaining the rational form of the transfer function while introducing a nonlinear frequency warping.[38] The impulse invariance method leverages the mapping $ z = e^{sT} $ to design digital infinite impulse response (IIR) filters from analog prototypes specified by Laplace transforms.[39] It achieves this by sampling the continuous-time impulse response $ h(t) $ at intervals $ T $ to form the discrete impulse response $ h[n] = h(nT) $, ensuring the digital filter matches the analog response at sampling points and avoiding the need for direct time-domain simulation.[39] For an analog transfer function expanded in partial fractions as $ H(s) = \sum_k \frac{K_k}{s - p_k} $, the corresponding Z-transform becomes
H(z)=kKk1epkTz1, H(z) = \sum_k \frac{K_k}{1 - e^{p_k T} z^{-1}},
where poles map directly as $ z_k = e^{p_k T} $, provided the analog signal is bandlimited to prevent aliasing.[39] A representative example is the unit step-exponential signal $ f(t) = e^{-at} u(t) $ for $ a > 0 $, with Laplace transform $ F(s) = \frac{1}{s + a} $.[37] Upon sampling, $ f(nT) = e^{-anT} u(n) $, and the Z-transform evaluates to the geometric series
X(z)=n=0(eaTz1)n=11eaTz1,z>eaT. X(z) = \sum_{n=0}^{\infty} (e^{-aT} z^{-1})^n = \frac{1}{1 - e^{-aT} z^{-1}}, \quad |z| > e^{-aT}.
This discrete form aligns with the continuous counterpart via $ z = e^{sT} $, shifting the pole from $ s = -a $ to $ z = e^{-aT} $.[39]

Mellin Transform

The Mellin transform of a function f(t)f(t) defined for 0<t<0 < t < \infty is given by
M{f}(s)=0f(t)ts1dt, \mathcal{M}\{f\}(s) = \int_0^\infty f(t) \, t^{s-1} \, dt,
where the integral converges in a vertical strip in the complex plane depending on ff.[40] This transform is closely related to the Laplace transform through a logarithmic substitution that connects multiplicative structures in the original domain to additive ones. Specifically, if g(u)=f(eu)g(u) = f(e^{-u}), then the (unilateral) Laplace transform of g(u)g(u) yields
L{g}(s)=01f(v)vs1dv, \mathcal{L}\{g\}(s) = \int_0^1 f(v) \, v^{s-1} \, dv,
which corresponds to the Mellin transform of ff restricted to (0,1)(0,1) with the parameter ss unchanged; extending to the full line via the bilateral Laplace transform g(u)esudu\int_{-\infty}^\infty g(u) e^{-s u} \, du provides the complete Mellin transform M{f}(s)\mathcal{M}\{f\}(s).[40][41] An key analogy to the Laplace transform's convolution theorem arises in the Mellin domain, where the transform converts multiplicative convolutions into simple products. The multiplicative convolution of two functions is defined as
(fg)(t)=0f(τ)g(tτ)dττ, (f \star g)(t) = \int_0^\infty f(\tau) \, g\left(\frac{t}{\tau}\right) \, \frac{d\tau}{\tau},
and its Mellin transform satisfies M{fg}(s)=M{f}(s)M{g}(s)\mathcal{M}\{f \star g\}(s) = \mathcal{M}\{f\}(s) \cdot \mathcal{M}\{g\}(s), mirroring how the Laplace transform handles additive convolutions.[40] A representative example illustrates this connection: the Mellin transform of the exponential function f(t)=etf(t) = e^{-t} is the Gamma function Γ(s)=0etts1dt\Gamma(s) = \int_0^\infty e^{-t} \, t^{s-1} \, dt for Re(s)>0\operatorname{Re}(s) > 0. This ties directly to Laplace transforms of power-law functions, as L{ta1}(s)=Γ(a)sa\mathcal{L}\{t^{a-1}\}(s) = \Gamma(a) s^{-a} for Re(a)>0\operatorname{Re}(a) > 0 and Re(s)>0\operatorname{Re}(s) > 0, highlighting how both transforms leverage the Gamma function to relate exponential decays and algebraic behaviors across domains.[40] Historically, both the Laplace and Mellin transforms have been applied to solve integral equations, with the Laplace transform addressing additive (convolutive) kernels in physical problems and the Mellin transform handling multiplicative structures in analytic number theory and special functions, as systematized by Hjalmar Mellin in the late 19th century.[40]

Common Laplace Transforms

Table of Selected Transforms

The unilateral Laplace transform is employed in these tables, considering functions $ f(t) $ that are zero for $ t < 0 $, which aligns with common applications in engineering and physics where initial conditions are specified at $ t = 0 $.[42][43]
$ f(t) $$ F(s) $ROC
$ \delta(t) $$ 1 $All s
$ u(t) $$ \frac{1}{s} $$ \operatorname{Re}(s) > 0 $
$ t^{n} $ (n positive integer)$ \frac{n!}{s^{n+1}} $$ \operatorname{Re}(s) > 0 $
$ e^{at} u(t) $$ \frac{1}{s - a} $$ \operatorname{Re}(s) > \operatorname{Re}(a) $
$ t e^{at} u(t) $$ \frac{1}{(s - a)^2} $$ \operatorname{Re}(s) > \operatorname{Re}(a) $
$ \sin(\omega t) u(t) $$ \frac{\omega}{s^2 + \omega^2} $$ \operatorname{Re}(s) > 0 $
$ \cos(\omega t) u(t) $$ \frac{s}{s^2 + \omega^2} $$ \operatorname{Re}(s) > 0 $
$ e^{at} \sin(\omega t) u(t) $$ \frac{\omega}{(s - a)^2 + \omega^2} $$ \operatorname{Re}(s) > \operatorname{Re}(a) $
$ e^{at} \cos(\omega t) u(t) $$ \frac{s - a}{(s - a)^2 + \omega^2} $$ \operatorname{Re}(s) > \operatorname{Re}(a) $
$ \sinh(\omega t) u(t) $$ \frac{\omega}{s^2 - \omega^2} $$ \operatorname{Re}(s) >
$ \cosh(\omega t) u(t) $$ \frac{s}{s^2 - \omega^2} $$ \operatorname{Re}(s) >
This table includes basic functions and damped oscillatory forms frequently used in circuit analysis and control systems; scaling factors such as n! arise from the gamma function generalization for non-integer powers, though integer cases are standard in engineering.[44][42] In common engineering notation, the unit step $ u(t) $ is often omitted but implied for causal signals.[43] Composite transforms, such as those involving convolution, appear as products in the s-domain; for example, the convolution of two exponentials $ e^{a t} u(t) * e^{b t} u(t) $ yields $ \frac{1}{(s - a)(s - b)} $ for $ a \neq b $, with ROC $ \operatorname{Re}(s) > \max(\operatorname{Re}(a), \operatorname{Re}(b)) $.[42] Properties like the first shifting theorem allow modification of table entries, such as obtaining $ e^{at} f(t) $ from $ F(s - a) $. Using linearity, transforms of polynomials times exponentials combine basic entries: for instance, $ \mathcal{L}{(a t^2 + b t + c) e^{k t} u(t)} = a \cdot \frac{2!}{(s - k)^3} + b \cdot \frac{1!}{(s - k)^2} + \frac{c}{s - k} $, assuming $ \operatorname{Re}(s) > \operatorname{Re}(k) $.[44][42]

Derivation of Key Transforms

The Laplace transform of a function f(t)f(t) is defined as L{f(t)}(s)=0f(t)estdt\mathcal{L}\{f(t)\}(s) = \int_0^\infty f(t) e^{-st} \, dt, where the integral converges for Re(s)\operatorname{Re}(s) in a suitable half-plane.[45] This section derives several fundamental transforms by direct evaluation of the integral, assuming Re(s)>0\operatorname{Re}(s) > 0 unless otherwise specified for convergence. Consider the constant function f(t)=1f(t) = 1. The transform is
L{1}(s)=0estdt. \mathcal{L}\{1\}(s) = \int_0^\infty e^{-st} \, dt.
Evaluating the antiderivative gives
[ests]0=limT(esTs+1s)=1s, \left[ -\frac{e^{-st}}{s} \right]_0^\infty = \lim_{T \to \infty} \left( -\frac{e^{-sT}}{s} + \frac{1}{s} \right) = \frac{1}{s},
since the limit term vanishes for Re(s)>0\operatorname{Re}(s) > 0.[45] For the exponential function f(t)=eatf(t) = e^{at} with constant aa, substitute into the definition:
L{eat}(s)=0eatestdt=0e(sa)tdt. \mathcal{L}\{e^{at}\}(s) = \int_0^\infty e^{at} e^{-st} \, dt = \int_0^\infty e^{-(s-a)t} \, dt.
The integral evaluates to
[e(sa)tsa]0=1sa, \left[ -\frac{e^{-(s-a)t}}{s-a} \right]_0^\infty = \frac{1}{s-a},
converging for Re(s)>a\operatorname{Re}(s) > a. A substitution u=t+a/su = t + a/s can shift the exponent, but the direct integration confirms the result equivalently.[45] The transform of powers follows from repeated application of the differentiation theorem, which states that L{tf(t)}(s)=ddsL{f(t)}(s)\mathcal{L}\{t f(t)\}(s) = -\frac{d}{ds} \mathcal{L}\{f(t)\}(s), assuming the transform exists. Starting from L{1}(s)=1/s\mathcal{L}\{1\}(s) = 1/s, differentiate to get L{t}(s)=1/s2\mathcal{L}\{t\}(s) = 1/s^2. Repeating nn times yields L{tn}(s)=n!/sn+1\mathcal{L}\{t^n\}(s) = n!/s^{n+1} for positive integer nn. Thus, for the normalized form,
L{tnn!}(s)=1sn+1. \mathcal{L}\left\{\frac{t^n}{n!}\right\}(s) = \frac{1}{s^{n+1}}.
For non-integer exponents, the result generalizes via the Gamma function, where L{tα}(s)=Γ(α+1)/sα+1\mathcal{L}\{t^\alpha\}(s) = \Gamma(\alpha+1)/s^{\alpha+1} for Re(α)>1\operatorname{Re}(\alpha) > -1 and Re(s)>0\operatorname{Re}(s) > 0, with Γ(n+1)=n!\Gamma(n+1) = n! recovering the integer case.[46] For the sine function f(t)=sin(ωt)f(t) = \sin(\omega t) with real ω>0\omega > 0, use Euler's formula sin(ωt)=eiωteiωt2i\sin(\omega t) = \frac{e^{i \omega t} - e^{-i \omega t}}{2i}. By linearity of the transform,
L{sin(ωt)}(s)=12i(L{eiωt}(s)L{eiωt}(s)). \mathcal{L}\{\sin(\omega t)\}(s) = \frac{1}{2i} \left( \mathcal{L}\{e^{i \omega t}\}(s) - \mathcal{L}\{e^{-i \omega t}\}(s) \right).
The exponential transforms are L{eiωt}(s)=1/(siω)\mathcal{L}\{e^{i \omega t}\}(s) = 1/(s - i \omega) and L{eiωt}(s)=1/(s+iω)\mathcal{L}\{e^{-i \omega t}\}(s) = 1/(s + i \omega), valid for Re(s)>0\operatorname{Re}(s) > 0. Substituting gives
12i(1siω1s+iω)=12i(s+iω)(siω)(siω)(s+iω)=12i2iωs2+ω2=ωs2+ω2. \frac{1}{2i} \left( \frac{1}{s - i \omega} - \frac{1}{s + i \omega} \right) = \frac{1}{2i} \cdot \frac{(s + i \omega) - (s - i \omega)}{(s - i \omega)(s + i \omega)} = \frac{1}{2i} \cdot \frac{2 i \omega}{s^2 + \omega^2} = \frac{\omega}{s^2 + \omega^2}.
This real integral evaluation leverages the complex exponentials along the real axis, assuming basic properties of the complex plane for analytic continuation.[47]

Applications

Solving Differential Equations

The Laplace transform provides an effective method for solving linear ordinary differential equations (ODEs) with constant coefficients, particularly initial value problems (IVPs). The approach involves applying the Laplace transform to each term of the ODE, leveraging the differentiation property—which states that the transform of the nth derivative incorporates initial conditions directly—to convert the differential equation into an algebraic equation in the s-domain.[1] The resulting equation is solved for the transform of the unknown function, Y(s), and then the inverse Laplace transform yields the time-domain solution y(t).[48] This method is especially suited for ODEs where initial conditions are specified at t=0, as they appear as explicit terms in the transformed equation without requiring separate integration.[49] A representative example is the undamped harmonic oscillator governed by the second-order ODE
my(t)+ky(t)=0, m y''(t) + k y(t) = 0,
with initial conditions y(0) = y_0 and y'(0) = v_0, where m > 0 is the mass and k > 0 is the spring constant. Applying the Laplace transform gives
m(s2Y(s)sy0v0)+kY(s)=0, m \left( s^2 Y(s) - s y_0 - v_0 \right) + k Y(s) = 0,
which simplifies to
Y(s)=msy0+mv0ms2+k=y0ss2+ω2+v0ωωs2+ω2, Y(s) = \frac{m s y_0 + m v_0}{m s^2 + k} = y_0 \frac{s}{s^2 + \omega^2} + \frac{v_0}{\omega} \frac{\omega}{s^2 + \omega^2},
where ω=k/m\omega = \sqrt{k/m}. The inverse transform produces the oscillatory solution
y(t)=y0cos(ωt)+v0ωsin(ωt), y(t) = y_0 \cos(\omega t) + \frac{v_0}{\omega} \sin(\omega t),
recovering the classical result directly from the s-domain algebra. For forced systems, such as those with step or impulse inputs, the Laplace transform facilitates solution via the convolution theorem, where the output is the convolution of the input with the system's impulse response. For instance, in the harmonic oscillator with a unit step input u(t) = 1 for t ≥ 0, the solution involves convolving the input's transform 1/s with the transfer function 1/(m s^2 + k), yielding y(t) = (1/k) (1 - cos(ω t)) after inversion. Impulse inputs, represented by the Dirac delta, correspond to the impulse response itself, simplifying analysis of sudden disturbances.[50] Compared to classical methods like undetermined coefficients or variation of parameters, the Laplace transform offers advantages in directly embedding initial conditions as y(0) and y'(0) terms, avoiding iterative guessing of particular solutions, and efficiently handling discontinuous or piecewise inputs through standard transform tables.[48] For multi-variable systems, such as coupled ODEs, the method extends naturally by defining transfer functions H(s) = Y(s)/U(s), which ratio the output transform to the input transform, enabling block-diagonal algebraic manipulation and modular analysis of interconnected linear systems.[1]

Circuit Analysis

The Laplace transform facilitates the analysis of linear time-invariant electrical circuits by converting time-domain differential equations into algebraic equations in the s-domain, where circuit elements are represented by their impedances. This approach is particularly useful for handling initial conditions and transient responses in circuits containing resistors, inductors, and capacitors. In the s-domain, the equivalents for basic passive components incorporate initial energy storage. A resistor retains its impedance $ Z_R(s) = R $, as the voltage-current relationship $ v(t) = R i(t) $ transforms directly without initial conditions. An inductor's impedance is $ Z_L(s) = sL $, with the initial current $ i_L(0) $ accounted for by a series voltage source of value $ L i_L(0) $ (noting the sign convention); similarly, a capacitor's impedance is $ Z_C(s) = \frac{1}{sC} $, with the initial voltage $ v_C(0) $ accounted for by a parallel current source of value $ C v_C(0) $ or an equivalent series voltage source of $ \frac{v_C(0)}{s} $. These equivalents allow circuits to be analyzed using familiar techniques while embedding initial conditions as sources.[51][52] Kirchhoff's laws extend seamlessly to the s-domain. Kirchhoff's voltage law (KVL) states that the algebraic sum of s-domain voltages around a loop equals zero, treating impedances as algebraic elements. Likewise, Kirchhoff's current law (KCL) requires the sum of s-domain currents at a node to be zero. These principles enable nodal and mesh analysis in the s-domain, simplifying the solution of complex networks by replacing differential equations with linear algebraic systems.[53] Transfer functions, defined as the ratio of output to input in the s-domain, characterize circuit behavior such as filtering. For a low-pass RC filter with resistor $ R $ in series and capacitor $ C $ to ground, the transfer function is $ H(s) = \frac{V_o(s)}{V_i(s)} = \frac{1}{1 + sRC} $, where the pole at $ s = -\frac{1}{RC} $ determines the cutoff frequency and roll-off. This form reveals the circuit's attenuation of high frequencies while passing low ones.[54][55] Consider a series RLC circuit driven by a unit step input $ u(t) $, with transfer function $ H(s) = \frac{1/(LC)}{s^2 + \frac{R}{L}s + \frac{1}{LC}} $ for the voltage across the capacitor (assuming zero initial conditions). The poles of $ H(s) $, roots of the denominator $ s^2 + 2\zeta\omega_n s + \omega_n^2 = 0 $ where $ \omega_n = \frac{1}{\sqrt{LC}} $ and $ \zeta = \frac{R}{2\sqrt{L/C}} ,governthetransientresponse:overdamping(, govern the transient response: overdamping ( \zeta > 1 ,realpoles)yieldsexponentialdecaywithoutoscillation;criticaldamping(, real poles) yields exponential decay without oscillation; critical damping ( \zeta = 1 )providesthefastestnonoscillatorysettling;andunderdamping() provides the fastest non-oscillatory settling; and underdamping ( \zeta < 1 $, complex poles) produces damped sinusoidal ringing. The inverse Laplace transform of $ H(s) \cdot \frac{1}{s} $ yields the time-domain step response, illustrating these damping behaviors.[8][56] For arbitrary inputs, the transient output voltage $ v_o(t) $ is the convolution of the input $ v_i(t) $ with the impulse response $ h(t) $, the inverse Laplace transform of the transfer function $ H(s) $. In the s-domain, this corresponds to $ V_o(s) = H(s) V_i(s) $, enabling efficient computation of responses to complex signals like pulses or ramps through multiplication followed by inversion.[51][8]

Probability and Stochastic Processes

In probability theory, the Laplace transform plays a central role in characterizing the distribution of non-negative random variables through its connection to the moment-generating function. For a non-negative random variable XX with probability density function f(t)f(t), the Laplace transform is defined as L{f}(s)=0estf(t)dt=E[esX]\mathcal{L}\{f\}(s) = \int_0^\infty e^{-st} f(t) \, dt = E[e^{-sX}] for (s)>0\Re(s) > 0, which equals the moment-generating function MX(s)M_X(-s).[57] This equivalence facilitates the extraction of moments, as the kk-th moment E[Xk]E[X^k] is obtained from (1)kdkdskL{f}(s)s=0(-1)^k \frac{d^k}{ds^k} \mathcal{L}\{f\}(s) \big|_{s=0}.[58] The transform's analytic properties, such as uniqueness under mild conditions, ensure that it uniquely determines the distribution, aiding in convergence theorems like those for sums of independent random variables.[59] A canonical example is the exponential distribution with rate parameter λ>0\lambda > 0, where the density is f(t)=λeλtf(t) = \lambda e^{-\lambda t} for t0t \geq 0. Its Laplace transform is L{f}(s)=λs+λ\mathcal{L}\{f\}(s) = \frac{\lambda}{s + \lambda} for (s)>λ\Re(s) > -\lambda.[60] Differentiating this transform at s=0s = 0 yields the moments: the mean E[X]=1λE[X] = \frac{1}{\lambda} from the first derivative, the variance Var(X)=1λ2\mathrm{Var}(X) = \frac{1}{\lambda^2} from the second, and higher cumulants following the pattern for the exponential family.[61] This approach extends to hypoexponential distributions, sums of independent exponentials, where the transform is the product of individual transforms, simplifying moment calculations for phase-type distributions.[62] In stochastic processes, particularly continuous-time Markov chains like birth-death processes, the Laplace transform simplifies the analysis of the Kolmogorov forward equations, which describe the time evolution of state probabilities pj(t)=P(X(t)=jX(0)=i)p_j(t) = P(X(t) = j \mid X(0) = i). These equations form an infinite system of linear differential equations; applying the Laplace transform p~j(s)=0estpj(t)dt\tilde{p}_j(s) = \int_0^\infty e^{-st} p_j(t) \, dt converts them into a system of algebraic equations solvable via generating functions or matrix methods.[63] For a general birth-death process with birth rates λj\lambda_j and death rates μj\mu_j, the transformed equations are p~j(s)(s+λj+μj)=λj1p~j1(s)+μj+1p~j+1(s)+δji\tilde{p}_j(s) (s + \lambda_j + \mu_j) = \lambda_{j-1} \tilde{p}_{j-1}(s) + \mu_{j+1} \tilde{p}_{j+1}(s) + \delta_{j i}, where δ\delta is the Kronecker delta.[64] Steady-state probabilities are recovered by taking lims0+sp~j(s)\lim_{s \to 0^+} s \tilde{p}_j(s), often leading to the balance equations λjπj=μj+1πj+1\lambda_j \pi_j = \mu_{j+1} \pi_{j+1}. This method is particularly effective for transient analysis, avoiding numerical integration of the original ODEs.[65] Renewal theory leverages the Laplace transform to analyze waiting times and counting processes, where interarrival times XiX_i are i.i.d. positive random variables with distribution FF and Laplace transform f~(s)=L{dF}(s)\tilde{f}(s) = \mathcal{L}\{dF\}(s). The renewal function m(t)=E[N(t)]m(t) = E[N(t)], the expected number of renewals by time tt, satisfies the renewal equation m(t)=F(t)+0tm(tu)dF(u)m(t) = F(t) + \int_0^t m(t - u) \, dF(u); its Laplace transform is m~(s)=f~(s)1f~(s)\tilde{m}(s) = \frac{\tilde{f}(s)}{1 - \tilde{f}(s)} for (s)>0\Re(s) > 0.[66] This yields the elementary renewal theorem limtm(t)/t=1/μ\lim_{t \to \infty} m(t)/t = 1/\mu, where μ=E[Xi]=f~(0)\mu = E[X_i] = -\tilde{f}'(0), and higher moments via further differentiation. The transform of the waiting time (forward recurrence time) distribution has density u(t)=(1F(t))/μu(t) = (1 - F(t))/\mu, with u~(s)=(1f~(s))/(sμ)\tilde{u}(s) = (1 - \tilde{f}(s))/(s \mu), enabling analysis of residual lifetimes in reliability and queueing.[67] A prominent application is the M/M/1 queue, a birth-death process with constant arrival rate λ\lambda and service rate μ>λ\mu > \lambda, where ρ=λ/μ<1\rho = \lambda / \mu < 1 is the utilization. The Kolmogorov forward equations for state probabilities transform to algebraic relations, and solving for the transform of the expected number in the system L(t)=j=0jpj(t)L(t) = \sum_{j=0}^\infty j p_j(t) yields a form whose limit as s0+s \to 0^+ gives the steady-state L=ρ/(1ρ)L = \rho / (1 - \rho), consistent with the expression ρ/(s(1ρ))\rho / (s (1 - \rho)) in the transient transform analysis.[68] This steady-state value reflects the long-run average system occupancy, derived without inverting the full transform, and extends to transient insights via partial fraction decomposition of the generating function.[69]

Improper Integral Evaluation

The Laplace transform facilitates the evaluation of improper integrals over the interval [0, ∞) by converting them into expressions amenable to algebraic manipulation, particularly through the introduction of a damping parameter and subsequent differentiation. A typical procedure involves defining a parameterized integral $ I(a) = \int_0^\infty e^{-at} f(t) , dt $ for $ a > 0 $, which coincides with the Laplace transform $ \mathcal{L}{f}(a) $, ensuring convergence. Differentiation with respect to $ a $ under the integral sign—leveraging the property $ \frac{d}{da} \mathcal{L}{f}(a) = -\mathcal{L}{t f(t)}(a) $—simplifies the form, allowing inversion or limit-taking as $ a \to 0^+ $ to recover the original integral when it converges.[70][71] This differentiation technique, often associated with Feynman's trick in integral evaluation, is particularly effective for integrals related to the Gamma function. For instance, the parameterized form $ \int_0^\infty t^{b-1} e^{-at} , dt = \frac{\Gamma(b)}{a^b} $ for $ \operatorname{Re}(a) > 0 $ and $ \operatorname{Re}(b) > 0 $ directly yields the Gamma function value upon setting $ a = 1 $, as $ \Gamma(b) = \int_0^\infty t^{b-1} e^{-t} , dt $. Repeated differentiation with respect to the parameter $ a $ generates higher moments or related forms, such as deriving $ \mathcal{L}{t^{n}}(s) = \frac{n!}{s^{n+1}} $ from the base case $ n = 0 $.[72][71] The connection between the Laplace and Mellin transforms further aids in evaluating such integrals. The Mellin transform of $ e^{-t} $ is $ \mathcal{M}{e^{-t}}(z) = \int_0^\infty t^{z-1} e^{-t} , dt = \Gamma(z) $ for $ \operatorname{Re}(z) > 0 $. By the substitution $ t = u/s $, the Laplace integral $ \int_0^\infty t^{z-1} e^{-st} , dt = s^{-z} \Gamma(z) $ emerges as a scaled Mellin transform, enabling evaluation of power-law integrals via known Gamma values.[41] A concrete example is the evaluation of $ \int_0^\infty \frac{e^{-at} \sin(bt)}{t} , dt = \arctan(b/a) $ for $ a > 0 $. This follows from the known Laplace transform $ \mathcal{L}\left{ \frac{\sin(bt)}{t} \right}(s) = \arctan(b/s) $, obtained by integrating the transform of $ \sin(bt) $ with respect to a parameter or using the integral representation; substituting $ s = a $ yields the result directly. For the related integral $ \int_0^\infty \frac{e^{-at}}{1 + t} , dt $, it can be computed using the series expansion $ \frac{1}{1+t} = \sum_{n=0}^\infty (-1)^n t^n $ for $ t < 1 $ extended via analytic continuation, or via the known transform leading to the exponential integral $ - \operatorname{Ei}(-a) $, where the Laplace property handles term-by-term integration.[70] The Laplace transform of $ t^{-1/2} $ also proves useful for integrals involving the error function. Specifically, $ \mathcal{L}{ t^{-1/2} }(s) = \sqrt{\pi / s} $ for $ \operatorname{Re}(s) > 0 $, derived from the Gamma form with $ b = 1/2 $ since $ \Gamma(1/2) = \sqrt{\pi} $. This enables evaluation of integrals like those in diffusion problems, where the complementary error function appears; for example, the transform $ \mathcal{L}{ \operatorname{erf}(\sqrt{t}) }(s) = \frac{1}{s \sqrt{s+1}} $ allows inversion to compute $ \int_0^\infty \operatorname{erf}(\sqrt{t}) e^{-st} , dt $, and limits or parameters yield values for error function-laden improper integrals such as $ \int_0^\infty \frac{\operatorname{erf}(at)}{t} , dt = \frac{\pi}{2} \ln(1 + 1/a^2) $ through differentiation under the sign.[73]

History and Development

Early Contributions

The origins of the Laplace transform lie in the mid-18th century efforts to solve differential equations using generating functions and integral representations. Leonhard Euler laid foundational work in 1744 with his paper "De constructione aequationum," where he employed generating functions to address linear differential equations of higher order. These functions involved integrals resembling the modern form, such as X(x)eaxdx\int X(x) e^{ax} \, dx, serving as precursors by transforming differential problems into algebraic ones through exponential weighting. Euler's approach marked an early shift toward operational methods in analysis, emphasizing the utility of such integrals for series solutions without fully developing an inversion theory.[74] Building on Euler's ideas, Joseph-Louis Lagrange advanced related concepts in his work on probability theory, investigating expressions of the form eaxX(x)dx\int e^{ax} X(x) \, dx for integrating probability density functions. This contributed to the conceptual framework for transform methods by highlighting the role of exponential integrals, though Lagrange prioritized algebraic manipulation over explicit integral transforms for differential equations.[3] Pierre-Simon Laplace significantly expanded these precursors in the late 1770s and early 1780s, applying them to celestial mechanics. In memoirs from 1779 and 1782, Laplace utilized expansions involving the integral estf(t)dt\int e^{-st} f(t) \, dt to model planetary perturbations, transforming complex differential equations governing orbital variations into more tractable forms. This integral form allowed him to approximate solutions for irregular motions caused by gravitational interactions among planets, demonstrating the transform's power in handling infinite series and asymptotic behaviors. By 1785, Laplace formalized these techniques in the first volume of Mécanique Céleste, where the method proved instrumental in analyzing the stability of the solar system and secular perturbations, establishing the transform as a key tool in analytical mechanics.[75] Later, in 1809, Laplace provided a more systematic treatment in his treatise Théorie analytique des probabilités, where he defined the transform explicitly and applied it to solving indefinite integrals and probabilistic problems.[3] A notable application in the 1930s was by Joseph L. Doob, who used the unilateral form of the Laplace transform in probability theory. Doob adapted the transform for stochastic processes, using the integral from 0 to ∞ to derive moment-generating functions and analyze waiting times and renewal phenomena. This unilateral variant, emphasizing causality and non-negative domains, facilitated rigorous treatments of Markov chains and diffusion processes, bridging classical analysis with modern probability.[76]

Modern Extensions

In the late 19th century, Oliver Heaviside developed operational calculus as an intuitive method for manipulating differential operators in the s-domain to solve physical problems, particularly in electromagnetism and telegraphy, though it lacked mathematical rigor and relied on formal manipulations of divergent series.[77] This approach prefigured the Laplace transform's utility in engineering but was criticized by contemporaries for its non-rigorous nature, prompting later formalizations.[77] During the 1930s, mathematicians like Emil Post and David V. Widder provided rigorous foundations for Laplace transform inversion, with Post introducing a key inversion formula in 1930 that expressed the original function as a limit involving derivatives of the transform.[78] Widder extended this work, developing the Post-Widder inversion theorem and applying Tauberian theorems to derive asymptotic behaviors of functions from their transforms, enabling precise recovery of time-domain information for analytic functions.[79] These contributions shifted the Laplace transform from heuristic tool to a cornerstone of analysis, particularly for studying moment problems and boundary behaviors.[78] The 1940s marked widespread adoption of the Laplace transform in engineering, especially feedback control systems, where Harry Nyquist's 1932 stability criterion and Hendrik Bode's 1945 frequency-domain techniques leveraged the transform's s-plane representation to analyze amplifier stability and design compensators.[80] This era solidified its role in linear systems theory, facilitating the transition from time-domain differential equations to algebraic manipulations in the complex domain for practical applications like servomechanisms.[80] Subsequent extensions include the two-sided (bilateral) Laplace transform, which integrates over the entire real line and has found use in quantum mechanics for modeling time-symmetric processes and solving difference equations in quantum variational calculus.[81] The fractional Laplace transform, incorporating non-integer orders, addresses anomalous diffusion in complex media, where standard diffusion equations fail to capture sub- or super-diffusive behaviors, as shown in asymptotic analyses of fractional-order partial differential equations.[82] More recently, numerical inversion algorithms like the Weeks method, introduced in 1966 using Laguerre function expansions for efficient computation, have seen 2020s improvements through machine learning-based parameter optimization, reducing computational complexity and enhancing accuracy for high-dimensional inversions.[83]

References

User Avatar
No comments yet.