Hubbry Logo
Linear multistep methodLinear multistep methodMain
Open search
Linear multistep method
Community hub
Linear multistep method
logo
7 pages, 0 posts
0 subscribers
Be the first to start a discussion here.
Be the first to start a discussion here.
Linear multistep method
Linear multistep method
from Wikipedia

Linear multistep methods are used for the numerical solution of ordinary differential equations. Conceptually, a numerical method starts from an initial point and then takes a short step forward in time to find the next solution point. The process continues with subsequent steps to map out the solution. Single-step methods (such as Euler's method) refer to only one previous point and its derivative to determine the current value. Methods such as Runge–Kutta take some intermediate steps (for example, a half-step) to obtain a higher order method, but then discard all previous information before taking a second step. Multistep methods attempt to gain efficiency by keeping and using the information from previous steps rather than discarding it. Consequently, multistep methods refer to several previous points and derivative values. In the case of linear multistep methods, a linear combination of the previous points and derivative values is used.

Definitions

[edit]

Numerical methods for ordinary differential equations approximate solutions to initial value problems of the form

The result is approximations for the value of at discrete times : where is the time step (sometimes referred to as ) and is an integer.

Multistep methods use information from the previous steps to calculate the next value. In particular, a linear multistep method uses a linear combination of and to calculate the value of for the desired current step. Thus, a linear multistep method is a method of the form with . The coefficients and determine the method. The designer of the method chooses the coefficients, balancing the need to get a good approximation to the true solution against the desire to get a method that is easy to apply. Often, many coefficients are zero to simplify the method.

One can distinguish between explicit and implicit methods. If , then the method is called "explicit", since the formula can directly compute . If then the method is called "implicit", since the value of depends on the value of , and the equation must be solved for . Iterative methods such as Newton's method are often used to solve the implicit formula.

Sometimes an explicit multistep method is used to "predict" the value of . That value is then used in an implicit formula to "correct" the value. The result is a predictor–corrector method.

Examples

[edit]

Consider for an example the problem The exact solution is .

One-step Euler

[edit]

A simple numerical method is Euler's method: Euler's method can be viewed as an explicit multistep method for the degenerate case of one step.

This method, applied with step size on the problem , gives the following results:

Two-step Adams–Bashforth

[edit]

Euler's method is a one-step method. A simple multistep method is the two-step Adams–Bashforth method This method needs two values, and , to compute the next value, . However, the initial value problem provides only one value, . One possibility to resolve this issue is to use the computed by Euler's method as the second value. With this choice, the Adams–Bashforth method yields (rounded to four digits): The exact solution at is , so the two-step Adams–Bashforth method is more accurate than Euler's method. This is always the case if the step size is small enough.

Families of multistep methods

[edit]

Three families of linear multistep methods are commonly used: Adams–Bashforth methods, Adams–Moulton methods, and the backward differentiation formulas (BDFs).

Adams–Bashforth methods

[edit]

The Adams–Bashforth methods are explicit methods. The coefficients are and , while the are chosen such that the methods have order s (this determines the methods uniquely).

The Adams–Bashforth methods with s = 1, 2, 3, 4, 5 are (Hairer, Nørsett & Wanner 1993, §III.1; Butcher 2003, p. 103):

The coefficients can be determined as follows. Use polynomial interpolation to find the polynomial p of degree such that The Lagrange formula for polynomial interpolation yields The polynomial p is locally a good approximation of the right-hand side of the differential equation that is to be solved, so consider the equation instead. This equation can be solved exactly; the solution is simply the integral of p. This suggests taking The Adams–Bashforth method arises when the formula for p is substituted. The coefficients turn out to be given by Replacing by its interpolant p incurs an error of order hs, and it follows that the s-step Adams–Bashforth method has indeed order s (Iserles 1996, §2.1)

The Adams–Bashforth methods were designed by John Couch Adams to solve a differential equation modelling capillary action due to Francis Bashforth. Bashforth (1883) published his theory and Adams' numerical method (Goldstine 1977).

Adams–Moulton methods

[edit]

The Adams–Moulton methods are similar to the Adams–Bashforth methods in that they also have and . Again the b coefficients are chosen to obtain the highest order possible. However, the Adams–Moulton methods are implicit methods. By removing the restriction that , an s-step Adams–Moulton method can reach order , while an s-step Adams–Bashforth methods has only order s.

The Adams–Moulton methods with s = 0, 1, 2, 3, 4 are (Hairer, Nørsett & Wanner 1993, §III.1; Quarteroni, Sacco & Saleri 2000) listed, where the first two methods are the backward Euler method and the trapezoidal rule respectively:

The derivation of the Adams–Moulton methods is similar to that of the Adams–Bashforth method; however, the interpolating polynomial uses not only the points , as above, but also . The coefficients are given by

The Adams–Moulton methods are solely due to John Couch Adams, like the Adams–Bashforth methods. The name of Forest Ray Moulton became associated with these methods because he realized that they could be used in tandem with the Adams–Bashforth methods as a predictor-corrector pair (Moulton 1926); Milne (1926) had the same idea. Adams used Newton's method to solve the implicit equation (Hairer, Nørsett & Wanner 1993, §III.1).

Backward differentiation formulas (BDF)

[edit]

The BDF methods are implicit methods with and the other coefficients chosen such that the method attains order s (the maximum possible). These methods are especially used for the solution of stiff differential equations.

Analysis

[edit]

The central concepts in the analysis of linear multistep methods, and indeed any numerical method for differential equations, are convergence, order, and stability.

Consistency and order

[edit]

The first question is whether the method is consistent: is the difference equation a good approximation of the differential equation ? More precisely, a multistep method is consistent if the local truncation error goes to zero faster than the step size h as h goes to zero, where the local truncation error is defined to be the difference between the result of the method, assuming that all the previous values are exact, and the exact solution of the equation at time . A computation using Taylor series shows that a linear multistep method is consistent if and only if All the methods mentioned above are consistent (Hairer, Nørsett & Wanner 1993, §III.2).

If the method is consistent, then the next question is how well the difference equation defining the numerical method approximates the differential equation. A multistep method is said to have order p if the local error is of order as h goes to zero. This is equivalent to the following condition on the coefficients of the methods: The s-step Adams–Bashforth method has order s, while the s-step Adams–Moulton method has order (Hairer, Nørsett & Wanner 1993, §III.2).

These conditions are often formulated using the characteristic polynomials In terms of these polynomials, the above condition for the method to have order p becomes In particular, the method is consistent if it has order at least one, which is the case if and .

Stability and convergence

[edit]

The numerical solution of a one-step method depends on the initial condition , but the numerical solution of an s-step method depend on all the s starting values, . It is thus of interest whether the numerical solution is stable with respect to perturbations in the starting values. A linear multistep method is zero-stable for a certain differential equation on a given time interval, if a perturbation in the starting values of size ε causes the numerical solution over that time interval to change by no more than Kε for some value of K which does not depend on the step size h. This is called "zero-stability" because it is enough to check the condition for the differential equation (Süli & Mayers 2003, p. 332).

If the roots of the characteristic polynomial ρ all have modulus less than or equal to 1 and the roots of modulus 1 are of multiplicity 1, we say that the root condition is satisfied. A linear multistep method is zero-stable if and only if the root condition is satisfied (Süli & Mayers 2003, p. 335).

Now suppose that a consistent linear multistep method is applied to a sufficiently smooth differential equation and that the starting values all converge to the initial value as . Then, the numerical solution converges to the exact solution as if and only if the method is zero-stable. This result is known as the Dahlquist equivalence theorem, named after Germund Dahlquist; this theorem is similar in spirit to the Lax equivalence theorem for finite difference methods. Furthermore, if the method has order p, then the global error (the difference between the numerical solution and the exact solution at a fixed time) is (Süli & Mayers 2003, p. 340).

Furthermore, if the method is convergent, the method is said to be strongly stable if is the only root of modulus 1. If it is convergent and all roots of modulus 1 are not repeated, but there is more than one such root, it is said to be relatively stable. Note that 1 must be a root for the method to be convergent; thus convergent methods are always one of these two.

To assess the performance of linear multistep methods on stiff equations, consider the linear test equation y' = λy. A multistep method applied to this differential equation with step size h yields a linear recurrence relation with characteristic polynomial This polynomial is called the stability polynomial of the multistep method. If all of its roots have modulus less than one then the numerical solution of the multistep method will converge to zero and the multistep method is said to be absolutely stable for that value of hλ. The method is said to be A-stable if it is absolutely stable for all hλ with negative real part. The region of absolute stability is the set of all hλ for which the multistep method is absolutely stable (Süli & Mayers 2003, pp. 347 & 348). For more details, see the section on stiff equations and multistep methods.

Example

[edit]

Consider the Adams–Bashforth three-step method One characteristic polynomial is thus which has roots , and the conditions above are satisfied. As is the only root of modulus 1, the method is strongly stable.

The other characteristic polynomial is

First and second Dahlquist barriers

[edit]

These two results were proved by Germund Dahlquist and represent an important bound for the order of convergence and for the A-stability of a linear multistep method. The first Dahlquist barrier was proved in Dahlquist (1956) and the second in Dahlquist (1963).

First Dahlquist barrier

[edit]

The first Dahlquist barrier states that a zero-stable and linear q-step multistep method cannot attain an order of convergence greater than q + 1 if q is odd and greater than q + 2 if q is even. If the method is also explicit, then it cannot attain an order greater than q (Hairer, Nørsett & Wanner 1993, Thm III.3.5).

Second Dahlquist barrier

[edit]

The second Dahlquist barrier states that no explicit linear multistep methods are A-stable. Further, the maximal order of an (implicit) A-stable linear multistep method is 2. Among the A-stable linear multistep methods of order 2, the trapezoidal rule has the smallest error constant (Dahlquist 1963, Thm 2.1 and 2.2).

See also

[edit]

References

[edit]
[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
A linear multistep method (LMM) is a class of techniques for solving initial value problems of ordinary differential equations (ODEs), where the approximation of the solution at a future time step is obtained as a of previous solution values and their corresponding derivatives, weighted by step size. These methods, denoted in general form as j=0kαjyn+j=hj=0kβjf(tn+j,yn+j)\sum_{j=0}^{k} \alpha_j y_{n+j} = h \sum_{j=0}^{k} \beta_j f(t_{n+j}, y_{n+j}) for a kk-step method, contrast with single-step methods like Runge-Kutta by reusing computed values to improve efficiency, particularly for non-stiff problems. The origins of LMMs trace back to the explicit Adams methods, first developed by John Couch Adams in 1855 and published in 1883 in collaboration with Francis Bashforth for applications in capillary action theory, where polynomial interpolation of the integrand enables higher-order approximations without resolving the full ODE at each step. Subsequent advancements by Germund Dahlquist in the 1950s established foundational convergence and stability theories, proving that consistent and zero-stable LMMs converge to the true solution, while introducing barriers such as the maximum order of 2k2k for a kk-step method and limitations on A-stability for explicit variants. Key examples include the explicit Adams-Bashforth methods, which predict the next solution using past derivatives (e.g., the second-order form yn+1=yn+h2(3fnfn1)y_{n+1} = y_n + \frac{h}{2}(3f_n - f_{n-1})), and implicit Adams-Moulton methods, which incorporate the future derivative for greater accuracy (e.g., third-order: yn+1=yn+h12(5fn+1+8fnfn1)y_{n+1} = y_n + \frac{h}{12}(5f_{n+1} + 8f_n - f_{n-1})). For stiff ODEs, where explicit methods suffer from severe stability restrictions, backward differentiation formulas (BDFs)—initially proposed by Charles F. Curtiss and Joseph O. Hirschfelder in 1952—offer implicit, A-stable alternatives up to order 5, such as the second-order BDF yn+143yn+13yn1=23hfn+1y_{n+1} - \frac{4}{3}y_n + \frac{1}{3}y_{n-1} = \frac{2}{3} h f_{n+1}. These methods' order and stability properties, analyzed via characteristic polynomials, remain central to their design and application in scientific computing.

Basic Concepts

Definition and Formulation

Linear multistep methods are a class of numerical integrators designed to approximate the solution of initial value problems for ordinary differential equations (ODEs) of the form y=f(t,y)y' = f(t, y), where y(t0)=y0y(t_0) = y_0 and the integration is performed over the interval [t0,T][t_0, T]. These methods advance the solution by utilizing a of previous solution values and their corresponding right-hand side evaluations, enabling efficient computation by reusing information from multiple time steps. Unlike single-step methods, they incorporate data from several prior points to achieve higher accuracy while maintaining computational efficiency. The general formulation of a linear multistep method is given by j=0kαjyn+j=hj=0kβjf(tn+j,yn+j),\sum_{j=0}^k \alpha_j y_{n+j} = h \sum_{j=0}^k \beta_j f(t_{n+j}, y_{n+j}), where h>0h > 0 is the uniform step size, tn=t0+nht_n = t_0 + n h, and the coefficients {αj}j=0k\{\alpha_j\}_{j=0}^k and {βj}j=0k\{\beta_j\}_{j=0}^k are real numbers satisfying the normalization condition αk=1\alpha_k = 1. Here, kk denotes the number of steps involved in the , and the method has an associated pp. The left-hand side advances the solution values, while the right-hand side approximates the integral contribution from the ODE's forcing term. To initiate the method beyond the y0y_0, one or more starting values y1,,yky_1, \dots, y_k are typically obtained using single-step methods such as Runge-Kutta schemes. Linear multistep methods are classified as explicit if βk=0\beta_k = 0, in which case yn+ky_{n+k} can be solved for directly without , or implicit if βk0\beta_k \neq 0, requiring the solution of a nonlinear equation at each step, often via or . The characteristic polynomials associated with the method are the advancing polynomial ρ(ζ)=j=0kαjζj\rho(\zeta) = \sum_{j=0}^k \alpha_j \zeta^j for the solution values and the forcing polynomial σ(ζ)=j=0kβjζj\sigma(\zeta) = \sum_{j=0}^k \beta_j \zeta^j for the right-hand side evaluations; these s play a central role in analyzing the method's qualitative behavior.

Historical Background

The origins of linear multistep methods can be traced to the , where Leonhard Euler's from 1768 served as an early single-step precursor for of ordinary differential equations (ODEs). However, the first explicit multistep approaches were developed by British astronomer around 1855 and published in 1883, who created formulas to compute solutions for ODEs modeling in collaboration with Francis Bashforth, motivated by astronomical and physical computations requiring higher efficiency than single-step methods. These early methods laid the groundwork for using multiple previous points to approximate derivatives, improving accuracy for non-stiff problems without explicit reliance on advanced computing. The formalization of linear multistep methods as a general class occurred in the mid-20th century through the pioneering work of Swedish mathematician Germund Dahlquist. In his 1956 paper, Dahlquist established the foundational theory of convergence and stability for these methods, analyzing zero-stability and consistency conditions essential for reliable numerical solutions of ODEs. This was expanded in his 1958 doctoral dissertation at , titled "Stability and error bounds in the numerical integration of ordinary differential equations," which introduced the broad linear multistep framework and emphasized stability concepts for practical implementation. Dahlquist's contributions were driven by the need to extend single-step methods like Runge-Kutta for more efficient integration in scientific computing, particularly as electronic computers emerged. During the 1950s and , advancements built on Adams' explicit methods by incorporating predictor-corrector pairs, where an explicit predictor (like Adams-Bashforth) estimates the next value and an implicit corrector (like Adams-Moulton) refines it, enhancing accuracy and stability for non-stiff ODEs in applications. Concurrently, backward differentiation formulas (BDFs), first proposed by Charles F. Curtiss and Joseph O. Hirschfelder in 1952, were further developed and popularized in the by C. William Gear to address stiff ODEs, where traditional explicit methods fail due to stability restrictions; Gear's 1966 publications and 1971 monograph popularized variable-order BDF implementations for simulations in and circuit analysis. These developments were motivated by the computational demands of post-World War II , including adaptations for stiff systems arising in reactor modeling and . Key milestones include Dahlquist's 1963 paper, which analyzed special stability issues and introduced A-stability, alongside the first Dahlquist barrier limiting order for stable methods, discovered during this era as theoretical bounds on achievable accuracy. Practical testing of these methods gained traction in the era of the late 1940s and 1950s, where early computers facilitated large-scale solutions for ballistics and weather prediction, underscoring the shift from hand calculations to automated .

Introductory Examples

Forward Euler Method

The forward Euler method represents the simplest linear multistep method for solving initial value problems of the form y=f(t,y)y' = f(t, y), y(t0)=y0y(t_0) = y_0, functioning as a one-step explicit scheme with k=1k=1. In the general linear multistep framework, it takes the form yn+1yn=hf(tn,yn)y_{n+1} - y_n = h f(t_n, y_n), corresponding to coefficients α0=1\alpha_0 = -1, α1=1\alpha_1 = 1, β0=1\beta_0 = 1, and β1=0\beta_1 = 0. This method derives from the first-order Taylor expansion of the exact solution around tnt_n: y(tn+1)=y(tn)+hy(tn)+12h2y(ξn)y(t_{n+1}) = y(t_n) + h y'(t_n) + \frac{1}{2} h^2 y''(\xi_n) for some ξn(tn,tn+1)\xi_n \in (t_n, t_{n+1}), where y(tn)=f(tn,y(tn))y'(t_n) = f(t_n, y(t_n)). Truncating the higher-order term yields the approximation y(tn+1)y(tn)+hf(tn,y(tn))y(t_{n+1}) \approx y(t_n) + h f(t_n, y(t_n)), which discretizes the ODE using a forward difference. The local , assuming the solution is exact at tnt_n, arises from neglecting the O(h2)O(h^2) in the Taylor expansion, resulting in an of O(h2)O(h^2) per step and confirming the method's accuracy. Over multiple steps spanning a fixed interval, this leads to a global of O(h)O(h) due to accumulation. Implementation proceeds iteratively from the , advancing the solution at each discrete time point tn=t0+nht_n = t_0 + n h. The following illustrates the basic algorithm for computing the solution up to a final time tft_f, with N=(tft0)/hN = \lceil (t_f - t_0)/h \rceil steps:

function forward_euler(y0, t0, tf, h, f) N = ceil((tf - t0) / h) t = t0 y = y0 for n = 0 to N-1 y = y + h * f(t, y) t = t + h end return y, t // approximate y(tf) end

function forward_euler(y0, t0, tf, h, f) N = ceil((tf - t0) / h) t = t0 y = y0 for n = 0 to N-1 y = y + h * f(t, y) t = t + h end return y, t // approximate y(tf) end

This requires only the function ff and initial data, with no solver for nonlinear equations. The method's primary advantages lie in its simplicity and ease of implementation, requiring no prior solution values beyond the , which eliminates the need for startup procedures common in multistep methods. However, its low order results in a global error of O(h)O(h), causing significant accumulation for moderate step sizes and limiting its use to problems tolerant of reduced accuracy.

Two-Step Adams-Bashforth Method

The two-step Adams-Bashforth method is an explicit linear multistep technique designed for numerically solving non-stiff initial value problems of the form y=f(t,y)y' = f(t, y), y(t0)=y0y(t_0) = y_0. It advances the solution from two prior points by extrapolating the right-hand side function linearly to approximate the over the next step interval. This method belongs to the broader Adams family of explicit methods and is valued for its simplicity and efficiency when the underlying lacks , allowing straightforward evaluation of ff without solving nonlinear equations. The method's formula is
yn+1=yn+h(32f(tn,yn)12f(tn1,yn1)),y_{n+1} = y_n + h \left( \frac{3}{2} f(t_n, y_n) - \frac{1}{2} f(t_{n-1}, y_{n-1}) \right),
where hh is the step size. This arises from deriving an to y(tn+1)=y(tn)+tntn+1f(s,y(s))dsy(t_{n+1}) = y(t_n) + \int_{t_n}^{t_{n+1}} f(s, y(s)) \, ds by fitting a linear to ff at the points tn1t_{n-1} and tnt_n. Let u=(stn)/hu = (s - t_n)/h, so the interpolation points are at u=1u = -1 and u=0u = 0. The is p(u)=fn+(fnfn1)up(u) = f_n + (f_n - f_{n-1}) u, and integrating yields 01p(u)du=32fn12fn1\int_0^1 p(u) \, du = \frac{3}{2} f_n - \frac{1}{2} f_{n-1}, which is then multiplied by hh to obtain the update.
The two-step Adams-Bashforth method achieves second-order accuracy, meaning the global error is O(h2)O(h^2) under suitable conditions, with a local of 512h3y(ξ)\frac{5}{12} h^3 y'''(\xi) for some ξ(tn1,tn+1)\xi \in (t_{n-1}, t_{n+1}). To start the iteration, an initial approximation y1y_1 is needed beyond the given y0y_0; this is typically computed using a lower-order single-step method like the forward , after which subsequent steps follow the formula iteratively. The explicit nature makes it computationally inexpensive for non-stiff problems, where the stability region encompasses a portion of the left half-plane sufficient for many practical applications.

Major Families

Adams Methods

The Adams methods form a key family of linear multistep methods designed for the numerical integration of non-stiff ordinary differential equations (ODEs), comprising the explicit Adams-Bashforth (AB) methods and the implicit Adams-Moulton (AM) methods. These methods approximate the solution by integrating an interpolating polynomial fitted to the right-hand side function ff of the ODE y=f(t,y)y' = f(t, y). The AB methods interpolate using values at previous time points, while the AM methods incorporate the value at the new time point, making them implicit. Developed initially for astronomical and physical computations, the methods provide high-order accuracy with efficient use of function evaluations. The k-step Adams-Bashforth method is explicit and given by yn+1=yn+hj=0k1βjfnj,y_{n+1} = y_n + h \sum_{j=0}^{k-1} \beta_j f_{n-j}, where the coefficients βj\beta_j are chosen via of ff over the interval [tnk+1,tn][t_{n-k+1}, t_n] to exactly integrate polynomials up to degree kk, yielding local O(hk+1)O(h^{k+1}) and global order kk. The k-step Adams-Moulton method is implicit and formulated as yn+1=yn+h(βkfn+1+j=0k1βjfnj),y_{n+1} = y_n + h \left( \beta_k f_{n+1} + \sum_{j=0}^{k-1} \beta_j f_{n-j} \right), derived similarly but using backward over [tn+1k,tn+1][t_{n+1-k}, t_{n+1}] to achieve order k+1k+1. The implicit nature of AM requires solving a nonlinear equation at each step, typically via or Newton methods. A common strategy pairs AB and AM in a predictor-corrector framework, where the explicit AB method provides an initial guess y~n+1\tilde{y}_{n+1} (predictor) to evaluate fn+1f_{n+1}, which is then inserted into the AM formula for refinement (corrector). This can be applied once (PE) for simplicity or iteratively (e.g., PEC or PECE) for higher accuracy, with the AM step often using one higher order than the AB predictor to optimize while maintaining overall order. This , which leverages the strengths of both explicit and implicit accuracy, was formalized by Forest Ray Moulton for ballistic trajectory computations. The coefficients for low-order Adams methods (orders 1 to 4) are summarized below, where for AB the coefficients apply to fn,fn1,f_n, f_{n-1}, \dots; for AM, they apply to fn+1,fn,fn1,f_{n+1}, f_n, f_{n-1}, \dots. These values ensure the methods are exact for polynomials up to the specified order minus one.
OrderAB CoefficientsAM Coefficients
11111
232,12\frac{3}{2}, -\frac{1}{2}12,12\frac{1}{2}, \frac{1}{2}
32312,1612,512\frac{23}{12}, -\frac{16}{12}, \frac{5}{12}512,812,112\frac{5}{12}, \frac{8}{12}, -\frac{1}{12}
45524,5924,3724,924\frac{55}{24}, -\frac{59}{24}, \frac{37}{24}, -\frac{9}{24}924,1924,524,124\frac{9}{24}, \frac{19}{24}, -\frac{5}{24}, \frac{1}{24}
Adams methods are particularly effective for non-stiff ODEs, where the explicit AB variants avoid nonlinear solves and the implicit AM variants deliver superior accuracy per function evaluation in predictor-corrector mode—often one order higher than AB for comparable computational cost. Extensions to variable step sizes incorporate local error estimates from the predictor-corrector difference to adjust hh, enabling adaptive integration in solvers like MATLAB's ode113.

Backward Differentiation Formulas

The backward differentiation formulas (BDFs) constitute a family of implicit linear multistep methods designed for the numerical solution of ordinary differential equations (ODEs), particularly those exhibiting . These methods approximate the solution by fitting a to previous solution values and evaluating the at the current time step using backward differences, which enhances stability for systems with widely varying timescales. Unlike explicit methods, BDFs require solving an implicit at each step, making them suitable for stiff problems where explicit schemes would necessitate prohibitively small time steps. The general formulation of a kk-step BDF is given by j=0kαjyn+j=hβkfn+k,\sum_{j=0}^k \alpha_j y_{n+j} = h \beta_k f_{n+k}, where hh is the step size, αk=1\alpha_k = 1, βk>0\beta_k > 0, and the coefficients αj\alpha_j (for j=0,,k1j=0,\dots,k-1) and βk\beta_k are determined by requiring the method to have order kk, meaning the local is O(hk+1)O(h^{k+1}). These coefficients arise from the approximation to the , specifically using backward differences to express yn+ky'_{n+k} in terms of past and current solution values. The derivation of BDFs relies on the backward difference operator yn=ynyn1\nabla y_n = y_n - y_{n-1}, with higher-order differences defined recursively as kyn=(k1yn)\nabla^k y_n = \nabla (\nabla^{k-1} y_n). The method approximates yn+ky'_{n+k} by interpolating a polynomial through the points yn,yn1,,ynk+1y_{n}, y_{n-1}, \dots, y_{n-k+1} and differentiating at tn+kt_{n+k}, or equivalently using the generating function approach to satisfy the order conditions. This backward-focused differencing provides inherent damping of high-frequency oscillations, a key property for stiff systems. BDF methods are zero-stable for orders k6k \leq 6, A-stable for k=1,2k=1,2, and A(α)(\alpha)-stable for k=3k=3 to $6(with(with\alphadecreasingfromapproximatelydecreasing from approximately90^\circtoto18^\circ),allowingreliableintegrationalongthenegativerealaxisinthe[complexplane](/page/Complexplane),butbecomeunstablefor), allowing reliable integration along the negative real axis in the [complex plane](/page/Complex_plane), but become unstable for k \geq 7$ due to root loci entering the right half-plane. A commonly used example is the second-order BDF (BDF2), formulated as yn+143yn+13yn1=23hfn+1,y_{n+1} - \frac{4}{3} y_n + \frac{1}{3} y_{n-1} = \frac{2}{3} h f_{n+1}, which balances computational efficiency and accuracy for many stiff problems. Due to their implicit nature, BDF equations are solved iteratively using nonlinear solvers such as , often with approximations to handle the resulting systems efficiently. For stiff ODEs, BDFs offer significant advantages over explicit methods like Adams-Bashforth by providing better damping of high-frequency modes, enabling larger step sizes without numerical instability. This makes them ideal for applications in , circuit simulation, and other stiff scenarios. Modern implementations, such as those in the SUNDIALS suite, employ variable-order BDFs (typically up to order 5) with adaptive step control to optimize performance across problem scales.

Theoretical Properties

Consistency and Order Conditions

In linear multistep methods, consistency refers to the property that the local approaches zero as the step size hh tends to zero, ensuring the method approximates the exact solution of the underlying . For a general kk-step method given by j=0kαjyn+j=hj=0kβjf(tn+j,yn+j),\sum_{j=0}^k \alpha_j y_{n+j} = h \sum_{j=0}^k \beta_j f(t_{n+j}, y_{n+j}), with αk=1\alpha_k = 1 and not all αj,βj=0\alpha_j, \beta_j = 0, consistency requires the first-order conditions derived from Taylor expansion: j=0kαj=0\sum_{j=0}^k \alpha_j = 0 and j=0kjαj=j=0kβj\sum_{j=0}^k j \alpha_j = \sum_{j=0}^k \beta_j. These conditions ensure that constant and linear solutions are reproduced exactly, corresponding to the method having order at least 1. The order pp of the method is defined such that the local τ=O(hp+1)\tau = O(h^{p+1}), meaning the method matches the exact solution up to terms of order pp in the expansion. This is achieved when the characteristic polynomials ρ(ζ)=j=0kαjζj\rho(\zeta) = \sum_{j=0}^k \alpha_j \zeta^j and σ(ζ)=j=0kβjζj\sigma(\zeta) = \sum_{j=0}^k \beta_j \zeta^j satisfy the order conditions that their derivatives align up to order pp. Specifically, the C(q)C(q) conditions for q=0,1,,pq = 0, 1, \dots, p are Cq=1q!j=0kαjjq1(q1)!j=0kβjjq1=0(q1),C_q = \frac{1}{q!} \sum_{j=0}^k \alpha_j j^q - \frac{1}{(q-1)!} \sum_{j=0}^k \beta_j j^{q-1} = 0 \quad (q \geq 1), with C0=j=0kαj=0C_0 = \sum_{j=0}^k \alpha_j = 0, and Cp+10C_{p+1} \neq 0 serving as the principal error constant. These conditions generalize the consistency requirements, ensuring higher-order accuracy by equating coefficients in the expansion. To derive these conditions, assume the exact solution y(t)y(t) satisfies the method with yn+j=y(tn+j)y_{n+j} = y(t_{n+j}). Expand y(tn+j)y(t_{n+j}) and y(tn+j)y'(t_{n+j}) in Taylor series around tnt_n: y(tn+j)=m=0(jh)mm!y(m)(tn),y(tn+j)=m=0(jh)mm!y(m+1)(tn).y(t_{n+j}) = \sum_{m=0}^\infty \frac{(j h)^m}{m!} y^{(m)}(t_n), \quad y'(t_{n+j}) = \sum_{m=0}^\infty \frac{(j h)^m}{m!} y^{(m+1)}(t_n). Substituting into the method yields the local truncation error hτn=j=0kαjy(tn+j)hj=0kβjy(tn+j)=q=0hqCqy(q)(tn),h \tau_n = \sum_{j=0}^k \alpha_j y(t_{n+j}) - h \sum_{j=0}^k \beta_j y'(t_{n+j}) = \sum_{q=0}^\infty h^q C_q y^{(q)}(t_n), where the coefficients CqC_q must vanish for q=0q = 0 to pp to achieve order pp. This derivation shows that solving the system of equations for the αj\alpha_j and βj\beta_j to satisfy the C(q)=0C(q) = 0 conditions determines methods of desired order. For example, the forward , a 1-step explicit method with α0=1\alpha_0 = -1, α1=1\alpha_1 = 1, β0=1\beta_0 = 1, β1=0\beta_1 = 0, satisfies C0=0C_0 = 0 and C1=0C_1 = 0, confirming order p=1p=1, with error constant C2=1/2C_2 = 1/2. Higher-order methods require solving larger systems; for instance, a 2-step method of order 2 involves setting C0=C1=C2=0C_0 = C_1 = C_2 = 0 to find appropriate coefficients, as in the .

Stability and Convergence

Zero-stability is a crucial property for linear multistep methods, ensuring that the numerical solution does not diverge as the step size hh approaches zero due to accumulated errors or perturbations in values. It is defined by the roots of the first ρ(ζ)=j=0kαjζj=0\rho(\zeta) = \sum_{j=0}^k \alpha_j \zeta^j = 0. The method satisfies the root condition if all ζ\zeta obey ζ1|\zeta| \leq 1, and any root with ζ=1|\zeta| = 1 (except the simple root at ζ=1\zeta = 1) has multiplicity one; the root ζ=1\zeta = 1 must be simple to preserve the constant solution. This condition prevents parasitic solutions from growing exponentially, thereby bounding error propagation over many steps. Convergence of a linear multistep method, which guarantees that the numerical solution approaches the exact solution as h0h \to 0, is characterized by Dahlquist's equivalence theorem: the method converges if and only if it is both consistent and zero-stable. For a consistent method of order pp that is also zero-stable, the global error satisfies yny(tn)=O(hp)|y_n - y(t_n)| = O(h^p) uniformly on any finite interval [t0,T][t_0, T], assuming sufficiently accurate starting values. The proof relies on perturbation , where the numerical solution is viewed as a small deviation from the exact solution; zero-stability ensures that local truncation errors, which are O(hp+1)O(h^{p+1}) per step due to consistency, accumulate to a global error of O(hp)O(h^p) without amplification. Consistency serves as a prerequisite, providing the local accuracy needed for this error bound. While zero-stability addresses asymptotic behavior as h0h \to 0, absolute stability examines the method's performance for fixed h>0h > 0, particularly in problems with widely varying timescales. For the scalar test equation y=λyy' = \lambda y with Re(λ)<0\operatorname{Re}(\lambda) < 0, the method applied with step size hh yields the recurrence involving the stability function R(z)R(z), where z=hλz = h\lambda. The method is absolutely stable if R(z)1|R(z)| \leq 1, and for damping of the numerical solution, R(z)<1|R(z)| < 1 in the stability region S={zC:R(z)1}S = \{ z \in \mathbb{C} : |R(z)| \leq 1 \}, which lies in the left half of the complex plane. This region determines where the method mimics the exact solution's decay. The implications of absolute stability are particularly evident in distinguishing stiff from non-stiff problems. Explicit linear multistep methods, such as those in the Adams-Bashforth family, have limited stability regions that exclude parts of the negative real axis for large z|z|, leading to instability when hλ|h\lambda| is large—common in stiff equations with eigenvalues far in the left half-plane. Implicit methods, like , often possess larger stability regions encompassing significant portions of the left half-plane, enabling reliable integration of stiff systems without excessive step-size restrictions.

Advanced Limitations

Second Dahlquist Barrier

The second Dahlquist barrier establishes that no linear multistep method of order p>2p > 2 can be A-, meaning it cannot remain for all step sizes h>0h > 0 when applied to the test equation y=λyy' = \lambda y with Re(λ)<0\operatorname{Re}(\lambda) < 0. This limitation arises because A-stability requires the absolute stability region to encompass the entire negative half of the complex plane, a property incompatible with higher-order accuracy in this class of methods. The theorem was proved by Germund Dahlquist in his 1963 paper, marking a pivotal result in the stability analysis of numerical ODE solvers. The proof analyzes the stability function R(z)R(z), a rational function approximating eze^z up to order pp, but on the imaginary axis for A-stability. Specifically, A-stability implies R(iy)1|R(i y)| \leq 1 for all real yy, as this forms the boundary of the left half-plane. For a method of order p>2p > 2, asymptotically for large y|y|, R(iy)|R(i y)| grows like yp2>1|y|^{p-2} > 1, violating the boundedness condition. This contradiction is resolved by showing that only orders up to 2 permit the required bounded behavior, often using the bilinear transformation z=(ξ+1)/(ξ1)z = (\xi + 1)/(\xi - 1) and properties of analytic functions such as the Riesz-Herglotz theorem. This barrier has profound implications for solving stiff ordinary differential equations, where eigenvalues with large negative real parts demand A-stability to avoid severe step-size restrictions for stability. Explicit linear multistep methods, lacking implicit terms, cannot achieve A-stability even at order 1, as their stability regions are bounded and exclude parts of the left half-plane. Consequently, while implicit methods enable A-stability up to order 2, higher-order approximations for stiff problems must compromise on full A-stability, often relying on related concepts like A(\alpha)-stability. An important exception is the , a second-order implicit method given by yn+1=yn+h2(f(tn,yn)+f(tn+1,yn+1))y_{n+1} = y_n + \frac{h}{2} (f(t_n, y_n) + f(t_{n+1}, y_{n+1})), which is A-stable and attains the minimal error constant among such methods. This makes it a benchmark for balanced accuracy and stability in non-stiff to mildly stiff contexts.

L-Stability Barrier

The L-stability barrier establishes a fundamental limitation on the design of linear multistep methods for stiff ordinary differential equations, stating that no L-stable method of order p>2p > 2 exists. L-stability requires not only A-stability—where the stability function R(z)R(z) satisfies R(z)1|R(z)| \leq 1 for all zz in the left half-plane (z)0\Re(z) \leq 0—but also that R(z)0|R(z)| \to 0 as z|z| \to \infty with (z)0\Re(z) \leq 0, ensuring rapid damping of transient components in stiff systems. This barrier extends the earlier restriction on A-stability alone, emphasizing the stricter demands for handling severe where eigenvalues have large negative real parts. Dahlquist introduced this barrier in 1985 as an extension of his earlier work on A-stability limits, highlighting its role in refining for stiff ODEs. The proof relies on the structure of the stability function R(z)R(z) for linear multistep methods, a whose numerator and denominator are polynomials of degrees typically equal to the number of steps kk. For L-stability, as zz \to -\infty requires R(z)c/zR(z) \sim c/z for some constant cc, implying a specific pole at to achieve the necessary decay. However, achieving order p>2p > 2 imposes Taylor expansion conditions on R(z)R(z) around z=0z = 0 that conflict with this asymptotic behavior, resulting in a degree mismatch that prevents the residue from decaying sufficiently while maintaining A-stability across the entire left half-plane. This incompatibility arises because higher-order accuracy demands more terms in the expansion, which disrupts the required rational form for L-stability. This limitation has significant implications for practical methods like the backward differentiation formulas (BDFs), which are widely used for stiff problems to their good stability properties. The BDF methods of orders 1 and 2 are both A-stable and L-stable, allowing effective treatment of stiff systems with fast-decaying transients. In contrast, BDF methods of orders 3 through 6 are only A(\alpha)-stable for some α>0\alpha > 0 (with α\alpha increasing as order rises, reaching about 73° for order 6), meaning they are not fully A-stable and thus cannot be L-stable; this can lead to slower decay of numerical transients in highly stiff scenarios, potentially requiring smaller steps or causing oscillations. Higher-order BDFs (beyond 6) lose even A(\alpha)-stability entirely. A common practical is order reduction near infinity, where the method effectively behaves like a lower-order (L-stable) formula for components with large λh|\lambda| h, achieved through techniques such as blending or variable-order selection in solvers; this mitigates the barrier's impact without sacrificing overall efficiency for moderately stiff problems.

References

Add your contribution
Related Hubs
User Avatar
No comments yet.