Recent from talks
Contribute something
Nothing was collected or created yet.
Finite difference method
View on Wikipedia| Differential equations |
|---|
| Scope |
| Classification |
| Solution |
| People |
In numerical analysis, finite-difference methods (FDM) are a class of numerical techniques for solving differential equations by approximating derivatives with finite differences. Both the spatial domain and time domain (if applicable) are discretized, or broken into a finite number of intervals, and the values of the solution at the end points of the intervals are approximated by solving algebraic equations containing finite differences and values from nearby points.
Finite difference methods convert ordinary differential equations (ODE) or partial differential equations (PDE), which may be nonlinear, into a system of linear equations that can be solved by matrix algebra techniques. Modern computers can perform these linear algebra computations efficiently, and this, along with their relative ease of implementation, has led to the widespread use of FDM in modern numerical analysis.[1] Today, FDMs are one of the most common approaches to the numerical solution of PDE, along with finite element methods.[1]
Derive difference quotient from Taylor's polynomial
[edit]For a n-times differentiable function, by Taylor's theorem the Taylor series expansion is given as
Where n! denotes the factorial of n, and Rn(x) is a remainder term, denoting the difference between the Taylor polynomial of degree n and the original function.
Following is the process to derive an approximation for the first derivative of the function f by first truncating the Taylor polynomial plus remainder: Dividing across by h gives: Solving for :
Assuming that is sufficiently small, the approximation of the first derivative of f is:
This is similar to the definition of derivative, which is: except for the limit towards zero (the method is named after this).
Accuracy and order
[edit]The error in a method's solution is defined as the difference between the approximation and the exact analytical solution. The two sources of error in finite difference methods are round-off error, the loss of precision due to computer rounding of decimal quantities, and truncation error or discretization error, the difference between the exact solution of the original differential equation and the exact quantity assuming perfect arithmetic (no round-off).

To use a finite difference method to approximate the solution to a problem, one must first discretize the problem's domain. This is usually done by dividing the domain into a uniform grid (see image). This means that finite-difference methods produce sets of discrete numerical approximations to the derivative, often in a "time-stepping" manner.
An expression of general interest is the local truncation error of a method. Typically expressed using Big-O notation, local truncation error refers to the error from a single application of a method. That is, it is the quantity if refers to the exact value and to the numerical approximation. The remainder term of the Taylor polynomial can be used to analyze local truncation error. Using the Lagrange form of the remainder from the Taylor polynomial for , which is the dominant term of the local truncation error can be discovered. For example, again using the forward-difference formula for the first derivative, knowing that , and with some algebraic manipulation, this leads to and further noting that the quantity on the left is the approximation from the finite difference method and that the quantity on the right is the exact quantity of interest plus a remainder, clearly that remainder is the local truncation error. A final expression of this example and its order is:
In this case, the local truncation error is proportional to the step sizes. The quality and duration of simulated FDM solution depends on the discretization equation selection and the step sizes (time and space steps). The data quality and simulation duration increase significantly with smaller step size.[2] Therefore, a reasonable balance between data quality and simulation duration is necessary for practical usage. Large time steps are useful for increasing simulation speed in practice. However, time steps which are too large may create instabilities and affect the data quality.[3][4]
The von Neumann and Courant-Friedrichs-Lewy criteria are often evaluated to determine the numerical model stability.[3][4][5][6]
Example: ordinary differential equation
[edit]For example, consider the ordinary differential equation The Euler method for solving this equation uses the finite difference quotient to approximate the differential equation by first substituting it for then applying a little algebra (multiplying both sides by h, and then adding to both sides) to get The last equation is a finite-difference equation, and solving this equation gives an approximate solution to the differential equation.
Example: The heat equation
[edit]Consider the normalized heat equation in one dimension, with homogeneous Dirichlet boundary conditions
One way to numerically solve this equation is to approximate all the derivatives by finite differences. First partition the domain in space using a mesh and in time using a mesh . Assume a uniform partition both in space and in time, so the difference between two consecutive space points will be h and between two consecutive time points will be k. The points
will represent the numerical approximation of
Explicit method
[edit]
Using a forward difference at time and a second-order central difference for the space derivative at position (FTCS) gives the recurrence equation:
This is an explicit method for solving the one-dimensional heat equation.
One can obtain from the other values this way:
where
So, with this recurrence relation, and knowing the values at time n, one can obtain the corresponding values at time n+1. and must be replaced by the boundary conditions, in this example they are both 0.
This explicit method is known to be numerically stable and convergent whenever .[7] The numerical errors are proportional to the time step and the square of the space step:
Implicit method
[edit]
Using the backward difference at time and a second-order central difference for the space derivative at position (The Backward Time, Centered Space Method "BTCS") gives the recurrence equation:
This is an implicit method for solving the one-dimensional heat equation.
One can obtain from solving a system of linear equations:
The scheme is always numerically stable and convergent but usually more numerically intensive than the explicit method as it requires solving a system of numerical equations on each time step. The errors are linear over the time step and quadratic over the space step:
Crank–Nicolson method
[edit]Finally, using the central difference at time and a second-order central difference for the space derivative at position ("CTCS") gives the recurrence equation:
This formula is known as the Crank–Nicolson method.

One can obtain from solving a system of linear equations:
The scheme is always numerically stable and convergent but usually more numerically intensive as it requires solving a system of numerical equations on each time step. The errors are quadratic over both the time step and the space step:
Comparison
[edit]To summarize, usually the Crank–Nicolson scheme is the most accurate scheme for small time steps. For larger time steps, the implicit scheme works better since it is less computationally demanding. The explicit scheme is the least accurate and can be unstable, but is also the easiest to implement and the least numerically intensive.
Here is an example. The figures below present the solutions given by the above methods to approximate the heat equation
with the boundary condition
The exact solution is
Example: The Laplace operator
[edit]The (continuous) Laplace operator in -dimensions is given by . The discrete Laplace operator depends on the dimension .
In 1D the Laplace operator is approximated as This approximation is usually expressed via the following stencil and which represents a symmetric, tridiagonal matrix. For an equidistant grid one gets a Toeplitz matrix.
The 2D case shows all the characteristics of the more general n-dimensional case. Each second partial derivative needs to be approximated similar to the 1D case which is usually given by the following stencil
Consistency
[edit]Consistency of the above-mentioned approximation can be shown for highly regular functions, such as . The statement is
To prove this, one needs to substitute Taylor Series expansions up to order 3 into the discrete Laplace operator.
Properties
[edit]Subharmonic
[edit]Similar to continuous subharmonic functions one can define subharmonic functions for finite-difference approximations
Mean value
[edit]One can define a general stencil of positive type via
If is (discrete) subharmonic then the following mean value property holds where the approximation is evaluated on points of the grid, and the stencil is assumed to be of positive type.
A similar mean value property also holds for the continuous case.
Maximum principle
[edit]For a (discrete) subharmonic function the following holds where are discretizations of the continuous domain , respectively the boundary .
A similar maximum principle also holds for the continuous case.
The SBP-SAT method
[edit]The SBP-SAT (summation by parts - simultaneous approximation term) method is a stable and accurate technique for discretizing and imposing boundary conditions of a well-posed linear partial differential equation using high order finite differences.[8][9]
The method is based on finite differences where the differentiation operators exhibit summation-by-parts properties. Typically, these operators consist of differentiation matrices with central difference stencils in the interior with carefully chosen one-sided boundary stencils designed to mimic integration-by-parts in the discrete setting. Using the SAT technique, the boundary conditions of the PDE are imposed weakly, where the boundary values are "pulled" towards the desired conditions rather than exactly fulfilled. If the tuning parameters (inherent to the SAT technique) are chosen properly, the resulting system of ODE's will exhibit similar energy behavior as the continuous PDE, i.e. the system has no non-physical energy growth. This guarantees stability if an integration scheme with a stability region that includes parts of the imaginary axis, such as the fourth order Runge-Kutta method, is used. This makes the SAT technique an attractive method of imposing boundary conditions for higher order finite difference methods, in contrast to for example the injection method, which typically will not be stable if high order differentiation operators are used.
See also
[edit]- Finite element method
- Finite difference
- Finite difference time domain
- Infinite difference method
- Stencil (numerical analysis)
- Finite difference coefficients
- Five-point stencil
- Lax–Richtmyer theorem
- Finite difference methods for option pricing
- Upwind differencing scheme for convection
- Central differencing scheme
- Discrete Poisson equation
- Discrete Laplace operator
References
[edit]- ^ a b Christian Grossmann; Hans-G. Roos; Martin Stynes (2007). Numerical Treatment of Partial Differential Equations. Springer Science & Business Media. p. 23. ISBN 978-3-540-71584-9.
- ^ Arieh Iserles (2008). A first course in the numerical analysis of differential equations. Cambridge University Press. p. 23. ISBN 9780521734905.
- ^ a b Hoffman JD; Frankel S (2001). Numerical methods for engineers and scientists. CRC Press, Boca Raton.
- ^ a b Jaluria Y; Atluri S (1994). "Computational heat transfer". Computational Mechanics. 14 (5): 385–386. Bibcode:1994CompM..14..385J. doi:10.1007/BF00377593. S2CID 119502676.
- ^ Majumdar P (2005). Computational methods for heat and mass transfer (1st ed.). Taylor and Francis, New York.
- ^ Smith GD (1985). Numerical solution of partial differential equations: finite difference methods (3rd ed.). Oxford University Press.
- ^ Crank, J. The Mathematics of Diffusion. 2nd Edition, Oxford, 1975, p. 143.
- ^ Bo Strand (1994). "Summation by Parts for Finite Difference Approximations for d/dx". Journal of Computational Physics. 110 (1): 47–67. Bibcode:1994JCoPh.110...47S. doi:10.1006/jcph.1994.1005.
- ^ Mark H. Carpenter; David I. Gottlieb; Saul S. Abarbanel (1994). "Time-stable boundary conditions for finite-difference schemes solving hyperbolic systems: Methodology and application to high-order compact schemes". Journal of Computational Physics. 111 (2): 220–236. Bibcode:1994JCoPh.111..220C. doi:10.1006/jcph.1994.1057. hdl:2060/19930013937.
Further reading
[edit]- K.W. Morton and D.F. Mayers, Numerical Solution of Partial Differential Equations, An Introduction. Cambridge University Press, 2005.
- Autar Kaw and E. Eric Kalu, Numerical Methods with Applications, (2008) [1]. Contains a brief, engineering-oriented introduction to FDM (for ODEs) in Chapter 08.07.
- John Strikwerda (2004). Finite Difference Schemes and Partial Differential Equations (2nd ed.). SIAM. ISBN 978-0-89871-639-9.
- Smith, G. D. (1985), Numerical Solution of Partial Differential Equations: Finite Difference Methods, 3rd ed., Oxford University Press
- Peter Olver (2013). Introduction to Partial Differential Equations. Springer. Chapter 5: Finite differences. ISBN 978-3-319-02099-0..
- Randall J. LeVeque, Finite Difference Methods for Ordinary and Partial Differential Equations, SIAM, 2007.
- Sergey Lemeshevsky, Piotr Matus, Dmitriy Poliakov(Eds): "Exact Finite-Difference Schemes", De Gruyter (2016). DOI: https://doi.org/10.1515/9783110491326 .
- Mikhail Shashkov: Conservative Finite-Difference Methods on General Grids, CRC Press, ISBN 0-8493-7375-1 (1996).
Finite difference method
View on GrokipediaFundamentals
Overview and Definition
The finite difference method is a numerical technique used to approximate solutions to differential equations by discretizing continuous domains into a finite grid of points and replacing derivatives with discrete difference quotients.[1] This approach transforms the continuous problem into a system of algebraic equations that can be solved computationally, making it particularly suitable for problems where analytical solutions are unavailable or impractical.[5] It assumes a basic familiarity with differential equations but focuses on the grid-based discretization process, where the domain is divided into evenly spaced points to enable finite approximations of rates of change. In standard notation, consider a one-dimensional domain discretized with grid spacing , yielding points for integer , and corresponding function values . The first derivative can be approximated using forward, backward, or central differences: the forward difference is , the backward difference is , and the central difference is .[7] For the second derivative , a common central difference approximation is .[5] These operators provide the foundational building blocks for constructing finite difference schemes across higher dimensions and more complex equations. The origins of finite differences trace back to 18th-century interpolation techniques developed by Isaac Newton and Leonhard Euler, who used discrete differences to approximate continuous functions.[8] The method was formalized for solving partial differential equations in the early 20th century, notably by Lewis Fry Richardson in his 1910 work on approximate arithmetical solutions to physical problems involving differential equations.[9] This laid the groundwork for modern applications in numerical analysis.Derivation from Taylor Series
The finite difference approximations for derivatives are derived from the Taylor series expansion of a smooth function around a given point. For a function that is sufficiently differentiable, the Taylor theorem states that for some between and .[10] Rearranging the expansion for yields the forward difference approximation to the first derivative: [11] The backward difference approximation is obtained analogously from the expansion for some between and , giving [12] The central difference for the first derivative combines the expansions for and , where the linear terms cancel, resulting in [13] For the second derivative, adding the expansions for and eliminates the odd-powered terms, yielding the central difference formula [14] Approximations for higher-order derivatives follow a similar process, employing Taylor expansions at multiple grid points and matching coefficients through the method of undetermined coefficients to achieve the desired accuracy order.[15]Accuracy and Convergence Order
The truncation error in finite difference methods represents the discrepancy between the exact mathematical derivative and its discrete approximation on a finite grid.[16] This error arises primarily from the truncation of the infinite Taylor series expansion used to derive the approximations.[17] The leading error terms are obtained from the remainders in Taylor expansions. For the forward difference approximation of the first derivative, for some , the truncation error is .[17] The backward difference, , similarly yields an error.[17] In contrast, the central difference for the first derivative, for some , has a leading error of .[18] For the second derivative using the central difference, the truncation error is also .[18] The order of accuracy of a finite difference approximation is defined as the largest integer such that the truncation error is as the grid spacing approaches zero.[19] Higher-order methods reduce the error more rapidly with decreasing , enabling more efficient computations for a given accuracy level.[19] For schemes applied to partial differential equations, consistency requires that the local truncation error vanishes as .[4] This property ensures that the discrete equations approach the continuous PDE in the limit, forming a prerequisite for convergence of the numerical solution to the exact one.[4] Although reducing diminishes truncation error, it exacerbates round-off errors from finite-precision arithmetic, as more grid points amplify accumulation of floating-point inaccuracies.[4] Optimal accuracy thus involves selecting an that balances these competing error sources.[4]Applications to Ordinary Differential Equations
First-Order Initial Value Problems
The finite difference method provides a framework for numerically solving first-order initial value problems by approximating the derivative with discrete differences on a uniform grid. The model problem is the ordinary differential equation with initial condition over the interval .[20] To apply the method, the interval is discretized into equal steps of size , yielding grid points for , where approximates . This approach relies on finite difference quotients to replace the continuous derivative, enabling a step-by-step march from the initial condition.[21] The forward Euler method employs the forward difference quotient , which, when substituted into the differential equation, yields the explicit update formula with .[22] This scheme is straightforward to implement and computationally efficient, as each step requires only evaluation of the known function at the current point. The local truncation error, arising from the Taylor series approximation of the exact solution, is , while the accumulated global error over the interval is , establishing it as a first-order accurate method.[23] In contrast, the backward Euler method uses the backward difference quotient , leading to the implicit relation This requires solving an equation for at each step, which may involve nonlinear solvers if is nonlinear, but the method offers improved stability for certain problems.[24] The trapezoidal rule combines the forward and backward approaches by averaging the right-hand side evaluations: This implicit scheme achieves second-order accuracy, with local truncation error and global error , providing a balance between precision and the need for iterative solution at each step.[25] A basic pseudocode implementation for the forward Euler method is as follows:function forward_euler(f, x0, u0, k, N):
x = x0
u = u0
for n = 0 to N-1:
u_next = u + k * f(x, u)
x = x + k
u = u_next
return x, u
function forward_euler(f, x0, u0, k, N):
x = x0
u = u0
for n = 0 to N-1:
u_next = u + k * f(x, u)
x = x + k
u = u_next
return x, u
