Hubbry Logo
Function of several real variablesFunction of several real variablesMain
Open search
Function of several real variables
Community hub
Function of several real variables
logo
7 pages, 0 posts
0 subscribers
Be the first to start a discussion here.
Be the first to start a discussion here.
Function of several real variables
Function of several real variables
from Wikipedia

In mathematical analysis and its applications, a function of several real variables or real multivariate function is a function with more than one argument, with all arguments being real variables. This concept extends the idea of a function of a real variable to several variables. The "input" variables take real values, while the "output", also called the "value of the function", may be real or complex. However, the study of the complex-valued functions may be easily reduced to the study of the real-valued functions, by considering the real and imaginary parts of the complex function; therefore, unless explicitly specified, only real-valued functions will be considered in this article.

The domain of a function of n variables is the subset of for which the function is defined. As usual, the domain of a function of several real variables is supposed to contain a nonempty open subset of .

General definition

[edit]
n = 1
n = 2
n = 3
Functions f(x1, x2, …, xn) of n variables, plotted as graphs in the space Rn + 1. The domains are the red n-dimensional regions, the images are the purple n-dimensional curves.

A real-valued function of n real variables is a function that takes as input n real numbers, commonly represented by the variables x1, x2, …, xn, for producing another real number, the value of the function, commonly denoted f(x1, x2, …, xn). For simplicity, in this article a real-valued function of several real variables will be simply called a function. To avoid any ambiguity, the other types of functions that may occur will be explicitly specified.

Some functions are defined for all real values of the variables (one says that they are everywhere defined), but some other functions are defined only if the value of the variable are taken in a subset X of Rn, the domain of the function, which is always supposed to contain an open subset of Rn. In other words, a real-valued function of n real variables is a function

such that its domain X is a subset of Rn that contains a nonempty open set.

An element of X being an n-tuple (x1, x2, …, xn) (usually delimited by parentheses), the general notation for denoting functions would be f((x1, x2, …, xn)). The common usage, much older than the general definition of functions between sets, is to not use double parentheses and to simply write f(x1, x2, …, xn).

It is also common to abbreviate the n-tuple (x1, x2, …, xn) by using a notation similar to that for vectors, like boldface x, underline x, or overarrow x. This article will use bold.

A simple example of a function in two variables could be:

which is the volume V of a cone with base area A and height h measured perpendicularly from the base. The domain restricts all variables to be positive since lengths and areas must be positive.

For an example of a function in two variables:

where a and b are real non-zero constants. Using the three-dimensional Cartesian coordinate system, where the xy plane is the domain R2 and the z axis is the codomain R, one can visualize the image to be a two-dimensional plane, with a slope of a in the positive x direction and a slope of b in the positive y direction. The function is well-defined at all points (x, y) in R2. The previous example can be extended easily to higher dimensions:

for p non-zero real constants a1, a2, …, ap, which describes a p-dimensional hyperplane.

The Euclidean norm:

is also a function of n variables which is everywhere defined, while

is defined only for x ≠ (0, 0, …, 0).

For a non-linear example function in two variables:

which takes in all points in X, a disk of radius 8 "punctured" at the origin (x, y) = (0, 0) in the plane R2, and returns a point in R. The function does not include the origin (x, y) = (0, 0), if it did then f would be ill-defined at that point. Using a 3d Cartesian coordinate system with the xy-plane as the domain R2, and the z axis the codomain R, the image can be visualized as a curved surface.

The function can be evaluated at the point (x, y) = (2, 3) in X:

However, the function couldn't be evaluated at, say

since these values of x and y do not satisfy the domain's rule.

Image

[edit]

The image of a function f(x1, x2, …, xn) is the set of all values of f when the n-tuple (x1, x2, …, xn) runs in the whole domain of f. For a continuous (see below for a definition) real-valued function which has a connected domain, the image is either an interval or a single value. In the latter case, the function is a constant function.

The preimage of a given real number c is called a level set. It is the set of the solutions of the equation f(x1, x2, …, xn) = c.

Domain

[edit]

The domain of a function of several real variables is a subset of Rn that is sometimes, but not always, explicitly defined. In fact, if one restricts the domain X of a function f to a subset YX, one gets formally a different function, the restriction of f to Y, which is denoted . In practice, it is often (but not always) not harmful to identify f and , and to omit the restrictor |Y.

Conversely, it is sometimes possible to enlarge naturally the domain of a given function, for example by continuity or by analytic continuation.

Moreover, many functions are defined in such a way that it is difficult to specify explicitly their domain. For example, given a function f, it may be difficult to specify the domain of the function If f is a multivariate polynomial, (which has as a domain), it is even difficult to test whether the domain of g is also . This is equivalent to test whether a polynomial is always positive, and is the object of an active research area (see Positive polynomial).

Algebraic structure

[edit]

The usual operations of arithmetic on the reals may be extended to real-valued functions of several real variables in the following way:

  • For every real number r, the constant function is everywhere defined.
  • For every real number r and every function f, the function: has the same domain as f (or is everywhere defined if r = 0).
  • If f and g are two functions of respective domains X and Y such that XY contains a nonempty open subset of Rn, then and are functions that have a domain containing XY.

It follows that the functions of n variables that are everywhere defined and the functions of n variables that are defined in some neighbourhood of a given point both form commutative algebras over the reals (R-algebras). This is a prototypical example of a function space.

One may similarly define

which is a function only if the set of the points (x1, …,xn) in the domain of f such that f(x1, …, xn) ≠ 0 contains an open subset of Rn. This constraint implies that the above two algebras are not fields.

Univariable functions associated with a multivariable function

[edit]

One can easily obtain a function in one real variable by giving a constant value to all but one of the variables. For example, if (a1, …, an) is a point of the interior of the domain of the function f, we can fix the values of x2, …, xn to a2, …, an respectively, to get a univariable function

whose domain contains an interval centered at a1. This function may also be viewed as the restriction of the function f to the line defined by the equations xi = ai for i = 2, …, n.

Other univariable functions may be defined by restricting f to any line passing through (a1, …, an). These are the functions

where the ci are real numbers that are not all zero.

In next section, we will show that, if the multivariable function is continuous, so are all these univariable functions, but the converse is not necessarily true.

Continuity and limit

[edit]

Until the second part of 19th century, only continuous functions were considered by mathematicians. At that time, the notion of continuity was elaborated for the functions of one or several real variables a rather long time before the formal definition of a topological space and a continuous map between topological spaces. As continuous functions of several real variables are ubiquitous in mathematics, it is worth to define this notion without reference to the general notion of continuous maps between topological space.

For defining the continuity, it is useful to consider the distance function of Rn, which is an everywhere defined function of 2n real variables:

A function f is continuous at a point a = (a1, …, an) which is interior to its domain, if, for every positive real number ε, there is a positive real number φ such that |f(x) − f(a)| < ε for all x such that d(x a) < φ. In other words, φ may be chosen small enough for having the image by f of the ball of radius φ centered at a contained in the interval of length 2ε centered at f(a). A function is continuous if it is continuous at every point of its domain.

If a function is continuous at f(a), then all the univariate functions that are obtained by fixing all the variables xi except one at the value ai, are continuous at f(a). The converse is false; this means that all these univariate functions may be continuous for a function that is not continuous at f(a). For an example, consider the function f such that f(0, 0) = 0, and is otherwise defined by

The functions xf(x, 0) and yf(0, y) are both constant and equal to zero, and are therefore continuous. The function f is not continuous at (0, 0), because, if ε < 1/2 and y = x2 ≠ 0, we have f(x, y) = 1/2, even if |x| is very small. Although not continuous, this function has the further property that all the univariate functions obtained by restricting it to a line passing through (0, 0) are also continuous. In fact, we have

for λ ≠ 0.

The limit at a point of a real-valued function of several real variables is defined as follows.[1] Let a = (a1, a2, …, an) be a point in topological closure of the domain X of the function f. The function, f has a limit L when x tends toward a, denoted

if the following condition is satisfied: For every positive real number ε > 0, there is a positive real number δ > 0 such that

for all x in the domain such that

If the limit exists, it is unique. If a is in the interior of the domain, the limit exists if and only if the function is continuous at a. In this case, we have

When a is in the boundary of the domain of f, and if f has a limit at a, the latter formula allows to "extend by continuity" the domain of f to a.

Symmetry

[edit]

A symmetric function is a function f that is unchanged when two variables xi and xj are interchanged:

where i and j are each one of 1, 2, …, n. For example:

is symmetric in x, y, z since interchanging any pair of x, y, z leaves f unchanged, but is not symmetric in all of x, y, z, t, since interchanging t with x or y or z gives a different function.

Function composition

[edit]

Suppose the functions

or more compactly ξ = ξ(x), are all defined on a domain X. As the n-tuple x = (x1, x2, …, xn) varies in X, a subset of Rn, the m-tuple ξ = (ξ1, ξ2, …, ξm) varies in another region Ξ a subset of Rm. To restate this:

Then, a function ζ of the functions ξ(x) defined on Ξ,

is a function composition defined on X,[2] in other terms the mapping

Note the numbers m and n do not need to be equal.

For example, the function

defined everywhere on R2 can be rewritten by introducing

which is also everywhere defined in R3 to obtain

Function composition can be used to simplify functions, which is useful for carrying out multiple integrals and solving partial differential equations.

Calculus

[edit]

Elementary calculus is the calculus of real-valued functions of one real variable, and the principal ideas of differentiation and integration of such functions can be extended to functions of more than one real variable; this extension is multivariable calculus.

Partial derivatives

[edit]

Partial derivatives can be defined with respect to each variable:

Partial derivatives themselves are functions, each of which represents the rate of change of f parallel to one of the x1, x2, …, xn axes at all points in the domain (if the derivatives exist and are continuous—see also below). A first derivative is positive if the function increases along the direction of the relevant axis, negative if it decreases, and zero if there is no increase or decrease. Evaluating a partial derivative at a particular point in the domain gives the rate of change of the function at that point in the direction parallel to a particular axis, a real number.

For real-valued functions of a real variable, y = f(x), its ordinary derivative dy/dx is geometrically the gradient of the tangent line to the curve y = f(x) at all points in the domain. Partial derivatives extend this idea to tangent hyperplanes to a curve.

The second order partial derivatives can be calculated for every pair of variables:

Geometrically, they are related to the local curvature of the function's image at all points in the domain. At any point where the function is well-defined, the function could be increasing along some axes, and/or decreasing along other axes, and/or not increasing or decreasing at all along other axes.

This leads to a variety of possible stationary points: global or local maxima, global or local minima, and saddle points—the multidimensional analogue of inflection points for real functions of one real variable. The Hessian matrix is a matrix of all the second order partial derivatives, which are used to investigate the stationary points of the function, important for mathematical optimization.

In general, partial derivatives of higher order p have the form:

where p1, p2, …, pn are each integers between 0 and p such that p1 + p2 + ⋯ + pn = p, using the definitions of zeroth partial derivatives as identity operators:

The number of possible partial derivatives increases with p, although some mixed partial derivatives (those with respect to more than one variable) are superfluous, because of the symmetry of second order partial derivatives. This reduces the number of partial derivatives to calculate for some p.

Multivariable differentiability

[edit]

A function f(x) is differentiable in a neighborhood of a point a if there is an n-tuple of numbers dependent on a in general, A(a) = (A1(a), A2(a), …, An(a)), so that:[3]

where as . This means that if f is differentiable at a point a, then f is continuous at x = a, although the converse is not true - continuity in the domain does not imply differentiability in the domain. If f is differentiable at a then the first order partial derivatives exist at a and:

for i = 1, 2, …, n, which can be found from the definitions of the individual partial derivatives, so the partial derivatives of f exist.

Assuming an n-dimensional analogue of a rectangular Cartesian coordinate system, these partial derivatives can be used to form a vectorial linear differential operator, called the gradient (also known as "nabla" or "del") in this coordinate system:

used extensively in vector calculus, because it is useful for constructing other differential operators and compactly formulating theorems in vector calculus.

Then substituting the gradient f (evaluated at x = a) with a slight rearrangement gives:

where · denotes the dot product. This equation represents the best linear approximation of the function f at all points x within a neighborhood of a. For infinitesimal changes in f and x as xa:

which is defined as the total differential, or simply differential, of f, at a. This expression corresponds to the total infinitesimal change of f, by adding all the infinitesimal changes of f in all the xi directions. Also, df can be construed as a covector with basis vectors as the infinitesimals dxi in each direction and partial derivatives of f as the components.

Geometrically f is perpendicular to the level sets of f, given by f(x) = c which for some constant c describes an (n − 1)-dimensional hypersurface. The differential of a constant is zero:

in which dx is an infinitesimal change in x in the hypersurface f(x) = c, and since the dot product of f and dx is zero, this means f is perpendicular to dx.

In arbitrary curvilinear coordinate systems in n dimensions, the explicit expression for the gradient would not be so simple - there would be scale factors in terms of the metric tensor for that coordinate system. For the above case used throughout this article, the metric is just the Kronecker delta and the scale factors are all 1.

Differentiability classes

[edit]

If all first order partial derivatives evaluated at a point a in the domain:

exist and are continuous for all a in the domain, f has differentiability class C1. In general, if all order p partial derivatives evaluated at a point a:

exist and are continuous, where p1, p2, …, pn, and p are as above, for all a in the domain, then f is differentiable to order p throughout the domain and has differentiability class C p.

If f is of differentiability class C, f has continuous partial derivatives of all order and is called smooth. If f is an analytic function and equals its Taylor series about any point in the domain, the notation Cω denotes this differentiability class.

Multiple integration

[edit]

Definite integration can be extended to multiple integration over the several real variables with the notation;

where each region R1, R2, …, Rn is a subset of or all of the real line:

and their Cartesian product gives the region to integrate over as a single set:

an n-dimensional hypervolume. When evaluated, a definite integral is a real number if the integral converges in the region R of integration (the result of a definite integral may diverge to infinity for a given region, in such cases the integral remains ill-defined). The variables are treated as "dummy" or "bound" variables which are substituted for numbers in the process of integration.

The integral of a real-valued function of a real variable y = f(x) with respect to x has geometric interpretation as the area bounded by the curve y = f(x) and the x-axis. Multiple integrals extend the dimensionality of this concept: assuming an n-dimensional analogue of a rectangular Cartesian coordinate system, the above definite integral has the geometric interpretation as the n-dimensional hypervolume bounded by f(x) and the x1, x2, …, xn axes, which may be positive, negative, or zero, depending on the function being integrated (if the integral is convergent).

While bounded hypervolume is a useful insight, the more important idea of definite integrals is that they represent total quantities within space. This has significance in applied mathematics and physics: if f is some scalar density field and x are the position vector coordinates, i.e. some scalar quantity per unit n-dimensional hypervolume, then integrating over the region R gives the total amount of quantity in R. The more formal notions of hypervolume is the subject of measure theory. Above we used the Lebesgue measure, see Lebesgue integration for more on this topic.

Theorems

[edit]

With the definitions of multiple integration and partial derivatives, key theorems can be formulated, including the fundamental theorem of calculus in several real variables (namely Stokes' theorem), integration by parts in several real variables, the symmetry of higher partial derivatives and Taylor's theorem for multivariable functions. Evaluating a mixture of integrals and partial derivatives can be done by using theorem differentiation under the integral sign.

Vector calculus

[edit]

One can collect a number of functions each of several real variables, say

into an m-tuple, or sometimes as a column vector or row vector, respectively:

all treated on the same footing as an m-component vector field, and use whichever form is convenient. All the above notations have a common compact notation y = f(x). The calculus of such vector fields is vector calculus. For more on the treatment of row vectors and column vectors of multivariable functions, see matrix calculus.

Implicit functions

[edit]

A real-valued implicit function of several real variables is not written in the form "y = f(…)". Instead, the mapping is from the space Rn + 1 to the zero element in R (just the ordinary zero 0):

is an equation in all the variables. Implicit functions are a more general way to represent functions, since if:

then we can always define:

but the converse is not always possible, i.e. not all implicit functions have an explicit form.

For example, using interval notation, let

Choosing a 3-dimensional (3D) Cartesian coordinate system, this function describes the surface of a 3D ellipsoid centered at the origin (x, y, z) = (0, 0, 0) with constant semi-major axes a, b, c, along the positive x, y and z axes respectively. In the case a = b = c = r, we have a sphere of radius r centered at the origin. Other conic section examples which can be described similarly include the hyperboloid and paraboloid, more generally so can any 2D surface in 3D Euclidean space. The above example can be solved for x, y or z; however it is much tidier to write it in an implicit form.

For a more sophisticated example:

for non-zero real constants A, B, C, ω, this function is well-defined for all (t, x, y, z), but it cannot be solved explicitly for these variables and written as "t =", "x =", etc.

The implicit function theorem of more than two real variables deals with the continuity and differentiability of the function, as follows.[4] Let ϕ(x1, x2, …, xn) be a continuous function with continuous first order partial derivatives, and let ϕ evaluated at a point (a, b) = (a1, a2, …, an, b) be zero:

and let the first partial derivative of ϕ with respect to y evaluated at (a, b) be non-zero:

Then, there is an interval [y1, y2] containing b, and a region R containing (a, b), such that for every x in R there is exactly one value of y in [y1, y2] satisfying ϕ(x, y) = 0, and y is a continuous function of x so that ϕ(x, y(x)) = 0. The total differentials of the functions are:

Substituting dy into the latter differential and equating coefficients of the differentials gives the first order partial derivatives of y with respect to xi in terms of the derivatives of the original function, each as a solution of the linear equation

for i = 1, 2, …, n.

Complex-valued function of several real variables

[edit]

A complex-valued function of several real variables may be defined by relaxing, in the definition of the real-valued functions, the restriction of the codomain to the real numbers, and allowing complex values.

If f(x1, …, xn) is such a complex valued function, it may be decomposed as

where g and h are real-valued functions. In other words, the study of the complex valued functions reduces easily to the study of the pairs of real valued functions.

This reduction works for the general properties. However, for an explicitly given function, such as:

the computation of the real and the imaginary part may be difficult.

Applications

[edit]

Multivariable functions of real variables arise inevitably in engineering and physics, because observable physical quantities are real numbers (with associated units and dimensions), and any one physical quantity will generally depend on a number of other quantities.

Examples of real-valued functions of several real variables

[edit]

Examples in continuum mechanics include the local mass density ρ of a mass distribution, a scalar field which depends on the spatial position coordinates (here Cartesian to exemplify), r = (x, y, z), and time t:

Similarly for electric charge density for electrically charged objects, and numerous other scalar potential fields.

Another example is the velocity field, a vector field, which has components of velocity v = (vx, vy, vz) that are each multivariable functions of spatial coordinates and time similarly:

Similarly for other physical vector fields such as electric fields and magnetic fields, and vector potential fields.

Another important example is the equation of state in thermodynamics, an equation relating pressure P, temperature T, and volume V of a fluid, in general it has an implicit form:

The simplest example is the ideal gas law:

where n is the number of moles, constant for a fixed amount of substance, and R the gas constant. Much more complicated equations of state have been empirically derived, but they all have the above implicit form.

Real-valued functions of several real variables appear pervasively in economics. In the underpinnings of consumer theory, utility is expressed as a function of the amounts of various goods consumed, each amount being an argument of the utility function. The result of maximizing utility is a set of demand functions, each expressing the amount demanded of a particular good as a function of the prices of the various goods and of income or wealth. In producer theory, a firm is usually assumed to maximize profit as a function of the quantities of various goods produced and of the quantities of various factors of production employed. The result of the optimization is a set of demand functions for the various factors of production and a set of supply functions for the various products; each of these functions has as its arguments the prices of the goods and of the factors of production.

Examples of complex-valued functions of several real variables

[edit]

Some "physical quantities" may be actually complex valued - such as complex impedance, complex permittivity, complex permeability, and complex refractive index. These are also functions of real variables, such as frequency or time, as well as temperature.

In two-dimensional fluid mechanics, specifically in the theory of the potential flows used to describe fluid motion in 2d, the complex potential

is a complex valued function of the two spatial coordinates x and y, and other real variables associated with the system. The real part is the velocity potential and the imaginary part is the stream function.

The spherical harmonics occur in physics and engineering as the solution to Laplace's equation, as well as the eigenfunctions of the z-component angular momentum operator, which are complex-valued functions of real-valued spherical polar angles:

In quantum mechanics, the wavefunction is necessarily complex-valued, but is a function of real spatial coordinates (or momentum components), as well as time t:

where each is related by a Fourier transform.

See also

[edit]

References

[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
A function of several real variables, also known as a multivariable function, is a mapping from a subset of Euclidean space Rn\mathbb{R}^n (where n2n \geq 2) to Rm\mathbb{R}^m, where the input is an ordered tuple or vector of nn real numbers and the output is either a single real number (scalar-valued) or a tuple of mm real numbers (vector-valued). These functions extend the ideas of single-variable calculus to higher dimensions, allowing the modeling of phenomena that depend on multiple independent factors, such as temperature distributions in a room or profit maximization in economics. In mathematical analysis, they form the foundation for studying limits, continuity, differentiability, and integration in multiple dimensions. The domain of such a function is typically a in Rn\mathbb{R}^n, often visualized as a 2D area for n=2n=2 or generalized to higher-dimensional volumes, while the graph of a scalar-valued function f:RnRf: \mathbb{R}^n \to \mathbb{R} forms a hypersurface in Rn+1\mathbb{R}^{n+1}. For instance, in two variables, the graph is a surface in 3D space, and tools like level curves (sets where f(x,y)=cf(x,y) = c) or traces (intersections with coordinate planes) aid in understanding its shape. Continuity at a point aRna \in \mathbb{R}^n requires that the limit of f(x)f(x) as xx approaches aa (in the Euclidean norm) equals f(a)f(a), independent of the approach path, generalizing the single-variable notion. Differentiability extends to partial derivatives (with respect to each variable, holding others fixed) and the , represented by the matrix for vector-valued functions, which provides the best at a point. A function is differentiable at aa if limh0f(a+h)f(a)Df(a)hh=0\lim_{h \to 0} \frac{\|f(a + h) - f(a) - Df(a) \cdot h\|}{\|h\|} = 0, where Df(a)Df(a) is the linear transformation (). This concept is crucial for optimization, where critical points are found by setting partial derivatives to zero, with applications in physics (e.g., fields) and (e.g., ). Integration of functions of several variables involves multiple integrals over regions in Rn\mathbb{R}^n, such as double integrals for area under surfaces or triple integrals for volume, enabling computations like mass or center of gravity in applied settings. These integrals, often evaluated using iterated integrals via Fubini's theorem, underpin advanced topics like vector calculus theorems (Green's, Stokes', divergence) and partial differential equations. Overall, functions of several real variables are indispensable in modeling real-world multivariable dependencies across , , and .

Definition and Fundamentals

Formal Definition

A function of several real variables is formally defined as a mapping f:DRnRmf: D \subseteq \mathbb{R}^n \to \mathbb{R}^m, where n2n \geq 2 and m1m \geq 1, with DD denoting a nonempty subset of the Euclidean space Rn\mathbb{R}^n. This setup encompasses both scalar-valued functions (when m=1m = 1) and vector-valued functions (when m>1m > 1), providing a framework for analyzing dependencies among multiple inputs. Unlike functions of a single real variable, which map from subsets of R\mathbb{R} to R\mathbb{R} or Rm\mathbb{R}^m, functions of several real variables operate over higher-dimensional domains, enabling the study of phenomena in spaces like planes (n=2n=2) or volumes (n=3n=3), where interactions between variables introduce new geometric and analytical complexities. In the vector-valued case, the function takes the general form f(x1,x2,,xn)=(f1(x1,x2,,xn),,fm(x1,x2,,xn))f(x_1, x_2, \dots, x_n) = (f_1(x_1, x_2, \dots, x_n), \dots, f_m(x_1, x_2, \dots, x_n)), where each component fi:DRf_i: D \to \mathbb{R} is a real-valued function of the nn variables, and x=(x1,,xn)Dx = (x_1, \dots, x_n) \in D. The origins of this concept trace back to the 18th century, when Leonhard Euler and extended analytical methods to functions depending on multiple variables, particularly in their foundational work on the .

Domain and Codomain

In the context of functions of several real variables, the f:DRmf: D \to \mathbb{R}^m, where DRnD \subseteq \mathbb{R}^n, is the set DD from which the inputs are drawn, typically chosen as a nonempty of Rn\mathbb{R}^n to ensure the function is well-defined. Often, DD is taken to be an open of Rn\mathbb{R}^n to facilitate such as differentiability, though closed or other subsets may be used depending on the application. Common examples include open balls B(a,r)={xRnxa<r}B(a, r) = \{ x \in \mathbb{R}^n \mid \|x - a\| < r \}, which are bounded and connected regions centered at a point aa with radius r>0r > 0, and rectangular domains such as products of open intervals, like (a1,b1)××(an,bn)(a_1, b_1) \times \cdots \times (a_n, b_n), which provide simple Cartesian structures for computation. More generally, domains can be open sets in Rn\mathbb{R}^n that form manifolds without boundary, allowing for smooth extensions of single-variable concepts to higher dimensions. The of ff is specified as Rm\mathbb{R}^m, the target space in which the outputs reside, though in some contexts it may extend to Cm\mathbb{C}^m if complex values are considered, but for real-variable functions, Rm\mathbb{R}^m is standard. The actual outputs form the (or range) f(D)Rmf(D) \subseteq \mathbb{R}^m, which is the of the codomain attained by applying ff to elements of DD, and this image may be proper if ff is not surjective. This distinction ensures that the codomain provides an upper bound on possible values, while the image captures the function's effective reach, as seen in examples like the projection functions where the image fills the entire codomain. Domains for functions of several real variables often exhibit specific topological and measure-theoretic properties that influence their suitability for further study, such as integration or optimization. Connectedness means DD cannot be partitioned into disjoint nonempty open subsets, ensuring a single "piece" for global behavior analysis, and is commonly assumed for open domains like or rectangles. Boundedness requires DD to fit within some of finite radius, which aids in arguments when combined with closure, as in closed that are compact in Rn\mathbb{R}^n. For readiness in Riemann integration, domains are frequently required to be Jordan measurable, meaning their boundary has Jordan measure zero; bounded open sets with piecewise smooth boundaries, such as rectangles or , satisfy this property. The graph of ff is the set Γf={(x,y)Rn×Rmy=f(x),xD}\Gamma_f = \{ (x, y) \in \mathbb{R}^n \times \mathbb{R}^m \mid y = f(x), \, x \in D \}, a subset of Rn+m\mathbb{R}^{n+m} that embeds the function's behavior, and for sufficiently regular ff, it forms a hypersurface of dimension nn in this higher-dimensional space. This contrasts with the domain DD, which is solely the input space, as the graph incorporates both inputs and outputs to visualize the mapping relation introduced formally elsewhere. Level sets, for scalar-valued functions (m=1m=1), are the subsets Lc={xDf(x)=c}L_c = \{ x \in D \mid f(x) = c \} for constants cRc \in \mathbb{R}, which partition the domain into regions of constant output and typically form hypersurfaces of dimension n1n-1 within DD, distinct from both the domain and graph by focusing on preimages rather than the full mapping or product structure.

Notation and Graphical Representation

Functions of several real variables are typically denoted using subscripted variables for scalar-valued functions, such as f(x1,x2,,xn)f(x_1, x_2, \dots, x_n) where x=(x1,x2,,xn)Rn\mathbf{x} = (x_1, x_2, \dots, x_n) \in \mathbb{R}^n and f:RnRf: \mathbb{R}^n \to \mathbb{R}. This notation emphasizes the independent variables explicitly. For vector-valued functions f:RnRm\mathbf{f}: \mathbb{R}^n \to \mathbb{R}^m, boldface is often used, as in f(x)\mathbf{f}(\mathbf{x}), to distinguish the output as a vector. Coordinate-free forms, such as f(x)f(\mathbf{x}) without explicit components, are also common in more abstract contexts to highlight structure. Graphical representation aids in understanding these functions, particularly for low dimensions. For scalar functions f:R2Rf: \mathbb{R}^2 \to \mathbb{R}, the graph is a surface in R3\mathbb{R}^3, plotted as z=f(x,y)z = f(x, y) to visualize height variations over the domain. Contour plots, or level curves, depict sets where f(x,y)=kf(x, y) = k for constant kk, providing a 2D projection that reveals gradients and critical points without full 3D rendering. For vector-valued functions f:RnRn\mathbf{f}: \mathbb{R}^n \to \mathbb{R}^n, such as vector fields, visualization uses arrow plots where each arrow at point x\mathbf{x} represents f(x)\mathbf{f}(\mathbf{x}), illustrating direction and magnitude; streamlines may trace integral curves for flow interpretation. Visualizing functions for n>3n > 3 faces inherent limitations due to human perception confined to three spatial dimensions, making direct graphs impossible. Common techniques include slicing, where some variables are fixed to reduce dimensionality (e.g., traces by setting x3=cx_3 = c), or projections onto lower-dimensional subspaces to approximate behavior. Level sets generalize contours to higher dimensions as hypersurfaces where f(x)=kf(\mathbf{x}) = k. Software tools facilitate these representations; for instance, MATLAB supports surface and quiver plots for 3D surfaces and vector fields, while Python's Matplotlib library offers similar capabilities for contour and arrow visualizations without requiring custom code for basic rendering.

Basic Properties

Continuity and Limits

In the context of functions from Rn\mathbb{R}^n to Rm\mathbb{R}^m, the limit of a function f:DRnRmf: D \subseteq \mathbb{R}^n \to \mathbb{R}^m as xa\mathbf{x} \to \mathbf{a} (where aRn\mathbf{a} \in \mathbb{R}^n and a\mathbf{a} may or may not be in DD) is defined using the ϵ\epsilon-δ\delta criterion. Specifically, limxaf(x)=L\lim_{\mathbf{x} \to \mathbf{a}} f(\mathbf{x}) = L if for every ϵ>0\epsilon > 0, there exists δ>0\delta > 0 such that if 0<xa<δ0 < \|\mathbf{x} - \mathbf{a}\| < \delta, then f(x)L<ϵ\|f(\mathbf{x}) - L\| < \epsilon, where \|\cdot\| denotes the Euclidean norm. This definition generalizes the single-variable case by considering neighborhoods in Rn\mathbb{R}^n as open balls centered at a\mathbf{a}, excluding a\mathbf{a} itself to focus on approaching behavior. A function ff is continuous at aD\mathbf{a} \in D if limxaf(x)=f(a)\lim_{\mathbf{x} \to \mathbf{a}} f(\mathbf{x}) = f(\mathbf{a}), which, by the ϵ\epsilon-δ\delta definition, means for every ϵ>0\epsilon > 0, there exists δ>0\delta > 0 such that if xa<δ\|\mathbf{x} - \mathbf{a}\| < \delta and xD\mathbf{x} \in D, then f(x)f(a)<ϵ\|f(\mathbf{x}) - f(\mathbf{a})\| < \epsilon. Continuity is pointwise, holding at individual points, but uniform continuity strengthens this to apply across the entire domain: for every ϵ>0\epsilon > 0, there exists δ>0\delta > 0 such that for all x,yD\mathbf{x}, \mathbf{y} \in D with xy<δ\|\mathbf{x} - \mathbf{y}\| < \delta, f(x)f(y)<ϵ\|f(\mathbf{x}) - f(\mathbf{y})\| < \epsilon, independent of the specific points. Pointwise continuity does not imply uniform continuity on unbounded domains, though continuous functions on compact subsets of Rn\mathbb{R}^n are uniformly continuous. Unlike single-variable limits, multivariable limits can depend on the path taken to approach a\mathbf{a}, complicating existence. For instance, consider f(x,y)=xyx2+y2f(x,y) = \frac{xy}{x^2 + y^2} as (x,y)(0,0)(x,y) \to (0,0). Along the x-axis (y=0y=0), the limit is 0; along the y-axis (x=0x=0), it is also 0; but along y=xy=x, it is 12\frac{1}{2}. Since different paths yield different values, the limit does not exist. Such path dependence arises because Rn\mathbb{R}^n for n2n \geq 2 allows infinitely many approach directions, requiring consistency across all for the limit to exist. An equivalent sequential characterization states that limxaf(x)=L\lim_{\mathbf{x} \to \mathbf{a}} f(\mathbf{x}) = L if and only if for every sequence {xk}k=1\{\mathbf{x}_k\}_{k=1}^\infty in D{a}D \setminus \{\mathbf{a}\} with xka\mathbf{x}_k \to \mathbf{a}, we have f(xk)Lf(\mathbf{x}_k) \to L. This is useful for proving non-existence: if two sequences approaching a\mathbf{a} give subsequences of ff converging to different limits, the overall limit fails. Limits are unique when they exist, and for vector-valued functions, the limit holds if and only if it holds componentwise.

Symmetry Properties

In the context of functions f:RnRf: \mathbb{R}^n \to \mathbb{R}, symmetry properties generalize concepts from single-variable calculus to higher dimensions, capturing invariances under geometric transformations such as reflections and translations. These properties are fundamental in analysis, aiding in simplification of integrals, Fourier representations, and understanding function behavior over symmetric domains. An even function satisfies f(x)=f(x)f(-x) = f(x) for all xRnx \in \mathbb{R}^n, where x=(x1,,xn)-x = (-x_1, \dots, -x_n) denotes componentwise negation; this condition implies symmetry under reflection through the origin, extending the one-dimensional notion to invariance across the origin in all directions or, more generally, reflections over coordinate hyperplanes. For example, the function f(x)=x2f(x) = \|x\|^2 is even, as f(x)=i=1n(xi)2=i=1nxi2=f(x)f(-x) = \sum_{i=1}^n (-x_i)^2 = \sum_{i=1}^n x_i^2 = f(x), reflecting rotational symmetry combined with evenness. This property preserves under addition and multiplication of even functions, facilitating decompositions in harmonic analysis. An odd function, in contrast, obeys f(x)=f(x)f(-x) = -f(x) for all xRnx \in \mathbb{R}^n, corresponding to antisymmetry with respect to the origin, such that the graph is invariant under 180-degree rotation about the origin. A key implication is that f(0)=0f(0) = 0 if the origin is in the domain, assuming continuity at the origin; for instance, f(x)=x1x2xnf(x) = x_1 x_2 \cdots x_n is odd in Rn\mathbb{R}^n, as negating all components yields the negative value. Products of odd and even functions yield odd functions, and sums of odd functions remain odd, which is useful for parity arguments in integration over symmetric regions. Radial symmetry arises when a function depends solely on the Euclidean norm x=i=1nxi2\|x\| = \sqrt{\sum_{i=1}^n x_i^2}
Add your contribution
Related Hubs
User Avatar
No comments yet.