Hubbry Logo
Vector spaceVector spaceMain
Open search
Vector space
Community hub
Vector space
logo
8 pages, 0 posts
0 subscribers
Be the first to start a discussion here.
Be the first to start a discussion here.
Contribute something
Vector space
Vector space
from Wikipedia
Vector addition and scalar multiplication: a vector v (blue) is added to another vector w (red, upper illustration). Below, w is stretched by a factor of 2, yielding the sum v + 2w.

In mathematics and physics, a vector space (also called a linear space) is a set whose elements, often called vectors, can be added together and multiplied ("scaled") by numbers called scalars. The operations of vector addition and scalar multiplication must satisfy certain requirements, called vector axioms. Real vector spaces and complex vector spaces are kinds of vector spaces based on different kinds of scalars: real numbers and complex numbers. Scalars can also be, more generally, elements of any field.

Vector spaces generalize Euclidean vectors, which allow modeling of physical quantities (such as forces and velocity) that have not only a magnitude, but also a direction. The concept of vector spaces is fundamental for linear algebra, together with the concept of matrices, which allows computing in vector spaces. This provides a concise and synthetic way for manipulating and studying systems of linear equations.

Vector spaces are characterized by their dimension, which, roughly speaking, specifies the number of independent directions in the space. This means that, for two vector spaces over a given field and with the same dimension, the properties that depend only on the vector-space structure are exactly the same (technically the vector spaces are isomorphic). A vector space is finite-dimensional if its dimension is a natural number. Otherwise, it is infinite-dimensional, and its dimension is an infinite cardinal. Finite-dimensional vector spaces occur naturally in geometry and related areas. Infinite-dimensional vector spaces occur in many areas of mathematics. For example, polynomial rings are countably infinite-dimensional vector spaces, and many function spaces have the cardinality of the continuum as a dimension.

Many vector spaces that are considered in mathematics are also endowed with other structures. This is the case of algebras, which include field extensions, polynomial rings, associative algebras and Lie algebras. This is also the case of topological vector spaces, which include function spaces, inner product spaces, normed spaces, Hilbert spaces and Banach spaces.

Definition and basic properties

[edit]

In this article, vectors are represented in boldface to distinguish them from scalars.[nb 1][1]

A vector space over a field F is a non-empty set V together with a binary operation and a binary function that satisfy the eight axioms listed below. In this context, the elements of V are commonly called vectors, and the elements of F are called scalars.[2]

  • The binary operation, called vector addition or simply addition assigns to any two vectors v and w in V. a third vector in V which is commonly written as v + w, and called the sum of these two vectors.
  • The binary function, called scalar multiplication, assigns to any scalar a in F and any vector v in V another vector in V, which is denoted av.[nb 2]

To have a vector space, the eight following axioms must be satisfied for every u, v and w in V, and a and b in F.[3]

Axiom Statement
Associativity of vector addition u + (v + w) = (u + v) + w
Commutativity of vector addition u + v = v + u
Identity element of vector addition There exists an element 0V, called the zero vector, such that v + 0 = v for all vV.
Inverse elements of vector addition For every vV, there exists an element vV, called the additive inverse of v, such that v + (−v) = 0.
Compatibility of scalar multiplication with field multiplication a(bv) = (ab)v [nb 3]
Identity element of scalar multiplication 1v = v, where 1 denotes the multiplicative identity in F.
Distributivity of scalar multiplication with respect to vector addition   a(u + v) = au + av
Distributivity of scalar multiplication with respect to field addition (a + b)v = av + bv

When the scalar field is the real numbers, the vector space is called a real vector space, and when the scalar field is the complex numbers, the vector space is called a complex vector space.[4] These two cases are the most common ones, but vector spaces with scalars in an arbitrary field F are also commonly considered. Such a vector space is called an F-vector space or a vector space over F.[5]

An equivalent definition of a vector space can be given, which is much more concise but less elementary: the first four axioms (related to vector addition) say that a vector space is an abelian group under addition, and the four remaining axioms (related to the scalar multiplication) say that this operation defines a ring homomorphism from the field F into the endomorphism ring of this group.[6] Specifically, the distributivity of scalar multiplication with respect to vector addition means that multiplication by a scalar a is an endomorphism of the group. The remaining three axiom establish that the function that maps a scalar a to the multiplication by a is a ring homomorphism from the field to the endomorphism ring of the group.

Subtraction of two vectors can be defined as

Direct consequences of the axioms include that, for every and one has

  • implies or

Even more concisely, a vector space is a module over a field.[7]

Bases, vector coordinates, and subspaces

[edit]
A vector v in R2 (blue) expressed in terms of different bases: using the standard basis of R2: v = xe1 + ye2 (black), and using a different, non-orthogonal basis: v = f1 + f2 (red).
Linear combination
Given a set G of elements of a F-vector space V, a linear combination of elements of G is an element of V of the form where and The scalars are called the coefficients of the linear combination.[8]
Linear independence
The elements of a subset G of a F-vector space V are said to be linearly independent if no element of G can be written as a linear combination of the other elements of G. Equivalently, they are linearly independent if two linear combinations of elements of G define the same element of V if and only if they have the same coefficients. Also equivalently, they are linearly independent if a linear combination results in the zero vector if and only if all its coefficients are zero.[9]
Linear subspace
A linear subspace or vector subspace W of a vector space V is a non-empty subset of V that is closed under vector addition and scalar multiplication; that is, the sum of two elements of W and the product of an element of W by a scalar belong to W.[10] This implies that every linear combination of elements of W belongs to W. A linear subspace is a vector space for the induced addition and scalar multiplication; this means that the closure property implies that the axioms of a vector space are satisfied.[11]
The closure property also implies that every intersection of linear subspaces is a linear subspace.[11]
Linear span
Given a subset G of a vector space V, the linear span or simply the span of G is the smallest linear subspace of V that contains G, in the sense that it is the intersection of all linear subspaces that contain G. The span of G is also the set of all linear combinations of elements of G.
If W is the span of G, one says that G spans or generates W, and that G is a spanning set or a generating set of W.[12]
Basis and dimension
A subset of a vector space is a basis if its elements are linearly independent and span the vector space.[13] Every vector space has at least one basis, or many in general (see Basis (linear algebra) § Proof that every vector space has a basis).[14] Moreover, all bases of a vector space have the same cardinality, which is called the dimension of the vector space (see Dimension theorem for vector spaces).[15] This is a fundamental property of vector spaces, which is detailed in the remainder of the section.

Bases are a fundamental tool for the study of vector spaces, especially when the dimension is finite. In the infinite-dimensional case, the existence of infinite bases, often called Hamel bases, depends on the axiom of choice. It follows that, in general, no base can be explicitly described.[16] For example, the real numbers form an infinite-dimensional vector space over the rational numbers, for which no specific basis is known.

Consider a basis of a vector space V of dimension n over a field F. The definition of a basis implies that every may be written with in F, and that this decomposition is unique. The scalars are called the coordinates of v on the basis. They are also said to be the coefficients of the decomposition of v on the basis. One also says that the n-tuple of the coordinates is the coordinate vector of v on the basis, since the set of the n-tuples of elements of F is a vector space for componentwise addition and scalar multiplication, whose dimension is n.

The one-to-one correspondence between vectors and their coordinate vectors maps vector addition to vector addition and scalar multiplication to scalar multiplication. It is thus a vector space isomorphism, which allows translating reasonings and computations on vectors into reasonings and computations on their coordinates.[17]

History

[edit]

Vector spaces stem from affine geometry, via the introduction of coordinates in the plane or three-dimensional space. Around 1636, French mathematicians René Descartes and Pierre de Fermat founded analytic geometry by identifying solutions to an equation of two variables with points on a plane curve.[18] To achieve geometric solutions without using coordinates, Bolzano introduced, in 1804, certain operations on points, lines, and planes, which are predecessors of vectors.[19] Möbius (1827) introduced the notion of barycentric coordinates.[20] Bellavitis (1833) introduced an equivalence relation on directed line segments that share the same length and direction which he called equipollence.[21] A Euclidean vector is then an equivalence class of that relation.[22]

Vectors were reconsidered with the presentation of complex numbers by Argand and Hamilton and the inception of quaternions by the latter.[23] They are elements in R2 and R4; treating them using linear combinations goes back to Laguerre in 1867, who also defined systems of linear equations.

In 1857, Cayley introduced the matrix notation which allows for harmonization and simplification of linear maps. Around the same time, Grassmann studied the barycentric calculus initiated by Möbius. He envisaged sets of abstract objects endowed with operations.[24] In his work, the concepts of linear independence and dimension, as well as scalar products are present. Grassmann's 1844 work exceeds the framework of vector spaces as well since his considering multiplication led him to what are today called algebras. Italian mathematician Peano was the first to give the modern definition of vector spaces and linear maps in 1888,[25] although he called them "linear systems".[26] Peano's axiomatization allowed for vector spaces with infinite dimension, but Peano did not develop that theory further. In 1897, Salvatore Pincherle adopted Peano's axioms and made initial inroads into the theory of infinite-dimensional vector spaces.[27]

An important development of vector spaces is due to the construction of function spaces by Henri Lebesgue. This was later formalized by Banach and Hilbert, around 1920.[28] At that time, algebra and the new field of functional analysis began to interact, notably with key concepts such as spaces of p-integrable functions and Hilbert spaces.[29]

Examples

[edit]

Arrows in the plane

[edit]
Vector addition: the sum v + w (black) of the vectors v (blue) and w (red) is shown.
Scalar multiplication: the multiples v and 2w are shown.

The first example of a vector space consists of arrows in a fixed plane, starting at one fixed point. This is used in physics to describe forces or velocities.[30] Given any two such arrows, v and w, the parallelogram spanned by these two arrows contains one diagonal arrow that starts at the origin, too. This new arrow is called the sum of the two arrows, and is denoted v + w. In the special case of two arrows on the same line, their sum is the arrow on this line whose length is the sum or the difference of the lengths, depending on whether the arrows have the same direction. Another operation that can be done with arrows is scaling: given any positive real number a, the arrow that has the same direction as v, but is dilated or shrunk by multiplying its length by a, is called multiplication of v by a. It is denoted av. When a is negative, av is defined as the arrow pointing in the opposite direction instead.[31]

The following shows a few examples: if a = 2, the resulting vector aw has the same direction as w, but is stretched to the double length of w (the second image). Equivalently, 2w is the sum w + w. Moreover, (−1)v = −v has the opposite direction and the same length as v (blue vector pointing down in the second image).

Ordered pairs of numbers

[edit]

A second key example of a vector space is provided by pairs of real numbers x and y. The order of the components x and y is significant, so such a pair is also called an ordered pair. Such a pair is written as (x, y). The sum of two such pairs and the multiplication of a pair with a number is defined as follows:[32]

The first example above reduces to this example if an arrow is represented by a pair of Cartesian coordinates of its endpoint.

Coordinate space

[edit]

The simplest example of a vector space over a field F is the field F itself with its addition viewed as vector addition and its multiplication viewed as scalar multiplication. More generally, all n-tuples (sequences of length n) of elements ai of F form a vector space that is usually denoted Fn and called a coordinate space.[33] The case n = 1 is the above-mentioned simplest example, in which the field F is also regarded as a vector space over itself. The case F = R and n = 2 (so R2) reduces to the previous example.

Complex numbers and other field extensions

[edit]

The set of complex numbers C, numbers that can be written in the form x + iy for real numbers x and y where i is the imaginary unit, form a vector space over the reals with the usual addition and multiplication: (x + iy) + (a + ib) = (x + a) + i(y + b) and c ⋅ (x + iy) = (cx) + i(cy) for real numbers x, y, a, b and c. The various axioms of a vector space follow from the fact that the same rules hold for complex number arithmetic. The example of complex numbers is essentially the same as (that is, it is isomorphic to) the vector space of ordered pairs of real numbers mentioned above: if we think of the complex number x + i y as representing the ordered pair (x, y) in the complex plane then we see that the rules for addition and scalar multiplication correspond exactly to those in the earlier example.

More generally, field extensions provide another class of examples of vector spaces, particularly in algebra and algebraic number theory: a field F containing a smaller field E is an E-vector space, by the given multiplication and addition operations of F.[34] For example, the complex numbers are a vector space over R, and the field extension is a vector space over Q.

Function spaces

[edit]
Addition of functions: the sum of the sine and the exponential function is with .

Functions from any fixed set Ω to a field F also form vector spaces, by performing addition and scalar multiplication pointwise. That is, the sum of two functions f and g is the function given by and similarly for multiplication. Such function spaces occur in many geometric situations, when Ω is the real line or an interval, or other subsets of R. Many notions in topology and analysis, such as continuity, integrability or differentiability are well-behaved with respect to linearity: sums and scalar multiples of functions possessing such a property still have that property.[35] Therefore, the set of such functions are vector spaces, whose study belongs to functional analysis.

Linear equations

[edit]

Systems of homogeneous linear equations are closely tied to vector spaces.[36] For example, the solutions of are given by triples with arbitrary and They form a vector space: sums and scalar multiples of such triples still satisfy the same ratios of the three variables; thus they are solutions, too. Matrices can be used to condense multiple linear equations as above into one vector equation, namely

where is the matrix containing the coefficients of the given equations, is the vector denotes the matrix product, and is the zero vector. In a similar vein, the solutions of homogeneous linear differential equations form vector spaces. For example,

yields where and are arbitrary constants, and is the natural exponential function.

Linear maps and matrices

[edit]

The relation of two vector spaces can be expressed by linear map or linear transformation. They are functions that reflect the vector space structure, that is, they preserve sums and scalar multiplication: for all and in all in [37]

An isomorphism is a linear map f : VW such that there exists an inverse map g : WV, which is a map such that the two possible compositions fg : WW and gf : VV are identity maps. Equivalently, f is both one-to-one (injective) and onto (surjective).[38] If there exists an isomorphism between V and W, the two spaces are said to be isomorphic; they are then essentially identical as vector spaces, since all identities holding in V are, via f, transported to similar ones in W, and vice versa via g.

Describing an arrow vector v by its coordinates x and y yields an isomorphism of vector spaces.

For example, the arrows in the plane and the ordered pairs of numbers vector spaces in the introduction above (see § Examples) are isomorphic: a planar arrow v departing at the origin of some (fixed) coordinate system can be expressed as an ordered pair by considering the x- and y-component of the arrow, as shown in the image at the right. Conversely, given a pair (x, y), the arrow going by x to the right (or to the left, if x is negative), and y up (down, if y is negative) turns back the arrow v.[39]

Linear maps VW between two vector spaces form a vector space HomF(V, W), also denoted L(V, W), or 𝓛(V, W).[40] The space of linear maps from V to F is called the dual vector space, denoted V.[41] Via the injective natural map VV∗∗, any vector space can be embedded into its bidual; the map is an isomorphism if and only if the space is finite-dimensional.[42]

Once a basis of V is chosen, linear maps f : VW are completely determined by specifying the images of the basis vectors, because any element of V is expressed uniquely as a linear combination of them.[43] If dim V = dim W, a 1-to-1 correspondence between fixed bases of V and W gives rise to a linear map that maps any basis element of V to the corresponding basis element of W. It is an isomorphism, by its very definition.[44] Therefore, two vector spaces over a given field are isomorphic if their dimensions agree and vice versa. Another way to express this is that any vector space over a given field is completely classified (up to isomorphism) by its dimension, a single number. In particular, any n-dimensional F-vector space V is isomorphic to Fn. However, there is no "canonical" or preferred isomorphism; an isomorphism φ : FnV is equivalent to the choice of a basis of V, by mapping the standard basis of Fn to V, via φ.

Matrices

[edit]
A typical matrix

Matrices are a useful notion to encode linear maps.[45] They are written as a rectangular array of scalars as in the image at the right. Any m-by-n matrix gives rise to a linear map from Fn to Fm, by the following where denotes summation, or by using the matrix multiplication of the matrix with the coordinate vector :

Moreover, after choosing bases of V and W, any linear map f : VW is uniquely represented by a matrix via this assignment.[46]

The volume of this parallelepiped is the absolute value of the determinant of the 3-by-3 matrix formed by the vectors r1, r2, and r3.

The determinant det (A) of a square matrix A is a scalar that tells whether the associated map is an isomorphism or not: to be so it is sufficient and necessary that the determinant is nonzero.[47] The linear transformation of Rn corresponding to a real n-by-n matrix is orientation preserving if and only if its determinant is positive.

Eigenvalues and eigenvectors

[edit]

Endomorphisms, linear maps f : VV, are particularly important since in this case vectors v can be compared with their image under f, f(v). Any nonzero vector v satisfying λv = f(v), where λ is a scalar, is called an eigenvector of f with eigenvalue λ.[48] Equivalently, v is an element of the kernel of the difference fλ · Id (where Id is the identity map VV). If V is finite-dimensional, this can be rephrased using determinants: f having eigenvalue λ is equivalent to By spelling out the definition of the determinant, the expression on the left hand side can be seen to be a polynomial function in λ, called the characteristic polynomial of f.[49] If the field F is large enough to contain a zero of this polynomial (which automatically happens for F algebraically closed, such as F = C) any linear map has at least one eigenvector. The vector space V may or may not possess an eigenbasis, a basis consisting of eigenvectors. This phenomenon is governed by the Jordan canonical form of the map.[50] The set of all eigenvectors corresponding to a particular eigenvalue of f forms a vector space known as the eigenspace corresponding to the eigenvalue (and f) in question.

Basic constructions

[edit]

In addition to the above concrete examples, there are a number of standard linear algebraic constructions that yield vector spaces related to given ones.

Subspaces and quotient spaces

[edit]
A line passing through the origin (blue, thick) in R3 is a linear subspace. It is the intersection of two planes (green and yellow).

A nonempty subset of a vector space that is closed under addition and scalar multiplication (and therefore contains the -vector of ) is called a linear subspace of , or simply a subspace of , when the ambient space is unambiguously a vector space.[51][nb 4] Subspaces of are vector spaces (over the same field) in their own right. The intersection of all subspaces containing a given set of vectors is called its span, and it is the smallest subspace of containing the set . Expressed in terms of elements, the span is the subspace consisting of all the linear combinations of elements of .[52]

Linear subspace of dimension 1 and 2 are referred to as a line (also vector line), and a plane respectively. If W is an n-dimensional vector space, any subspace of dimension 1 less, i.e., of dimension is called a hyperplane.[53]

The counterpart to subspaces are quotient vector spaces.[54] Given any subspace , the quotient space (" modulo ") is defined as follows: as a set, it consists of where is an arbitrary vector in . The sum of two such elements and is , and scalar multiplication is given by . The key point in this definition is that if and only if the difference of and lies in .[nb 5] This way, the quotient space "forgets" information that is contained in the subspace .

The kernel of a linear map consists of vectors that are mapped to in .[55] The kernel and the image are subspaces of and , respectively.[56]

An important example is the kernel of a linear map for some fixed matrix . The kernel of this map is the subspace of vectors such that , which is precisely the set of solutions to the system of homogeneous linear equations belonging to . This concept also extends to linear differential equations where the coefficients are functions in too. In the corresponding map the derivatives of the function appear linearly (as opposed to , for example). Since differentiation is a linear procedure (that is, and for a constant ) this assignment is linear, called a linear differential operator. In particular, the solutions to the differential equation form a vector space (over R or C).[57]

The existence of kernels and images is part of the statement that the category of vector spaces (over a fixed field ) is an abelian category, that is, a corpus of mathematical objects and structure-preserving maps between them (a category) that behaves much like the category of abelian groups.[58] Because of this, many statements such as the first isomorphism theorem (also called rank–nullity theorem in matrix-related terms) and the second and third isomorphism theorem can be formulated and proven in a way very similar to the corresponding statements for groups.

Direct product and direct sum

[edit]

The direct product of vector spaces and the direct sum of vector spaces are two ways of combining an indexed family of vector spaces into a new vector space.

The direct product of a family of vector spaces consists of the set of all tuples , which specify for each index in some index set an element of .[59] Addition and scalar multiplication is performed componentwise. A variant of this construction is the direct sum (also called coproduct and denoted ), where only tuples with finitely many nonzero vectors are allowed. If the index set is finite, the two constructions agree, but in general they are different.

Tensor product

[edit]

The tensor product or simply of two vector spaces and is one of the central notions of multilinear algebra which deals with extending notions such as linear maps to several variables. A map from the Cartesian product is called bilinear if is linear in both variables and That is to say, for fixed the map is linear in the sense above and likewise for fixed

Commutative diagram depicting the universal property of the tensor product

The tensor product is a particular vector space that is a universal recipient of bilinear maps as follows. It is defined as the vector space consisting of finite (formal) sums of symbols called tensors subject to the rules[60] These rules ensure that the map from the to that maps a tuple to is bilinear. The universality states that given any vector space and any bilinear map there exists a unique map shown in the diagram with a dotted arrow, whose composition with equals [61] This is called the universal property of the tensor product, an instance of the method—much used in advanced abstract algebra—to indirectly define objects by specifying maps from or to this object.

Vector spaces with additional structure

[edit]

From the point of view of linear algebra, vector spaces are completely understood insofar as any vector space over a given field is characterized, up to isomorphism, by its dimension. However, vector spaces per se do not offer a framework to deal with the question—crucial to analysis—whether a sequence of functions converges to another function. Likewise, linear algebra is not adapted to deal with infinite series, since the addition operation allows only finitely many terms to be added. Therefore, the needs of functional analysis require considering additional structures.[62]

A vector space may be given a partial order under which some vectors can be compared.[63] For example, -dimensional real space can be ordered by comparing its vectors componentwise. Ordered vector spaces, for example Riesz spaces, are fundamental to Lebesgue integration, which relies on the ability to express a function as a difference of two positive functions where denotes the positive part of and the negative part.[64]

Normed vector spaces and inner product spaces

[edit]

"Measuring" vectors is done by specifying a norm, a datum which measures lengths of vectors, or by an inner product, which measures angles between vectors. Norms and inner products are denoted and respectively. The datum of an inner product entails that lengths of vectors can be defined too, by defining the associated norm Vector spaces endowed with such data are known as normed vector spaces and inner product spaces, respectively.[65]

Coordinate space can be equipped with the standard dot product: In this reflects the common notion of the angle between two vectors and by the law of cosines: Because of this, two vectors satisfying are called orthogonal. An important variant of the standard dot product is used in Minkowski space: endowed with the Lorentz product[66] In contrast to the standard dot product, it is not positive definite: also takes negative values, for example, for Singling out the fourth coordinate—corresponding to time, as opposed to three space-dimensions—makes it useful for the mathematical treatment of special relativity. Note that in other conventions time is often written as the first, or "zeroeth" component so that the Lorentz product is written

Topological vector spaces

[edit]

Convergence questions are treated by considering vector spaces carrying a compatible topology, a structure that allows one to talk about elements being close to each other.[67] Compatible here means that addition and scalar multiplication have to be continuous maps. Roughly, if and in , and in vary by a bounded amount, then so do and [nb 6] To make sense of specifying the amount a scalar changes, the field also has to carry a topology in this context; a common choice is the reals or the complex numbers.

In such topological vector spaces one can consider series of vectors. The infinite sum denotes the limit of the corresponding finite partial sums of the sequence of elements of For example, the could be (real or complex) functions belonging to some function space in which case the series is a function series. The mode of convergence of the series depends on the topology imposed on the function space. In such cases, pointwise convergence and uniform convergence are two prominent examples.[68]

Unit "spheres" in consist of plane vectors of norm 1. Depicted are the unit spheres in different -norms, for and The bigger diamond depicts points of 1-norm equal to 2.

A way to ensure the existence of limits of certain infinite series is to restrict attention to spaces where any Cauchy sequence has a limit; such a vector space is called complete. Roughly, a vector space is complete provided that it contains all necessary limits. For example, the vector space of polynomials on the unit interval equipped with the topology of uniform convergence is not complete because any continuous function on can be uniformly approximated by a sequence of polynomials, by the Weierstrass approximation theorem.[69] In contrast, the space of all continuous functions on with the same topology is complete.[70] A norm gives rise to a topology by defining that a sequence of vectors converges to if and only if Banach and Hilbert spaces are complete topological vector spaces whose topologies are given, respectively, by a norm and an inner product. Their study—a key piece of functional analysis—focuses on infinite-dimensional vector spaces, since all norms on finite-dimensional topological vector spaces give rise to the same notion of convergence.[71] The image at the right shows the equivalence of the -norm and -norm on as the unit "balls" enclose each other, a sequence converges to zero in one norm if and only if it so does in the other norm. In the infinite-dimensional case, however, there will generally be inequivalent topologies, which makes the study of topological vector spaces richer than that of vector spaces without additional data.

From a conceptual point of view, all notions related to topological vector spaces should match the topology. For example, instead of considering all linear maps (also called functionals) maps between topological vector spaces are required to be continuous.[72] In particular, the (topological) dual space consists of continuous functionals (or to ). The fundamental Hahn–Banach theorem is concerned with separating subspaces of appropriate topological vector spaces by continuous functionals.[73]

Banach spaces

[edit]

Banach spaces, introduced by Stefan Banach, are complete normed vector spaces.[74]

A first example is the vector space consisting of infinite vectors with real entries whose -norm given by

The topologies on the infinite-dimensional space are inequivalent for different For example, the sequence of vectors in which the first components are and the following ones are converges to the zero vector for but does not for but

More generally than sequences of real numbers, functions are endowed with a norm that replaces the above sum by the Lebesgue integral

The space of integrable functions on a given domain (for example an interval) satisfying and equipped with this norm are called Lebesgue spaces, denoted [nb 7]

These spaces are complete.[75] (If one uses the Riemann integral instead, the space is not complete, which may be seen as a justification for Lebesgue's integration theory.[nb 8]) Concretely this means that for any sequence of Lebesgue-integrable functions with satisfying the condition there exists a function belonging to the vector space such that

Imposing boundedness conditions not only on the function, but also on its derivatives leads to Sobolev spaces.[76]

Hilbert spaces

[edit]
The succeeding snapshots show summation of 1 to 5 terms in approximating a periodic function (blue) by finite sum of sine functions (red).

Complete inner product spaces are known as Hilbert spaces, in honor of David Hilbert.[77] The Hilbert space with inner product given by where denotes the complex conjugate of [78][nb 9] is a key case.

By definition, in a Hilbert space, any Cauchy sequence converges to a limit. Conversely, finding a sequence of functions with desirable properties that approximate a given limit function is equally crucial. Early analysis, in the guise of the Taylor approximation, established an approximation of differentiable functions by polynomials.[79] By the Stone–Weierstrass theorem, every continuous function on can be approximated as closely as desired by a polynomial.[80] A similar approximation technique by trigonometric functions is commonly called Fourier expansion, and is much applied in engineering. More generally, and more conceptually, the theorem yields a simple description of what "basic functions", or, in abstract Hilbert spaces, what basic vectors suffice to generate a Hilbert space in the sense that the closure of their span (that is, finite linear combinations and limits of those) is the whole space. Such a set of functions is called a basis of its cardinality is known as the Hilbert space dimension.[nb 10] Not only does the theorem exhibit suitable basis functions as sufficient for approximation purposes, but also together with the Gram–Schmidt process, it enables one to construct a basis of orthogonal vectors.[81] Such orthogonal bases are the Hilbert space generalization of the coordinate axes in finite-dimensional Euclidean space.

The solutions to various differential equations can be interpreted in terms of Hilbert spaces. For example, a great many fields in physics and engineering lead to such equations, and frequently solutions with particular physical properties are used as basis functions, often orthogonal.[82] As an example from physics, the time-dependent Schrödinger equation in quantum mechanics describes the change of physical properties in time by means of a partial differential equation, whose solutions are called wavefunctions.[83] Definite values for physical properties such as energy, or momentum, correspond to eigenvalues of a certain (linear) differential operator and the associated wavefunctions are called eigenstates. The spectral theorem decomposes a linear compact operator acting on functions in terms of these eigenfunctions and their eigenvalues.[84]

Algebras over fields

[edit]
A hyperbola, given by the equation The coordinate ring of functions on this hyperbola is given by an infinite-dimensional vector space over

General vector spaces do not possess a multiplication between vectors. A vector space equipped with an additional bilinear operator defining the multiplication of two vectors is an algebra over a field (or F-algebra if the field F is specified).[85]

For example, the set of all polynomials forms an algebra known as the polynomial ring: using that the sum of two polynomials is a polynomial, they form a vector space; they form an algebra since the product of two polynomials is again a polynomial. Rings of polynomials (in several variables) and their quotients form the basis of algebraic geometry, because they are rings of functions of algebraic geometric objects.[86]

Another crucial example are Lie algebras, which are neither commutative nor associative, but the failure to be so is limited by the constraints ( denotes the product of and ):

  • (anticommutativity), and
  • (Jacobi identity).[87]

Examples include the vector space of -by- matrices, with the commutator of two matrices, and endowed with the cross product.

The tensor algebra is a formal way of adding products to any vector space to obtain an algebra.[88] As a vector space, it is spanned by symbols, called simple tensors where the degree varies. The multiplication is given by concatenating such symbols, imposing the distributive law under addition, and requiring that scalar multiplication commute with the tensor product ⊗, much the same way as with the tensor product of two vector spaces introduced in the above section on tensor products. In general, there are no relations between and Forcing two such elements to be equal leads to the symmetric algebra, whereas forcing yields the exterior algebra.[89]

[edit]

Vector bundles

[edit]
A Möbius strip. Locally, it looks like U × R.

A vector bundle is a family of vector spaces parametrized continuously by a topological space X.[90] More precisely, a vector bundle over X is a topological space E equipped with a continuous map such that for every x in X, the fiber π−1(x) is a vector space. The case dim V = 1 is called a line bundle. For any vector space V, the projection X × VX makes the product X × V into a "trivial" vector bundle. Vector bundles over X are required to be locally a product of X and some (fixed) vector space V: for every x in X, there is a neighborhood U of x such that the restriction of π to π−1(U) is isomorphic[nb 11] to the trivial bundle U × VU. Despite their locally trivial character, vector bundles may (depending on the shape of the underlying space X) be "twisted" in the large (that is, the bundle need not be (globally isomorphic to) the trivial bundle X × V). For example, the Möbius strip can be seen as a line bundle over the circle S1 (by identifying open intervals with the real line). It is, however, different from the cylinder S1 × R, because the latter is orientable whereas the former is not.[91]

Properties of certain vector bundles provide information about the underlying topological space. For example, the tangent bundle consists of the collection of tangent spaces parametrized by the points of a differentiable manifold. The tangent bundle of the circle S1 is globally isomorphic to S1 × R, since there is a global nonzero vector field on S1.[nb 12] In contrast, by the hairy ball theorem, there is no (tangent) vector field on the 2-sphere S2 which is everywhere nonzero.[92] K-theory studies the isomorphism classes of all vector bundles over some topological space.[93] In addition to deepening topological and geometrical insight, it has purely algebraic consequences, such as the classification of finite-dimensional real division algebras: R, C, the quaternions H and the octonions O.

The cotangent bundle of a differentiable manifold consists, at every point of the manifold, of the dual of the tangent space, the cotangent space. Sections of that bundle are known as differential one-forms.

Modules

[edit]

Modules are to rings what vector spaces are to fields: the same axioms, applied to a ring R instead of a field F, yield modules.[94] The theory of modules, compared to that of vector spaces, is complicated by the presence of ring elements that do not have multiplicative inverses. For example, modules need not have bases, as the Z-module (that is, abelian group) Z/2Z shows; those modules that do (including all vector spaces) are known as free modules. Nevertheless, a vector space can be compactly defined as a module over a ring which is a field, with the elements being called vectors. Some authors use the term vector space to mean modules over a division ring.[95] The algebro-geometric interpretation of commutative rings via their spectrum allows the development of concepts such as locally free modules, the algebraic counterpart to vector bundles.

Affine and projective spaces

[edit]
An affine plane (light blue) in R3. It is a two-dimensional subspace shifted by a vector x (red).

Roughly, affine spaces are vector spaces whose origins are not specified.[96] More precisely, an affine space is a set with a free transitive vector space action. In particular, a vector space is an affine space over itself, by the map If W is a vector space, then an affine subspace is a subset of W obtained by translating a linear subspace V by a fixed vector xW; this space is denoted by x + V (it is a coset of V in W) and consists of all vectors of the form x + v for vV. An important example is the space of solutions of a system of inhomogeneous linear equations generalizing the homogeneous case discussed in the above section on linear equations, which can be found by setting in this equation.[97] The space of solutions is the affine subspace x + V where x is a particular solution of the equation, and V is the space of solutions of the homogeneous equation (the nullspace of A).

The set of one-dimensional subspaces of a fixed finite-dimensional vector space V is known as projective space; it may be used to formalize the idea of parallel lines intersecting at infinity.[98] Grassmannians and flag manifolds generalize this by parametrizing linear subspaces of fixed dimension k and flags of subspaces, respectively.

Notes

[edit]

Citations

[edit]
  1. ^ Lang 2002.
  2. ^ Brown 1991, p. 86.
  3. ^ Roman 2005, ch. 1, p. 27.
  4. ^ Brown 1991, p. 87.
  5. ^ Springer 2000, p. 185; Brown 1991, p. 86.
  6. ^ Atiyah & Macdonald 1969, p. 17.
  7. ^ Bourbaki 1998, §1.1, Definition 2.
  8. ^ Brown 1991, p. 94.
  9. ^ Brown 1991, pp. 99–101.
  10. ^ Brown 1991, p. 92.
  11. ^ a b Stoll & Wong 1968, p. 14.
  12. ^ Roman 2005, pp. 41–42.
  13. ^ Lang 1987, p. 10–11; Anton & Rorres 2010, p. 212.
  14. ^ Blass 1984.
  15. ^ Joshi 1989, p. 450.
  16. ^ Heil 2011, p. 126.
  17. ^ Halmos 1948, p. 12.
  18. ^ Bourbaki 1969, ch. "Algèbre linéaire et algèbre multilinéaire", pp. 78–91.
  19. ^ Bolzano 1804.
  20. ^ Möbius 1827.
  21. ^ Bellavitis 1833.
  22. ^ Dorier 1995.
  23. ^ Hamilton 1853.
  24. ^ Grassmann 2000.
  25. ^ Peano 1888, ch. IX.
  26. ^ Guo 2021.
  27. ^ Moore 1995, pp. 268–271.
  28. ^ Banach 1922.
  29. ^ Dorier 1995; Moore 1995.
  30. ^ Kreyszig 2020, p. 355.
  31. ^ Kreyszig 2020, p. 358–359.
  32. ^ Jain 2001, p. 11.
  33. ^ Lang 1987, ch. I.1.
  34. ^ Lang 2002, ch. V.1.
  35. ^ Lang 1993, ch. XII.3., p. 335.
  36. ^ Lang 1987, ch. VI.3..
  37. ^ Roman 2005, ch. 2, p. 45.
  38. ^ Lang 1987, ch. IV.4, Corollary, p. 106.
  39. ^ Nicholson 2018, ch. 7.3.
  40. ^ Lang 1987, Example IV.2.6.
  41. ^ Lang 1987, ch. VI.6.
  42. ^ Halmos 1974, p. 28, Ex. 9.
  43. ^ Lang 1987, Theorem IV.2.1, p. 95.
  44. ^ Roman 2005, Th. 2.5 and 2.6, p. 49.
  45. ^ Lang 1987, ch. V.1.
  46. ^ Lang 1987, ch. V.3., Corollary, p. 106.
  47. ^ Lang 1987, Theorem VII.9.8, p. 198.
  48. ^ Roman 2005, ch. 8, p. 135–156.
  49. ^ & Lang 1987, ch. IX.4.
  50. ^ Roman 2005, ch. 8, p. 140.
  51. ^ Roman 2005, ch. 1, p. 29.
  52. ^ Roman 2005, ch. 1, p. 35.
  53. ^ Nicholson 2018, ch. 10.4.
  54. ^ Roman 2005, ch. 3, p. 64.
  55. ^ Lang 1987, ch. IV.3..
  56. ^ Roman 2005, ch. 2, p. 48.
  57. ^ Nicholson 2018, ch. 7.4.
  58. ^ Mac Lane 1998.
  59. ^ Roman 2005, ch. 1, pp. 31–32.
  60. ^ Lang 2002, ch. XVI.1.
  61. ^ Roman (2005), Th. 14.3. See also Yoneda lemma.
  62. ^ Rudin 1991, p.3.
  63. ^ Schaefer & Wolff 1999, pp. 204–205.
  64. ^ Bourbaki 2004, ch. 2, p. 48.
  65. ^ Roman 2005, ch. 9.
  66. ^ Naber 2003, ch. 1.2.
  67. ^ Treves 1967; Bourbaki 1987.
  68. ^ Schaefer & Wolff 1999, p. 7.
  69. ^ Kreyszig 1989, §4.11-5
  70. ^ Kreyszig 1989, §1.5-5
  71. ^ Choquet 1966, Proposition III.7.2.
  72. ^ Treves 1967, p. 34–36.
  73. ^ Lang 1983, Cor. 4.1.2, p. 69.
  74. ^ Treves 1967, ch. 11.
  75. ^ Treves 1967, Theorem 11.2, p. 102.
  76. ^ Evans 1998, ch. 5.
  77. ^ Treves 1967, ch. 12.
  78. ^ Dennery & Krzywicki 1996, p.190.
  79. ^ Lang 1993, Th. XIII.6, p. 349.
  80. ^ Lang 1993, Th. III.1.1.
  81. ^ Choquet 1966, Lemma III.16.11.
  82. ^ Kreyszig 1999, Chapter 11.
  83. ^ Griffiths 1995, Chapter 1.
  84. ^ Lang 1993, ch. XVII.3.
  85. ^ Lang 2002, ch. III.1, p. 121.
  86. ^ Eisenbud 1995, ch. 1.6.
  87. ^ Varadarajan 1974.
  88. ^ Lang 2002, ch. XVI.7.
  89. ^ Lang 2002, ch. XVI.8.
  90. ^ Spivak 1999, ch. 3.
  91. ^ Kreyszig 1991, §34, p. 108.
  92. ^ Eisenberg & Guy 1979.
  93. ^ Atiyah 1989.
  94. ^ Artin 1991, ch. 12.
  95. ^ Grillet 2007.
  96. ^ Meyer 2000, Example 5.13.5, p. 436.
  97. ^ Meyer 2000, Exercise 5.13.15–17, p. 442.
  98. ^ Coxeter 1987.

References

[edit]
[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
A vector space, also known as a linear space, is a fundamental consisting of a set VV of elements called vectors, together with two operations: and by elements from a field FF (such as the real numbers R\mathbb{R} or complex numbers C\mathbb{C}). These operations must satisfy ten axioms, including closure under and , associativity and commutativity of , the existence of an (zero vector) and inverses, and distributivity of over and . This framework generalizes the properties of arrows in , allowing vectors to represent not just geometric directions and magnitudes but also abstract quantities like functions, polynomials, or matrices. The concept of a vector space forms the cornerstone of linear algebra, enabling the study of linear transformations, systems of linear equations, and properties such as basis, , and . For instance, the of a vector space is the number of vectors in a basis—a maximal linearly independent spanning set—providing a measure of its "size" independent of the choice of basis. Subspaces, which are subsets that are themselves vector spaces under the induced operations, play a crucial role in decomposing complex spaces into simpler components. Vector spaces have broad applications across , physics, , and , underpinning models in , where state spaces are Hilbert spaces (complete inner product spaces), and in , where high-dimensional datasets are treated as points in vector spaces for techniques like . In , signals and images are represented as vectors, with linear operators (matrices) modeling filters and transformations. Their formalization extends intuitive geometric ideas into rigorous theory, facilitating solutions to differential equations and optimization problems in fields like and .

Formal definition

Axioms of vector spaces

A vector space VV over a field FF is a nonempty set whose elements are called vectors, equipped with two binary operations: vector addition, which combines two vectors to produce another vector in VV, and , which combines an element (scalar) of FF with a vector to produce another vector in VV. Common choices for the field FF include the real numbers R\mathbb{R} or the complex numbers C\mathbb{C}. The operations must satisfy ten axioms, ensuring consistent algebraic behavior. For vectors u,v,wV\mathbf{u}, \mathbf{v}, \mathbf{w} \in V and scalars α,βF\alpha, \beta \in F, the addition operation is denoted u+v\mathbf{u} + \mathbf{v} and the scalar multiplication by αv\alpha \mathbf{v}. These axioms are:
  1. Closure under addition: u+vV\mathbf{u} + \mathbf{v} \in V for all u,vV\mathbf{u}, \mathbf{v} \in V.
  2. Associativity of addition: (u+v)+w=u+(v+w)(\mathbf{u} + \mathbf{v}) + \mathbf{w} = \mathbf{u} + (\mathbf{v} + \mathbf{w}) for all u,v,wV\mathbf{u}, \mathbf{v}, \mathbf{w} \in V.
  3. Commutativity of addition: u+v=v+u\mathbf{u} + \mathbf{v} = \mathbf{v} + \mathbf{u} for all u,vV\mathbf{u}, \mathbf{v} \in V.
  4. Existence of zero vector: There exists a vector 0V\mathbf{0} \in V such that u+0=u\mathbf{u} + \mathbf{0} = \mathbf{u} for all uV\mathbf{u} \in V.
  5. Existence of additive inverses: For each uV\mathbf{u} \in V, there exists uV-\mathbf{u} \in V such that u+(u)=0\mathbf{u} + (-\mathbf{u}) = \mathbf{0}.
  6. Closure under scalar multiplication: αvV\alpha \mathbf{v} \in V for all αF\alpha \in F and vV\mathbf{v} \in V.
  7. Distributivity of scalar multiplication over vector addition: α(u+v)=αu+αv\alpha (\mathbf{u} + \mathbf{v}) = \alpha \mathbf{u} + \alpha \mathbf{v} for all αF\alpha \in F and u,vV\mathbf{u}, \mathbf{v} \in V.
  8. Distributivity of scalar multiplication over field addition: (α+β)v=αv+βv(\alpha + \beta) \mathbf{v} = \alpha \mathbf{v} + \beta \mathbf{v} for all α,βF\alpha, \beta \in F and vV\mathbf{v} \in V.
  9. Compatibility with field multiplication: α(βv)=(αβ)v\alpha (\beta \mathbf{v}) = (\alpha \beta) \mathbf{v} for all α,βF\alpha, \beta \in F and vV\mathbf{v} \in V.
  10. Multiplicative identity: 1v=v1 \cdot \mathbf{v} = \mathbf{v} for all vV\mathbf{v} \in V, where 1 is the multiplicative identity in FF.
The first five axioms establish that (V,+)(V, +) forms an abelian group under addition.

Abelian group structure under addition

In a vector space VV over a field FF, the set VV equipped with the binary operation of vector addition ++ forms an abelian group (V,+)(V, +). This group structure arises directly from the axioms governing addition in the definition of a vector space, which ensure closure under addition (i.e., u+vVu + v \in V for all u,vVu, v \in V), associativity of addition ((u+v)+w=u+(v+w)(u + v) + w = u + (v + w) for all u,v,wVu, v, w \in V), commutativity of addition (u+v=v+uu + v = v + u for all u,vVu, v \in V), the existence of an additive identity (the zero vector 0V0 \in V such that v+0=vv + 0 = v for all vVv \in V), and the existence of additive inverses (for each vVv \in V, there exists vV-v \in V such that v+(v)=0v + (-v) = 0). The identity 0+v=v0 + v = v for all vVv \in V follows from the axioms: 0+v=(v+(v))+v=v+((v)+v)=v+0=v0 + v = (v + (-v)) + v = v + ((-v) + v) = v + 0 = v, using associativity and the inverse property. The zero vector and additive inverses are unique. Suppose 00' is another element satisfying v+0=vv + 0' = v for all vVv \in V. Then 0+0=00 + 0' = 0, and using the group properties, 0=00' = 0. Similarly, if ww satisfies v+w=0v + w = 0, then w=vw = -v, as follows from adding v-v to both sides: w+v+(v)=0+(v)w + v + (-v) = 0 + (-v), so w+0=vw + 0 = -v, hence w=vw = -v. The additive group (V,+)(V, +) relates to the field's additive group (F,+)(F, +) in that both are abelian groups, with VV's structure extending FF's through the module action of scalars, preserving commutativity and other properties inherited from FF.

Basic properties and operations

Scalar multiplication properties

Scalar multiplication in a vector space VV over a field FF associates each scalar αF\alpha \in F and vector vV\mathbf{v} \in V with a vector αvV\alpha \mathbf{v} \in V, satisfying specific axioms that ensure compatibility with the underlying addition structure. These properties include distributivity over vector addition, given by α(u+w)=αu+αw\alpha (\mathbf{u} + \mathbf{w}) = \alpha \mathbf{u} + \alpha \mathbf{w} for all αF\alpha \in F and u,wV\mathbf{u}, \mathbf{w} \in V, which aligns the scaling operation with the Abelian group structure under addition. Similarly, distributivity over scalar addition holds: (α+β)v=αv+βv(\alpha + \beta) \mathbf{v} = \alpha \mathbf{v} + \beta \mathbf{v} for all α,βF\alpha, \beta \in F and vV\mathbf{v} \in V. Homogeneity, or compatibility with field multiplication, ensures that scalar multiplications compose appropriately: (αβ)v=α(βv)(\alpha \beta) \mathbf{v} = \alpha (\beta \mathbf{v}) for all α,βF\alpha, \beta \in F and vV\mathbf{v} \in V. The multiplicative identity in the field acts as the identity for scalar multiplication: 1v=v1 \cdot \mathbf{v} = \mathbf{v} for all vV\mathbf{v} \in V. Additionally, multiplication by the zero scalar yields the zero vector: 0v=00 \cdot \mathbf{v} = \mathbf{0} for all vV\mathbf{v} \in V. To see this, note that v=1v=(1+0)v=1v+0v=v+0v\mathbf{v} = 1 \cdot \mathbf{v} = (1 + 0) \mathbf{v} = 1 \cdot \mathbf{v} + 0 \cdot \mathbf{v} = \mathbf{v} + 0 \cdot \mathbf{v}, so by the cancellation property of addition, 0v=00 \cdot \mathbf{v} = \mathbf{0}. These axioms lead to further corollaries, such as the behavior with the . Specifically, (1)v=v(-1) \mathbf{v} = -\mathbf{v} for all vV\mathbf{v} \in V, where v-\mathbf{v} is the of v\mathbf{v}. This follows from v+(1)v=(1+(1))v=0v=0\mathbf{v} + (-1) \mathbf{v} = (1 + (-1)) \mathbf{v} = 0 \cdot \mathbf{v} = \mathbf{0}, confirming that (1)v(-1) \mathbf{v} serves as the inverse. Another consequence is that α0=0\alpha \cdot \mathbf{0} = \mathbf{0} for all αF\alpha \in F, derived from α0=α(0+0)=α0+α0\alpha \cdot \mathbf{0} = \alpha (\mathbf{0} + \mathbf{0}) = \alpha \cdot \mathbf{0} + \alpha \cdot \mathbf{0}, implying α0=0\alpha \cdot \mathbf{0} = \mathbf{0} by cancellation. For a fixed scalar αF\alpha \in F, the mapping T:VVT: V \to V defined by T(v)=αvT(\mathbf{v}) = \alpha \mathbf{v} preserves both addition and , making it a linear transformation: T(u+w)=α(u+w)=αu+αw=T(u)+T(w)T(\mathbf{u} + \mathbf{w}) = \alpha (\mathbf{u} + \mathbf{w}) = \alpha \mathbf{u} + \alpha \mathbf{w} = T(\mathbf{u}) + T(\mathbf{w}) and T(βv)=α(βv)=(αβ)v=β(αv)=βT(v)T(\beta \mathbf{v}) = \alpha (\beta \mathbf{v}) = (\alpha \beta) \mathbf{v} = \beta (\alpha \mathbf{v}) = \beta T(\mathbf{v}) for all βF\beta \in F and u,w,vV\mathbf{u}, \mathbf{w}, \mathbf{v} \in V.

Vector addition identities

Vector addition in a vector space VV satisfies the axioms of an , ensuring that the operation is both commutative and associative. Commutativity states that for all vectors u,vV\mathbf{u}, \mathbf{v} \in V, u+v=v+u.\mathbf{u} + \mathbf{v} = \mathbf{v} + \mathbf{u}. This axiom guarantees that the order of does not affect the result, mirroring the observed in familiar examples like . Associativity further ensures that the grouping of vectors in a sum is irrelevant: for all u,v,wV\mathbf{u}, \mathbf{v}, \mathbf{w} \in V, (u+v)+w=u+(v+w).(\mathbf{u} + \mathbf{v}) + \mathbf{w} = \mathbf{u} + (\mathbf{v} + \mathbf{w}). This property allows for unambiguous extension of addition to any finite number of vectors, as the result remains consistent regardless of parenthesization. Like commutativity, associativity is a defining of the additive structure in vector spaces. From these axioms, along with the existence of the zero vector and additive inverses, several derived identities follow, including the cancellation law. This law asserts that if x+y=x+y\mathbf{x} + \mathbf{y} = \mathbf{x}' + \mathbf{y} for x,x,yV\mathbf{x}, \mathbf{x}', \mathbf{y} \in V, then x=x\mathbf{x} = \mathbf{x}'. To prove this, add the y-\mathbf{y} to both sides of the equation: (x+y)+(y)=(x+y)+(y).(\mathbf{x} + \mathbf{y}) + (-\mathbf{y}) = (\mathbf{x}' + \mathbf{y}) + (-\mathbf{y}). By associativity, this simplifies to x+(y+(y))=x+(y+(y)),\mathbf{x} + (\mathbf{y} + (-\mathbf{y})) = \mathbf{x}' + (\mathbf{y} + (-\mathbf{y})), and since y+(y)=0\mathbf{y} + (-\mathbf{y}) = \mathbf{0}, where 0\mathbf{0} is the zero vector, it further reduces to x+0=x+0.\mathbf{x} + \mathbf{0} = \mathbf{x}' + \mathbf{0}. Finally, as x+0=x\mathbf{x} + \mathbf{0} = \mathbf{x} and x+0=x\mathbf{x}' + \mathbf{0} = \mathbf{x}' by the zero vector axiom, x=x\mathbf{x} = \mathbf{x}'. This derivation relies solely on the group axioms for and underscores the uniqueness implied by the structure.

Examples

Coordinate spaces over fields

A coordinate space over a field FF, denoted FnF^n, is the set of all ordered nn-tuples (a1,,an)(a_1, \dots, a_n) where each aiFa_i \in F, equipped with componentwise vector addition defined by (a1,,an)+(b1,,bn)=(a1+b1,,an+bn)(a_1, \dots, a_n) + (b_1, \dots, b_n) = (a_1 + b_1, \dots, a_n + b_n) and defined by k(a1,,an)=(ka1,,kan)k(a_1, \dots, a_n) = (ka_1, \dots, ka_n) for kFk \in F. This structure satisfies the vector space axioms over FF, providing a fundamental example of a finite-dimensional vector space. Common instances include Rn\mathbb{R}^n over the real numbers and Cn\mathbb{C}^n over the complex numbers, where the operations inherit the field properties of R\mathbb{R} and C\mathbb{C}. To illustrate, consider R2\mathbb{R}^2 as a prototypical case. The axioms hold via componentwise operations: addition is commutative since (x1,y1)+(x2,y2)=(x1+x2,y1+y2)=(x2+x1,y2+y1)(x_1, y_1) + (x_2, y_2) = (x_1 + x_2, y_1 + y_2) = (x_2 + x_1, y_2 + y_1); associative by the field's ; the zero vector is (0,0)(0, 0); the additive inverse of (x,y)(x, y) is (x,y)(-x, -y); scalar multiplication distributes over in scalars and vectors, associates properly, and satisfies the identity 1(x,y)=(x,y)1 \cdot (x, y) = (x, y). Similarly, for R3\mathbb{R}^3, the verification is analogous, with operations on triples (x,y,z)(x, y, z) proceeding componentwise to confirm all eight axioms, leveraging the arithmetic of R\mathbb{R}. Geometrically, in R2\mathbb{R}^2, elements can be visualized as directed in the , originating from the origin, with represented by placing the tail of one at the head of the other to form a . This interpretation highlights the intuitive role of coordinate spaces in modeling physical quantities like displacement, though it emphasizes the without delving into metrics or inner products. The of FnF^n is nn, directly reflecting the number of independent coordinates required to specify each element.

Function spaces

Function spaces provide examples of infinite-dimensional vector spaces where the elements are functions, equipped with addition and . These spaces illustrate how abstract algebraic structures can apply to continuous or mappings, extending the concept of vectors beyond finite coordinates. A prominent example is the space C[0,1]C[0,1], consisting of all continuous real-valued functions on the closed interval [0,1][0,1], with the field of real numbers R\mathbb{R}. Addition and are defined : for functions f,gC[0,1]f, g \in C[0,1] and scalar αR\alpha \in \mathbb{R}, the sum (f+g)(x)=f(x)+g(x)(f + g)(x) = f(x) + g(x) and the scaled function (αf)(x)=αf(x)(\alpha f)(x) = \alpha f(x) for all x[0,1]x \in [0,1]. This structure ensures closure under these operations, as the sum and scalar multiple of continuous functions remain continuous. Another key example is the space P\mathbb{P} of all polynomials with real coefficients, viewed as functions from R\mathbb{R} to R\mathbb{R}. This space is closed under pointwise addition and scalar multiplication, since the sum of two polynomials is a polynomial and scalar multiplication distributes over coefficients. For instance, if p(x)=a0+a1x++anxnp(x) = a_0 + a_1 x + \cdots + a_n x^n and q(x)=b0+b1x++bmxmq(x) = b_0 + b_1 x + \cdots + b_m x^m, then (p+q)(x)=(a0+b0)+(a1+b1)x+(p + q)(x) = (a_0 + b_0) + (a_1 + b_1) x + \cdots, which is again a polynomial. To confirm these are vector spaces, the operations must satisfy the axioms outlined in the formal definition, such as associativity of , existence of a (the constant function 0), and distributivity of over vector . For C[0,1]C[0,1], the zero vector is the zero function, and additive inverses exist as (f)(x)=f(x)(-f)(x) = -f(x), which is continuous; similar verifications hold for commutativity and scalar properties, leveraging the field structure of R\mathbb{R}. In P\mathbb{P}, the zero polynomial serves as the , and inverses are obtained by negating coefficients, with all axioms following from polynomial arithmetic. Both spaces are thus infinite-dimensional, as they contain linearly independent sets of arbitrary finite size, such as monomials. A concrete illustration in P\mathbb{P} involves linear combinations of basis-like functions, such as the monomials 1,x,x21, x, x^2. Any quadratic polynomial, like 3+2xx23 + 2x - x^2, can be expressed as 31+2x+(1)x23 \cdot 1 + 2 \cdot x + (-1) \cdot x^2, demonstrating how scalar multiples and sums generate elements within the . This extends indefinitely to higher degrees, underscoring the infinite dimensionality.

Polynomial rings as vector spaces

The set of all polynomials with real coefficients, denoted R\mathbb{R}, forms a vector space over the field R\mathbb{R}. Each element of R\mathbb{R} is a polynomial of the form p(x)=k=0nakxkp(x) = \sum_{k=0}^n a_k x^k, where akRa_k \in \mathbb{R} for each kk and nn is a non-negative (possibly infinite in the sense of arbitrary degree, but each individual polynomial has finite degree). Vector addition is defined componentwise on the coefficients: for polynomials p(x)=k=0nakxkp(x) = \sum_{k=0}^n a_k x^k and q(x)=k=0mbkxkq(x) = \sum_{k=0}^m b_k x^k (padding with zeros if necessary), (p+q)(x)=k=0max(n,m)(ak+bk)xk(p + q)(x) = \sum_{k=0}^{\max(n,m)} (a_k + b_k) x^k. by λR\lambda \in \mathbb{R} scales the coefficients: (λp)(x)=k=0n(λak)xk(\lambda p)(x) = \sum_{k=0}^n (\lambda a_k) x^k. These operations satisfy the vector space axioms, with the zero polynomial as the and negation given by multiplying by 1-1. A for R\mathbb{R} is the set of monomials {1,x,x2,}\{1, x, x^2, \dots \}, which is countably infinite. Every p(x)=k=0nakxkp(x) = \sum_{k=0}^n a_k x^k can be uniquely expressed as a finite p(x)=a01+a1x++anxnp(x) = a_0 \cdot 1 + a_1 \cdot x + \dots + a_n \cdot x^n of these basis elements, confirming that they span R\mathbb{R} and are linearly independent (no finite nontrivial equals the zero ). The infinitude of this basis implies that R\mathbb{R} is infinite-dimensional as a vector space over R\mathbb{R}, meaning it has no finite basis. For each non-negative nn, the Rn\mathbb{R}_n consisting of all of degree at most nn is a finite-dimensional subspace of R\mathbb{R}. This subspace has basis {1,x,x2,,xn}\{1, x, x^2, \dots, x^n\} and n+1n+1, as any such polynomial is a unique of these n+1n+1 monomials. The collection of all such Rn\mathbb{R}_n for n=0,1,2,n = 0, 1, 2, \dots forms an increasing of subspaces whose union is R\mathbb{R}, underscoring the infinite-dimensional nature of the full . Although R\mathbb{R} is a ring under the usual polynomial multiplication, the vector space structure considered here focuses solely on addition and of coefficients, independent of the multiplicative operation. This perspective aligns R\mathbb{R} with other infinite-dimensional examples like certain function spaces, but emphasizes its algebraic simplicity via bases.

Basis, dimension, and coordinates

Linear independence and spanning sets

A linear combination of vectors v1,v2,,vkv_1, v_2, \dots, v_k in a vector space VV over a field FF is any vector of the form i=1kαivi\sum_{i=1}^k \alpha_i v_i, where αiF\alpha_i \in F are scalars. The set of all linear combinations of a nonempty subset S={v1,v2,,vk}VS = \{ v_1, v_2, \dots, v_k \} \subseteq V is called the span of SS, denoted span(S)\operatorname{span}(S). The subset SS is a spanning set for VV if span(S)=V\operatorname{span}(S) = V, meaning every vector in VV can be expressed as a of elements from SS. A subset {v1,v2,,vk}V\{ v_1, v_2, \dots, v_k \} \subseteq V is linearly independent if the only solution to the equation i=1kαivi=0\sum_{i=1}^k \alpha_i v_i = 0 is the trivial solution where all scalars αi=0\alpha_i = 0. Equivalently, the set is linearly dependent if there exists a nontrivial linear dependence relation i=1kαivi=0\sum_{i=1}^k \alpha_i v_i = 0 with at least one αi0\alpha_i \neq 0. In the coordinate space Rn\mathbb{R}^n over the field R\mathbb{R}, the standard basis S={e1,e2,,en}S = \{ e_1, e_2, \dots, e_n \}, where eie_i has a 1 in the ii-th position and 0 elsewhere, is a spanning set because every vector (x1,x2,,xn)(x_1, x_2, \dots, x_n) equals i=1nxiei\sum_{i=1}^n x_i e_i. This set is also linearly independent, as the equation i=1nαiei=0\sum_{i=1}^n \alpha_i e_i = 0 implies all αi=0\alpha_i = 0. However, a proper subset of SS, such as {e1,e2,,en1}\{ e_1, e_2, \dots, e_{n-1} \}, is linearly independent but does not span Rn\mathbb{R}^n, since vectors with nonzero nn-th coordinate cannot be expressed as their linear combinations.

Hamel bases and dimension

A basis for a vector space VV over a field KK is a BVB \subseteq V that is linearly independent and spans VV, meaning every element of VV can be expressed as a finite of elements from BB. In finite-dimensional spaces, bases consist of finitely many elements, but in infinite-dimensional spaces, bases may be infinite and are specifically termed Hamel bases to distinguish them from other notions of bases used in , such as Schauder bases. The concept of a Hamel basis originates from the work of Georg Hamel, who demonstrated its existence for R\mathbb{R} as a vector space over Q\mathbb{Q} in 1905. The existence of a Hamel basis for any vector space VV (including the zero space, where the empty set serves as the basis) is established using Zorn's lemma, a consequence of the axiom of choice in set theory. Consider the collection of all linearly independent subsets of VV, partially ordered by inclusion; any chain in this poset has an upper bound given by its union, which remains linearly independent. By Zorn's lemma, there exists a maximal linearly independent subset BB, and maximality implies that BB spans VV, as adjoining any additional vector would violate linear independence. The dimension of a vector space VV, denoted dim(V)\dim(V), is defined as the cardinality of any Hamel basis of VV; if this cardinality is finite, VV is finite-dimensional, whereas infinite cardinality indicates an infinite-dimensional space. This definition is unambiguous because any two Hamel bases of VV have the same cardinality: suppose B1B_1 and B2B_2 are bases with B1<B2|B_1| < |B_2|; then B1B_1 can be extended to a basis of cardinality B2|B_2|, but since B2B_2 spans VV, it cannot contain a linearly independent subset larger than B1|B_1| without contradicting the spanning property of B1B_1, leading to a contradiction. A key consequence is that any linearly independent subset SVS \subseteq V can be extended to a Hamel basis of VV. To see this, apply Zorn's lemma to the poset of linearly independent subsets containing SS, ordered by inclusion; chains have unions as upper bounds, so a maximal element BB containing SS must span VV. This extension property underscores the foundational role of bases in vector space theory.

Coordinate representations

In a finite-dimensional vector space VV over a field FF with basis B={e1,,en}B = \{e_1, \dots, e_n\}, every vector vVv \in V can be expressed uniquely as a linear combination v=i=1nxieiv = \sum_{i=1}^n x_i e_i, where the scalars xiFx_i \in F. The ordered tuple (x1,,xn)(x_1, \dots, x_n) is called the coordinate representation of vv with respect to BB, denoted BFn_B \in F^n. This coordinate map ϕB:FnV\phi_B: F^n \to V defined by ϕB(x1,,xn)=i=1nxiei\phi_B(x_1, \dots, x_n) = \sum_{i=1}^n x_i e_i is an isomorphism of vector spaces, ensuring a one-to-one correspondence between vectors in VV and their coordinate tuples. The uniqueness of coordinates follows directly from the definition of a basis: the set BB spans VV, so every vv has at least one representation as a linear combination, while linear independence ensures no two distinct combinations yield the same vector. In infinite-dimensional spaces, such uniqueness may fail without additional structure like Hamel bases, but in finite dimensions, it holds for any basis. To relate coordinates across different bases, suppose BB' is another basis for VV. Let PP be the n×nn \times n matrix over FF whose columns are the coordinate vectors [e1]B,,[en]B[e'_1]_B, \dots, [e'_n]_B, where {e1,,en}=B\{e'_1, \dots, e'_n\} = B'. Then the coordinates transform via B=PB1_{B'} = P^{-1} _B, since PP is invertible as the bases have the same cardinality. This formula introduces the change-of-basis matrix without altering the intrinsic properties of vv. A concrete example occurs in the real vector space Rn\mathbb{R}^n with the standard basis E={e1,,en}E = \{e_1, \dots, e_n\}, where eie_i has 1 in the ii-th position and 0 elsewhere. For any v=(v1,,vn)Rnv = (v_1, \dots, v_n) \in \mathbb{R}^n, the coordinates E=(v1,,vn)_E = (v_1, \dots, v_n) coincide with the usual component representation, simplifying computations in Euclidean space.

Subspaces and quotient spaces

Defining subspaces

A subspace of a vector space VV over a field FF is a subset WVW \subseteq V that contains the zero vector and is closed under vector addition and scalar multiplication by elements of FF. This means that for all u,vW\mathbf{u}, \mathbf{v} \in W and all cFc \in F, both u+vW\mathbf{u} + \mathbf{v} \in W and cuWc \mathbf{u} \in W. Equivalently, WW is a subspace if it is a vector space in its own right under the addition and scalar multiplication operations induced from VV. Common examples of subspaces include the trivial subspace {0}\{\mathbf{0}\}, which consists solely of the zero vector and satisfies the conditions vacuously, and the entire space VV itself. Another important example is the span of a subset SVS \subseteq V, denoted span(S)\operatorname{span}(S), which is the set of all finite linear combinations of elements from SS and forms the smallest subspace containing SS. Additionally, the solution set to a system of homogeneous linear equations Ax=0A\mathbf{x} = \mathbf{0}, where AA is a matrix over FF, is a subspace of the coordinate space, as it is closed under addition and scalar multiplication. A key property is that the intersection of any collection of subspaces of VV is itself a subspace of VV. To see this, note that the zero vector belongs to every subspace, and if u,v\mathbf{u}, \mathbf{v} are in the intersection, then u+v\mathbf{u} + \mathbf{v} and cuc\mathbf{u} remain in each original subspace, hence in their intersection. In contrast, the union of two subspaces is generally not a subspace unless one is contained in the other; for instance, the union of the x-axis and y-axis in R2\mathbb{R}^2 fails closure under addition, as (1,0)+(0,1)=(1,1)(1,0) + (0,1) = (1,1) lies outside both.

Quotient space construction

Given a subspace WW of a vector space VV over a field FF, the cosets of WW in VV are the sets of the form v+W={v+wwW}v + W = \{v + w \mid w \in W\} for vVv \in V. These cosets partition VV into equivalence classes, where two vectors v1,v2Vv_1, v_2 \in V are equivalent modulo WW if v1v2Wv_1 - v_2 \in W. The quotient space V/WV/W is the set of all such cosets {v+WvV}\{v + W \mid v \in V\}, equipped with vector space operations defined by (v+W)+(u+W)=(v+u)+W(v + W) + (u + W) = (v + u) + W for addition and α(v+W)=αv+W\alpha (v + W) = \alpha v + W for scalar multiplication by αF\alpha \in F. These operations are well-defined, independent of the choice of representatives, because if v+W=v+Wv' + W = v + W and u+W=u+Wu' + W = u + W, then v=v+w1v' = v + w_1 and u=u+w2u' = u + w_2 for some w1,w2Ww_1, w_2 \in W, so v+u+W=v+u+(w1+w2)+W=v+u+Wv' + u' + W = v + u + (w_1 + w_2) + W = v + u + W. To verify that V/WV/W is a vector space over FF, the operations satisfy the vector space axioms: addition is associative and commutative, with zero element 0+W=W0 + W = W and additive inverse v+W-v + W; scalar multiplication distributes over vector addition and field multiplication, and satisfies α(β(v+W))=(αβ)(v+W)\alpha (\beta (v + W)) = (\alpha \beta)(v + W) and 1(v+W)=v+W1 \cdot (v + W) = v + W. Closure follows from the definitions, and all properties inherit from those of VV. If VV is finite-dimensional, the dimension theorem states that dim(V)=dim(W)+dim(V/W)\dim(V) = \dim(W) + \dim(V/W). To see this, extend a basis of WW to a basis of VV, and the images of the additional basis vectors under the quotient map form a basis for V/WV/W. The natural projection π:VV/W\pi: V \to V/W given by π(v)=v+W\pi(v) = v + W is a surjective linear map with kernel WW.

Linear transformations

Definition and properties

A linear transformation, also known as a linear map, from a vector space VV to a vector space WW over the same field is a function T:VWT: V \to W that preserves vector addition and scalar multiplication. Specifically, for all u,vVu, v \in V and scalars α\alpha in the field, T(u+v)=T(u)+T(v)T(u + v) = T(u) + T(v) and T(αv)=αT(v)T(\alpha v) = \alpha T(v). The kernel of TT, denoted ker(T)\ker(T), is the set of all vectors in VV that map to the zero vector in WW, i.e., ker(T)={vVT(v)=0}\ker(T) = \{ v \in V \mid T(v) = 0 \}, which forms a subspace of VV. The image of TT, denoted im(T)\operatorname{im}(T), is the set of all vectors in WW that are outputs of TT, i.e., im(T)={T(v)vV}\operatorname{im}(T) = \{ T(v) \mid v \in V \}, which forms a subspace of WW. A key property is that TT is linear if and only if it preserves arbitrary finite linear combinations: for any finite collection of vectors v1,,vnVv_1, \dots, v_n \in V and scalars α1,,αn\alpha_1, \dots, \alpha_n, T(i=1nαivi)=i=1nαiT(vi)T\left( \sum_{i=1}^n \alpha_i v_i \right) = \sum_{i=1}^n \alpha_i T(v_i). Additionally, TT is injective (one-to-one) if and only if its kernel is the trivial subspace {0}\{0\}. An isomorphism between vector spaces VV and WW is a bijective linear map T:VWT: V \to W whose inverse T1:WVT^{-1}: W \to V is also linear. Such maps establish that VV and WW have the same structure as vector spaces.

Kernel, image, and rank-nullity theorem

For a linear transformation T:VWT: V \to W between vector spaces over a field, the kernel of TT, denoted ker(T)\ker(T), is the set {vVT(v)=0}\{v \in V \mid T(v) = 0\}. This set forms a subspace of the domain VV, as it is closed under addition and scalar multiplication, and contains the zero vector. Similarly, the image of TT, denoted im(T)\operatorname{im}(T), is the set {T(v)vV}\{T(v) \mid v \in V\}, which is a subspace of the codomain WW because the image of a linear combination is the linear combination of the images. The nullity of TT, denoted n(T)n(T), is defined as the dimension of ker(T)\ker(T). The rank of TT, denoted r(T)r(T), is the dimension of im(T)\operatorname{im}(T). These quantities measure the "degeneracy" and "reach" of the transformation, respectively: a higher nullity indicates more vectors are mapped to zero, while a higher rank reflects a larger subspace spanned by the outputs. The rank-nullity theorem states that if VV is finite-dimensional, then dim(V)=r(T)+n(T)\dim(V) = r(T) + n(T). This fundamental result connects the dimensions of the domain, kernel, and image, providing insight into the structure of linear maps. To prove the rank-nullity theorem, suppose dim(V)=n<\dim(V) = n < \infty and let k=n(T)=dim(ker(T))k = n(T) = \dim(\ker(T)). Choose a basis {u1,,uk}\{u_1, \dots, u_k\} for ker(T)\ker(T). Extend this to a basis {u1,,uk,v1,,vm}\{u_1, \dots, u_k, v_1, \dots, v_m\} for VV, where m=nkm = n - k. Since T(ui)=0T(u_i) = 0 for i=1,,ki = 1, \dots, k, the images T(v1),,T(vm)T(v_1), \dots, T(v_m) lie in im(T)\operatorname{im}(T). These images form a spanning set for im(T)\operatorname{im}(T), as any T(w)T(w) for wVw \in V can be expressed using the basis coefficients. Moreover, {T(v1),,T(vm)}\{T(v_1), \dots, T(v_m)\} is linearly independent: if cjT(vj)=0\sum c_j T(v_j) = 0, then T(cjvj)=0T(\sum c_j v_j) = 0, so cjvjker(T)\sum c_j v_j \in \ker(T), implying all cj=0c_j = 0 by basis extension properties. Thus, dim(im(T))=m\dim(\operatorname{im}(T)) = m, so r(T)=m=nkr(T) = m = n - k, and dim(V)=r(T)+n(T)\dim(V) = r(T) + n(T). As a consequence, if dim(V)=n<\dim(V) = n < \infty, then r(T)nr(T) \leq n, since n(T)0n(T) \geq 0. Additionally, r(T)dim(W)r(T) \leq \dim(W) because im(T)\operatorname{im}(T) is a subspace of WW. Therefore, r(T)min(n,dim(W))r(T) \leq \min(n, \dim(W)).

Matrix representations

Linear maps as matrices

Given finite-dimensional vector spaces VV and WW over the same field, with dimV=n\dim V = n and dimW=m\dim W = m, and ordered bases B={e1,,en}\mathcal{B} = \{e_1, \dots, e_n\} for VV and C={f1,,fm}\mathcal{C} = \{f_1, \dots, f_m\} for WW, any linear map T:VWT: V \to W can be represented by an m×nm \times n matrix AA relative to these bases. The columns of AA are the coordinate vectors [T(ei)]C[T(e_i)]_{\mathcal{C}} with respect to C\mathcal{C}, for i=1,,ni = 1, \dots, n. This matrix AA encodes the action of TT on coordinate representations: for any vector vVv \in V, the coordinate vector [T(v)]C[T(v)]_{\mathcal{C}} equals ABA _{\mathcal{B}}, where B_{\mathcal{B}} is the coordinate vector of vv with respect to B\mathcal{B}. This correspondence arises because T(v)=T(i=1nxiei)=i=1nxiT(ei)T(v) = T\left( \sum_{i=1}^n x_i e_i \right) = \sum_{i=1}^n x_i T(e_i), and expressing the images T(ei)T(e_i) in coordinates yields the matrix-vector product. The matrix AA is unique for the given bases B\mathcal{B} and C\mathcal{C}, as the coordinate representations [T(ei)]C[T(e_i)]_{\mathcal{C}} are uniquely determined by the spanning and linear independence properties of the bases. For a concrete example, consider the rotation linear map Rθ:R2R2R_\theta: \mathbb{R}^2 \to \mathbb{R}^2 by an angle θ\theta counterclockwise, with respect to the standard basis E={e1=(1,0),e2=(0,1)}\mathcal{E} = \{e_1 = (1,0), e_2 = (0,1)\}. The matrix of RθR_\theta is (cosθsinθsinθcosθ),\begin{pmatrix} \cos \theta & -\sin \theta \\ \sin \theta & \cos \theta \end{pmatrix}, since Rθ(e1)=(cosθ,sinθ)R_\theta(e_1) = (\cos \theta, \sin \theta) and Rθ(e2)=(sinθ,cosθ)R_\theta(e_2) = (-\sin \theta, \cos \theta). Applying this matrix to a vector (x,y)(x,y) yields the rotated coordinates (cosθxsinθy,sinθx+cosθy)(\cos \theta \cdot x - \sin \theta \cdot y, \sin \theta \cdot x + \cos \theta \cdot y).

Change of basis matrices

In a finite-dimensional vector space VV over a field F\mathbb{F}, consider two bases B={e1,,en}B = \{e_1, \dots, e_n\} and B={e1,,en}B' = \{e'_1, \dots, e'_n\}. The change-of-basis matrix PP from BB to BB' is the invertible n×nn \times n matrix whose columns are the coordinate vectors [ei]B[e'_i]_B of the vectors in the new basis BB' expressed with respect to the old basis BB. This matrix PP satisfies the relation that for any vector vVv \in V, the coordinates transform as B=PB1_{B'} = P^{-1} _B, ensuring that the vector representation remains consistent across bases since v=Biei=Bieiv = \sum _B^i e_i = \sum _{B'}^i e'_i. For a linear transformation T:VWT: V \to W between finite-dimensional vector spaces, with bases BB and BB' in VV, and bases CC and CC' in WW, let PP be the change-of-basis matrix from BB to BB' in VV, and let QQ be the change-of-basis matrix from CC to CC' in WW (defined analogously, with columns [fj]C[f'_j]_C). If AA is the matrix of TT with respect to the bases BB and CC, then the matrix AA' of TT with respect to BB' and CC' is given by the transformation A=Q1APA' = Q^{-1} A P. This formula arises from substituting the coordinate relations: [Tv]C=AB[T v]_C = A _B, so [Tv]C=Q1[Tv]C=Q1AB=Q1APB[T v]_{C'} = Q^{-1} [T v]_C = Q^{-1} A _B = Q^{-1} A P _{B'}, yielding AA' directly. When TT is an endomorphism on VV (so W=VW = V and the same basis change applies, with Q=PQ = P), the formula simplifies to the similarity transformation A=P1APA' = P^{-1} A P. This preserves key invariants like the trace, determinant, and characteristic polynomial of the matrix, as similarity reflects the intrinsic properties of the linear operator independent of basis choice. The proof follows from ensuring consistency with the linear action: the transformed matrix must satisfy T(ei)=jAjiejT(e'_i) = \sum_j A'_{ji} e'_j for all ii, which holds by expressing T(ei)T(e'_i) in the new basis using the original matrix AA and the coordinate conversions via PP.

Algebraic constructions

Direct sums and products

The external direct sum of two vector spaces VV and WW over the same field FF, denoted VWV \oplus W, consists of all ordered pairs (v,w)(v, w) with vVv \in V and wWw \in W, equipped with componentwise addition (v1,w1)+(v2,w2)=(v1+v2,w1+w2)(v_1, w_1) + (v_2, w_2) = (v_1 + v_2, w_1 + w_2) and scalar multiplication c(v,w)=(cv,cw)c(v, w) = (cv, cw) for cFc \in F. This construction forms a vector space whose elements can be thought of as formal combinations of vectors from VV and WW without overlap. In contrast, the internal direct sum arises within a single vector space VV that decomposes into the sum of two subspaces UU and WW, written V=UWV = U \oplus W, if every vector in VV can be uniquely expressed as v=u+wv = u + w with uUu \in U and wWw \in W, which holds precisely when U+W=VU + W = V and UW={0}U \cap W = \{0\}. This condition ensures that the decomposition is unique, distinguishing it from a general sum of subspaces. The internal direct sum corresponds to the external direct sum via the canonical isomorphism that identifies VV with UWU \oplus W when the conditions are met. For a finite collection of vector spaces, the direct product coincides with the direct sum up to isomorphism; specifically, the direct product i=1nVi\prod_{i=1}^n V_i is the set of all nn-tuples (v1,,vn)(v_1, \dots, v_n) with viViv_i \in V_i and componentwise operations, which is isomorphic to the direct sum i=1nVi\bigoplus_{i=1}^n V_i in the finite case. This equivalence simplifies constructions in finite dimensions, where the notions are often used interchangeably. If V=UWV = U \oplus W is an internal direct sum and {ui}\{u_i\} is a basis for UU while {wj}\{w_j\} is a basis for WW, then the union {ui}{wj}\{u_i\} \cup \{w_j\} forms a basis for VV. Consequently, the dimension satisfies dim(VW)=dimV+dimW\dim(V \oplus W) = \dim V + \dim W for the external direct sum, and similarly for internal decompositions.

Tensor products of vector spaces

The tensor product of two vector spaces VV and WW over a field KK, denoted VKWV \otimes_K W, is a vector space equipped with a bilinear map :V×WVKW\otimes: V \times W \to V \otimes_K W that satisfies the relations (v1+v2)w=v1w+v2w(v_1 + v_2) \otimes w = v_1 \otimes w + v_2 \otimes w, v(w1+w2)=vw1+vw2v \otimes (w_1 + w_2) = v \otimes w_1 + v \otimes w_2, and (αv)w=v(αw)=α(vw)(\alpha v) \otimes w = v \otimes (\alpha w) = \alpha (v \otimes w) for all v,v1,v2Vv, v_1, v_2 \in V, w,w1,w2Ww, w_1, w_2 \in W, and αK\alpha \in K. This construction generates VKWV \otimes_K W as the span of elements of the form vwv \otimes w, subject to these bilinearity conditions, ensuring that the map \otimes is KK-bilinear. The tensor product satisfies a universal property: for any vector space UU and any KK-bilinear map ϕ:V×WU\phi: V \times W \to U, there exists a unique KK-linear map ϕ~:VKWU\tilde{\phi}: V \otimes_K W \to U such that ϕ~(vw)=ϕ(v,w)\tilde{\phi}(v \otimes w) = \phi(v, w) for all vVv \in V, wWw \in W. This property characterizes the tensor product up to unique isomorphism and allows bilinear maps to factor uniquely through the linear map on the tensor product. Suppose {ei}i=1n\{e_i\}_{i=1}^n is a basis for a finite-dimensional vector space VV over KK and {fj}j=1m\{f_j\}_{j=1}^m is a basis for WW. Then {eifj}i=1,,n;j=1,,m\{e_i \otimes f_j\}_{i=1,\dots,n; j=1,\dots,m} forms a basis for VKWV \otimes_K W, consisting of nmnm elements. Consequently, the dimension of the tensor product is the product of the dimensions: dimK(VKW)=(dimKV)(dimKW)\dim_K(V \otimes_K W) = (\dim_K V) \cdot (\dim_K W). For example, over the field R\mathbb{R}, the tensor product RmRRn\mathbb{R}^m \otimes_\mathbb{R} \mathbb{R}^n is isomorphic as a vector space to Rmn\mathbb{R}^{mn}, where the isomorphism arises from mapping the standard basis elements eifje_i \otimes f_j to the standard basis of Rmn\mathbb{R}^{mn}.

Vector spaces with metric structure

Normed vector spaces

A normed vector space is a vector space VV over the real or complex numbers equipped with a norm :V[0,)\|\cdot\|: V \to [0, \infty), which measures the "length" of vectors and satisfies three fundamental axioms. These axioms ensure the norm behaves consistently with the vector space operations of addition and scalar multiplication. The norm satisfies positivity: v0\|v\| \geq 0 for all vVv \in V, with equality if and only if v=0v = 0; absolute homogeneity: αv=αv\|\alpha v\| = |\alpha| \|v\| for every scalar α\alpha and vector vVv \in V; and the triangle inequality: u+vu+v\|u + v\| \leq \|u\| + \|v\| for all u,vVu, v \in V. These properties make the norm a natural extension of intuitive notions of distance and size in familiar spaces like Rn\mathbb{R}^n. The norm induces a metric on VV defined by d(u,v)=uvd(u, v) = \|u - v\| for u,vVu, v \in V, which satisfies the axioms of a metric space: non-negativity, symmetry, the identity of indiscernibles, and the triangle inequality. This metric structure allows the application of topological concepts to vector spaces, such as convergence and continuity, while preserving linearity. Common examples include the Euclidean norm on Rn\mathbb{R}^n, given by x2=i=1nxi2\|x\|_2 = \sqrt{\sum_{i=1}^n x_i^2}
Add your contribution
Related Hubs
Contribute something
User Avatar
No comments yet.