Hubbry Logo
PolynomialPolynomialMain
Open search
Polynomial
Community hub
Polynomial
logo
7 pages, 0 posts
0 subscribers
Be the first to start a discussion here.
Be the first to start a discussion here.
Polynomial
Polynomial
from Wikipedia

In mathematics, a polynomial is a mathematical expression consisting of indeterminates (also called variables) and coefficients, that involves only the operations of addition, subtraction, multiplication and exponentiation to nonnegative integer powers, and has a finite number of terms.[1][2][3][4][5] An example of a polynomial of a single indeterminate is . An example with three indeterminates is .

Polynomials appear in many areas of mathematics and science. For example, they are used to form polynomial equations, which encode a wide range of problems, from elementary word problems to complicated scientific problems; they are used to define polynomial functions, which appear in settings ranging from basic chemistry and physics to economics and social science; and they are used in calculus and numerical analysis to approximate other functions. In advanced mathematics, polynomials are used to construct polynomial rings and algebraic varieties, which are central concepts in algebra and algebraic geometry.

Etymology

[edit]

The word polynomial joins two diverse roots: the Greek poly, meaning "many", and the Latin nomen, or "name". It was derived from the term binomial by replacing the Latin root bi- with the Greek poly-. That is, it means a sum of many terms (many monomials). The word polynomial was first used in the 17th century.[6]

Notation and terminology

[edit]
The graph of a polynomial function of degree 3

The occurring in a polynomial is commonly called a variable or an indeterminate. When the polynomial is considered as an expression, is a fixed symbol which does not have any value (its value is "indeterminate"). However, when one considers the function defined by the polynomial, then represents the argument of the function, and is therefore called a "variable". Many authors use these two words interchangeably.

A polynomial in the indeterminate is commonly denoted either as or as . Formally, the name of the polynomial is , not , but the use of the functional notation dates from a time when the distinction between a polynomial and the associated function was unclear. Moreover, the functional notation is often useful for specifying, in a single phrase, a polynomial and its indeterminate. For example, "let be a polynomial" is a shorthand for "let be a polynomial in the indeterminate ". On the other hand, when it is not necessary to emphasize the name of the indeterminate, many formulas are much simpler and easier to read if the name(s) of the indeterminate(s) do not appear at each occurrence of the polynomial.

The ambiguity of having two notations for a single mathematical object may be formally resolved by considering the general meaning of the functional notation for polynomials. If denotes a number, a variable, another polynomial, or, more generally, any expression, then denotes, by convention, the result of substituting for in . Thus, the polynomial defines the function which is the polynomial function associated to . Frequently, when using this notation, one supposes that is a number. However, one may use it over any domain where addition and multiplication are defined (that is, any ring). In particular, if is a polynomial then is also a polynomial.

More specifically, when is the indeterminate , then the image of by this function is the polynomial itself (substituting for does not change anything). In other words, which justifies formally the existence of two notations for the same polynomial.

Definition

[edit]

A polynomial expression is an expression that can be built from constants and symbols called variables or indeterminates by means of addition, multiplication and exponentiation to a non-negative integer power. The constants are generally numbers, but may be any expression that do not involve the indeterminates, and represent mathematical objects that can be added and multiplied. Two polynomial expressions are considered as defining the same polynomial if they may be transformed, one to the other, by applying the usual properties of commutativity, associativity and distributivity of addition and multiplication. For example and are two polynomial expressions that represent the same polynomial; so, one has the equality .

A polynomial in a single indeterminate x can always be written (or rewritten) in the form where are constants that are called the coefficients of the polynomial, and is the indeterminate.[7] The word "indeterminate" means that represents no particular value, although any value may be substituted for it. The mapping that associates the result of this substitution to the substituted value is a function, called a polynomial function.

This can be expressed more concisely by using summation notation: That is, a polynomial can either be zero or can be written as the sum of a finite number of non-zero terms. Each term consists of the product of a number – called the coefficient of the term[a] – and a finite number of indeterminates, raised to non-negative integer powers.

Classification

[edit]

The exponent on an indeterminate in a term is called the degree of that indeterminate in that term; the degree of the term is the sum of the degrees of the indeterminates in that term, and the degree of a polynomial is the largest degree of any term with nonzero coefficient.[8] Because , the degree of an indeterminate without a written exponent is one.

A term with no indeterminates and a polynomial with no indeterminates are called, respectively, a constant term and a constant polynomial.[b] The degree of a constant term and of a nonzero constant polynomial is . The degree of the zero polynomial (which has no terms at all) is generally treated as not defined (but see below).[9]

For example: is a term. The coefficient is , the indeterminates are and , the degree of is two, while the degree of is one. The degree of the entire term is the sum of the degrees of each indeterminate in it, so in this example the degree is .

Forming a sum of several terms produces a polynomial. For example, the following is a polynomial: It consists of three terms: the first is degree two, the second is degree one, and the third is degree zero.

Polynomials of small degree have been given specific names. A polynomial of degree zero is a constant polynomial, or simply a constant. Polynomials of degree one, two or three are respectively linear polynomials, quadratic polynomials and cubic polynomials.[8] For higher degrees, the specific names are not commonly used, although quartic polynomial (for degree four) and quintic polynomial (for degree five) are sometimes used. The names for the degrees may be applied to the polynomial or to its terms. For example, the term in is a linear term in a quadratic polynomial.

The polynomial , which may be considered to have no terms at all, is called the zero polynomial. Unlike other constant polynomials, its degree is not zero. Rather, the degree of the zero polynomial is either left explicitly undefined, or defined as negative (either −1 or ).[10] The zero polynomial is also unique in that it is the only polynomial in one indeterminate that has an infinite number of roots. The graph of the zero polynomial, , is the -axis.

In the case of polynomials in more than one indeterminate, a polynomial is called homogeneous of degree if all of its non-zero terms have degree . The zero polynomial is homogeneous, and, as a homogeneous polynomial, its degree is undefined.[c] For example, is homogeneous of degree . For more details, see homogeneous polynomials.

The commutative law of addition can be used to rearrange terms into any preferred order. In polynomials with one indeterminate, the terms are usually ordered according to degree, either in "descending powers of ", with the term of largest degree first, or in "ascending powers of ". The polynomial is written in descending powers of . The first term has coefficient , indeterminate , and exponent . In the second term, the coefficient is . The third term is a constant. Because the degree of a non-zero polynomial is the largest degree of any one term, this polynomial has degree two.[11]

Two terms with the same indeterminates raised to the same powers are called "similar terms" or "like terms", and they can be combined, using the distributive law, into a single term whose coefficient is the sum of the coefficients of the terms that were combined. It may happen that this makes the coefficient .[12]

Polynomials can be classified by the number of terms with nonzero coefficients, so that a one-term polynomial is called a monomial,[d] a two-term polynomial is called a binomial, and a three-term polynomial is called a trinomial. A polynomial with two or more terms is also called a multinomial.[13][14]

A real polynomial is a polynomial with real coefficients. When it is used to define a function, the domain is not so restricted. However, a real polynomial function is a function from the reals to the reals that is defined by a real polynomial. Similarly, an integer polynomial is a polynomial with integer coefficients, and a complex polynomial is a polynomial with complex coefficients.

A polynomial in one indeterminate is called a univariate polynomial, a polynomial in more than one indeterminate is called a multivariate polynomial.[15] A polynomial with two indeterminates is called a bivariate polynomial.[7] These notions refer more to the kind of polynomials one is generally working with than to individual polynomials; for instance, when working with univariate polynomials, one does not exclude constant polynomials (which may result from the subtraction of non-constant polynomials), although strictly speaking, constant polynomials do not contain any indeterminates at all. It is possible to further classify multivariate polynomials as bivariate, trivariate, and so on, according to the maximum number of indeterminates allowed. Again, so that the set of objects under consideration be closed under subtraction, a study of trivariate polynomials usually allows bivariate polynomials, and so on. It is also common to say simply "polynomials in , and ", listing the indeterminates allowed.

Operations

[edit]

Addition and subtraction

[edit]

Polynomials can be added using the associative law of addition (grouping all their terms together into a single sum), possibly followed by reordering (using the commutative law) and combining of like terms.[12][16] For example, if and then the sum can be reordered and regrouped as and then simplified to When polynomials are added together, the result is another polynomial.[17]

Subtraction of polynomials is similar.

Multiplication

[edit]

Polynomials can also be multiplied. To expand the product of two polynomials into a sum of terms, the distributive law is repeatedly applied, which results in each term of one polynomial being multiplied by every term of the other.[12] For example, if then Carrying out the multiplication in each term produces Combining similar terms yields which can be simplified to As in the example, the product of polynomials is always a polynomial.[17][9]

Composition

[edit]

Given a polynomial of a single variable and another polynomial of any number of variables, the composition is obtained by substituting each copy of the variable of the first polynomial by the second polynomial.[9] For example, if and then A composition may be expanded to a sum of terms using the rules for multiplication and division of polynomials. The composition of two polynomials is another polynomial.[18]

Division

[edit]

The division of one polynomial by another is not typically a polynomial. Instead, such ratios are a more general family of objects, called rational fractions, rational expressions, or rational functions, depending on context.[19] This is analogous to the fact that the ratio of two integers is a rational number, not necessarily an integer.[20][21] For example, the fraction is not a polynomial, and it cannot be written as a finite sum of powers of the variable .

For polynomials in one variable, there is a notion of Euclidean division of polynomials, generalizing the Euclidean division of integers.[e] This notion of the division results in two polynomials, a quotient and a remainder , such that and , where is the degree of . The quotient and remainder may be computed by any of several algorithms, including polynomial long division and synthetic division.[22]

When the denominator is monic and linear, that is, for some constant , then the polynomial remainder theorem asserts that the remainder of the division of by is the evaluation .[21] In this case, the quotient may be computed by Ruffini's rule, a special case of synthetic division.[23]

Factoring

[edit]

All polynomials with coefficients in a unique factorization domain (for example, the integers or a field) also have a factored form in which the polynomial is written as a product of irreducible polynomials and a constant. This factored form is unique up to the order of the factors and their multiplication by an invertible constant. In the case of the field of complex numbers, the irreducible factors are linear. Over the real numbers, they have the degree either one or two. Over the integers and the rational numbers the irreducible factors may have any degree.[24] For example, the factored form of is over the integers and the reals, and over the complex numbers.

The computation of the factored form, called factorization is, in general, too difficult to be done by hand-written computation. However, efficient polynomial factorization algorithms are available in most computer algebra systems.

Calculus

[edit]

Calculating derivatives and integrals of polynomials is particularly simple, compared to other kinds of functions. The derivative of the polynomial with respect to is the polynomial Similarly, the general antiderivative (or indefinite integral) of is where is an arbitrary constant. For example, antiderivatives of have the form .

For polynomials whose coefficients come from more abstract settings (for example, if the coefficients are integers modulo some prime number , or elements of an arbitrary ring), the formula for the derivative can still be interpreted formally, with the coefficient understood to mean the sum of copies of . For example, over the integers modulo , the derivative of the polynomial is the polynomial .[25]

Polynomial functions

[edit]

A polynomial function is a function that can be defined by evaluating a polynomial. More precisely, a function of one argument from a given domain is a polynomial function if there exists a polynomial that evaluates to for all x in the domain of (here, is a non-negative integer and are constant coefficients).[26] Generally, unless otherwise specified, polynomial functions have complex coefficients, arguments, and values. In particular, a polynomial, restricted to have real coefficients, defines a function from the complex numbers to the complex numbers. If the domain of this function is also restricted to the reals, the resulting function is a real function that maps reals to reals.

For example, the function , defined by is a polynomial function of one variable. Polynomial functions of several variables are similarly defined, using polynomials in more than one indeterminate, as in According to the definition of polynomial functions, there may be expressions that obviously are not polynomials but nevertheless define polynomial functions. An example is the expression which takes the same values as the polynomial on the interval , and thus both expressions define the same polynomial function on this interval.

Every polynomial function is continuous, smooth, and entire.

The evaluation of a polynomial is the computation of the corresponding polynomial function; that is, the evaluation consists of substituting a numerical value to each indeterminate and carrying out the indicated multiplications and additions.

For polynomials in one indeterminate, the evaluation is usually more efficient (lower number of arithmetic operations to perform) using Horner's method, which consists of rewriting the polynomial as

Graphs

[edit]

A polynomial function in one real variable can be represented by a graph.

  • The graph of the zero polynomial
    f(x) = 0
    is the x-axis.
  • The graph of a degree 0 polynomial
    f(x) = a0, where a0 ≠ 0,
    is a horizontal line with y-intercept a0
  • The graph of a degree 1 polynomial (or linear function)
    f(x) = a0 + a1x, where a1 ≠ 0,
    is an oblique line with y-intercept a0 and slope a1.
  • The graph of a degree 2 polynomial
    f(x) = a0 + a1x + a2x2, where a2 ≠ 0
    is a parabola.
  • The graph of a degree 3 polynomial
    f(x) = a0 + a1x + a2x2 + a3x3, where a3 ≠ 0
    is a cubic curve.
  • The graph of any polynomial with degree 2 or greater
    f(x) = a0 + a1x + a2x2 + ⋯ + anxn, where an ≠ 0 and n ≥ 2
    is a continuous non-linear curve.

A non-constant polynomial function tends to infinity when the variable increases indefinitely (in absolute value). If the degree is higher than one, the graph does not have any asymptote. It has two parabolic branches with vertical direction (one branch for positive x and one for negative x).

Polynomial graphs are analyzed in calculus using intercepts, slopes, concavity, and end behavior.

Equations

[edit]

A polynomial equation, also called an algebraic equation, is an equation of the form[27] For example, is a polynomial equation.

When considering equations, the indeterminates (variables) of polynomials are also called unknowns, and the solutions are the possible values of the unknowns for which the equality is true (in general more than one solution may exist). A polynomial equation stands in contrast to a polynomial identity like , where both expressions represent the same polynomial in different forms, and as a consequence any evaluation of both members gives a valid equality.

In elementary algebra, methods such as the quadratic formula are taught for solving all first degree and second degree polynomial equations in one variable. There are also formulas for the cubic and quartic equations. For higher degrees, the Abel–Ruffini theorem asserts that there can not exist a general formula in radicals. However, root-finding algorithms may be used to find numerical approximations of the roots of a polynomial expression of any degree.

The number of solutions of a polynomial equation with real coefficients may not exceed the degree, and equals the degree when the complex solutions are counted with their multiplicity. This fact is called the fundamental theorem of algebra.

Solving equations

[edit]

A root of a nonzero univariate polynomial P is a value a of x such that P(a) = 0. In other words, a root of P is a solution of the polynomial equation P(x) = 0 or a zero of the polynomial function defined by P. In the case of the zero polynomial, every number is a zero of the corresponding function, and the concept of root is rarely considered.

A number a is a root of a polynomial P if and only if the linear polynomial xa divides P, that is if there is another polynomial Q such that P = (xa) Q. It may happen that a power (greater than 1) of xa divides P; in this case, a is a multiple root of P, and otherwise a is a simple root of P. If P is a nonzero polynomial, there is a highest power m such that (xa)m divides P, which is called the multiplicity of a as a root of P. The number of roots of a nonzero polynomial P, counted with their respective multiplicities, cannot exceed the degree of P,[28] and equals this degree if all complex roots are considered (this is a consequence of the fundamental theorem of algebra). The coefficients of a polynomial and its roots are related by Vieta's formulas.

Some polynomials, such as x2 + 1, do not have any roots among the real numbers. If, however, the set of accepted solutions is expanded to the complex numbers, every non-constant polynomial has at least one root; this is the fundamental theorem of algebra. By successively dividing out factors xa, one sees that any polynomial with complex coefficients can be written as a constant (its leading coefficient) times a product of such polynomial factors of degree 1; as a consequence, the number of (complex) roots counted with their multiplicities is exactly equal to the degree of the polynomial.

There may be several meanings of "solving an equation". One may want to express the solutions as explicit numbers; for example, the unique solution of 2x − 1 = 0 is 1/2. This is, in general, impossible for equations of degree greater than one, and, since the ancient times, mathematicians have searched to express the solutions as algebraic expressions; for example, the golden ratio (1+5)/2 is the unique positive solution of x2x − 1 = 0 In the ancient times, they succeeded only for degrees one and two. For quadratic equations, the quadratic formula provides such expressions of the solutions. Since the 16th century, similar formulas (using cube roots in addition to square roots), although much more complicated, are known for equations of degree three and four (see cubic equation and quartic equation). But formulas for degree 5 and higher eluded researchers for several centuries. In 1824, Niels Henrik Abel proved the striking result that there are equations of degree 5 whose solutions cannot be expressed by a (finite) formula, involving only arithmetic operations and radicals (see Abel–Ruffini theorem). In 1830, Évariste Galois proved that most equations of degree higher than four cannot be solved by radicals, and showed that for each equation, one may decide whether it is solvable by radicals, and, if it is, solve it. This result marked the start of Galois theory and group theory, two important branches of modern algebra. Galois himself noted that the computations implied by his method were impracticable. Nevertheless, formulas for solvable equations of degrees 5 and 6 have been published (see quintic function and sextic equation).

When there is no algebraic expression for the roots, and when such an algebraic expression exists but is too complicated to be useful, the unique way of solving it is to compute numerical approximations of the solutions.[29] There are many methods for that; some are restricted to polynomials and others may apply to any continuous function. The most efficient algorithms allow solving easily (on a computer) polynomial equations of degree higher than 1,000 (see Root-finding algorithm).

For polynomials with more than one indeterminate, the combinations of values for the variables for which the polynomial function takes the value zero are generally called zeros instead of "roots". The study of the sets of zeros of polynomials is the object of algebraic geometry. For a set of polynomial equations with several unknowns, there are algorithms to decide whether they have a finite number of complex solutions, and, if this number is finite, for computing the solutions. See System of polynomial equations.

The special case where all the polynomials are of degree one is called a system of linear equations, for which another range of different solution methods exist, including the classical Gaussian elimination.

A polynomial equation for which one is interested only in the solutions which are integers is called a Diophantine equation. Solving Diophantine equations is generally a very hard task. It has been proved that there cannot be any general algorithm for solving them, or even for deciding whether the set of solutions is empty (see Hilbert's tenth problem). Some of the most famous problems that have been solved during the last fifty years are related to Diophantine equations, such as Fermat's Last Theorem.

Polynomial expressions

[edit]

Polynomials where indeterminates are substituted for some other mathematical objects are often considered, and sometimes have a special name.

Trigonometric polynomials

[edit]

A trigonometric polynomial is a finite linear combination of functions sin(nx) and cos(nx) with n taking on the values of one or more natural numbers.[30] The coefficients may be taken as real numbers, for real-valued functions.

If sin(nx) and cos(nx) are expanded in terms of sin(x) and cos(x), a trigonometric polynomial becomes a polynomial in the two variables sin(x) and cos(x) (using the multiple-angle formulae). Conversely, every polynomial in sin(x) and cos(x) may be converted, with Product-to-sum identities, into a linear combination of functions sin(nx) and cos(nx). This equivalence explains why linear combinations are called polynomials.

For complex coefficients, there is no difference between such a function and a finite Fourier series.

Trigonometric polynomials are widely used, for example in trigonometric interpolation applied to the interpolation of periodic functions. They are also used in the discrete Fourier transform.

Matrix polynomials

[edit]

A matrix polynomial is a polynomial with square matrices as variables.[31] Given an ordinary, scalar-valued polynomial this polynomial evaluated at a matrix A is where I is the identity matrix.[32]

A matrix polynomial equation is an equality between two matrix polynomials, which holds for the specific matrices in question. A matrix polynomial identity is a matrix polynomial equation which holds for all matrices A in a specified matrix ring Mn(R).

Exponential polynomials

[edit]

A bivariate polynomial where the second variable is substituted for an exponential function applied to the first variable, for example P(x, ex), may be called an exponential polynomial.

[edit]

Rational functions

[edit]

A rational fraction is the quotient (algebraic fraction) of two polynomials. Any algebraic expression that can be rewritten as a rational fraction is a rational function.

While polynomial functions are defined for all values of the variables, a rational function is defined only for the values of the variables for which the denominator is not zero.

The rational fractions include the Laurent polynomials, but do not limit denominators to powers of an indeterminate.

Laurent polynomials

[edit]

Laurent polynomials are like polynomials, but allow negative powers of the variable(s) to occur.

Power series

[edit]

Formal power series are like polynomials, but allow infinitely many non-zero terms to occur, so that they do not have finite degree. Unlike polynomials they cannot in general be explicitly and fully written down (just like irrational numbers cannot), but the rules for manipulating their terms are the same as for polynomials. Non-formal power series also generalize polynomials, but the multiplication of two power series may not converge.

Polynomial ring

[edit]

A polynomial f over a commutative ring R is a polynomial all of whose coefficients belong to R. It is straightforward to verify that the polynomials in a given set of indeterminates over R form a commutative ring, called the polynomial ring in these indeterminates, denoted in the univariate case and in the multivariate case.

One has So, most of the theory of the multivariate case can be reduced to an iterated univariate case.

The map from R to R[x] sending r to itself considered as a constant polynomial is an injective ring homomorphism, by which R is viewed as a subring of R[x]. In particular, R[x] is an algebra over R.

One can think of the ring R[x] as arising from R by adding one new element x to R, and extending in a minimal way to a ring in which x satisfies no other relations than the obligatory ones, plus commutation with all elements of R (that is xr = rx). To do this, one must add all powers of x and their linear combinations as well.

Formation of the polynomial ring, together with forming factor rings by factoring out ideals, are important tools for constructing new rings out of known ones. For instance, the ring (in fact field) of complex numbers, which can be constructed from the polynomial ring R[x] over the real numbers by factoring out the ideal of multiples of the polynomial x2 + 1. Another example is the construction of finite fields, which proceeds similarly, starting out with the field of integers modulo some prime number as the coefficient ring R (see modular arithmetic).

If R is commutative, then one can associate with every polynomial P in R[x] a polynomial function f with domain and range equal to R. (More generally, one can take domain and range to be any same unital associative algebra over R.) One obtains the value f(r) by substitution of the value r for the symbol x in P. One reason to distinguish between polynomials and polynomial functions is that, over some rings, different polynomials may give rise to the same polynomial function (see Fermat's little theorem for an example where R is the integers modulo p). This is not the case when R is the real or complex numbers, whence the two concepts are not always distinguished in analysis. An even more important reason to distinguish between polynomials and polynomial functions is that many operations on polynomials (like Euclidean division) require looking at what a polynomial is composed of as an expression rather than evaluating it at some constant value for x.

Divisibility

[edit]

If R is an integral domain and f and g are polynomials in R[x], it is said that f divides g or f is a divisor of g if there exists a polynomial q in R[x] such that f q = g. If then a is a root of f if and only divides f. In this case, the quotient can be computed using the polynomial long division.[33][34]

If F is a field and f and g are polynomials in F[x] with g ≠ 0, then there exist unique polynomials q and r in F[x] with and such that the degree of r is smaller than the degree of g (using the convention that the polynomial 0 has a negative degree). The polynomials q and r are uniquely determined by f and g. This is called Euclidean division, division with remainder or polynomial long division and shows that the ring F[x] is a Euclidean domain.

Analogously, prime polynomials (more correctly, irreducible polynomials) can be defined as non-zero polynomials which cannot be factorized into the product of two non-constant polynomials. In the case of coefficients in a ring, "non-constant" must be replaced by "non-constant or non-unit" (both definitions agree in the case of coefficients in a field). Any polynomial may be decomposed into the product of an invertible constant by a product of irreducible polynomials. If the coefficients belong to a field or a unique factorization domain this decomposition is unique up to the order of the factors and the multiplication of any non-unit factor by a unit (and division of the unit factor by the same unit). When the coefficients belong to integers, rational numbers or a finite field, there are algorithms to test irreducibility and to compute the factorization into irreducible polynomials (see Factorization of polynomials). These algorithms are not practicable for hand-written computation, but are available in any computer algebra system. Eisenstein's criterion can also be used in some cases to determine irreducibility.

Applications

[edit]

Positional notation

[edit]

In modern positional numbers systems, such as the decimal system, the digits and their positions in the representation of an integer, for example, 45, are a shorthand notation for a polynomial in the radix or base, in this case, 4 × 101 + 5 × 100. As another example, in radix 5, a string of digits such as 132 denotes the (decimal) number 1 × 52 + 3 × 51 + 2 × 50 = 42. This representation is unique. Let b be a positive integer greater than 1. Then every positive integer a can be expressed uniquely in the form

where m is a nonnegative integer and the r's are integers such that

0 < rm < b and 0 ≤ ri < b for i = 0, 1, . . . , m − 1.[35]

Interpolation and approximation

[edit]

The simple structure of polynomial functions makes them quite useful in analyzing general functions using polynomial approximations. An important example in calculus is Taylor's theorem, which roughly states that every differentiable function locally looks like a polynomial function, and the Stone–Weierstrass theorem, which states that every continuous function defined on a compact interval of the real axis can be approximated on the whole interval as closely as desired by a polynomial function. Practical methods of approximation include polynomial interpolation and the use of splines.[36]

Other applications

[edit]

Polynomials are frequently used to encode information about some other object. The characteristic polynomial of a matrix or linear operator contains information about the operator's eigenvalues. The minimal polynomial of an algebraic element records the simplest algebraic relation satisfied by that element. The chromatic polynomial of a graph counts the number of proper colourings of that graph.

The term "polynomial", as an adjective, can also be used for quantities or functions that can be written in polynomial form. For example, in computational complexity theory the phrase polynomial time means that the time it takes to complete an algorithm is bounded by a polynomial function of some variable, such as the size of the input.

History

[edit]

Determining the roots of polynomials, or "solving algebraic equations", is among the oldest problems in mathematics. However, the elegant and practical notation we use today only developed beginning in the 15th century. Before that, equations were written out in words. For example, an algebra problem from the Chinese Arithmetic in Nine Sections, c. 200 BCE, begins "Three sheafs of good crop, two sheafs of mediocre crop, and one sheaf of bad crop are sold for 29 dou." We would write 3x + 2y + z = 29.

History of the notation

[edit]

The earliest known use of the equal sign is in Robert Recorde's The Whetstone of Witte, 1557. The signs + for addition, − for subtraction, and the use of a letter for an unknown appear in Michael Stifel's Arithemetica integra, 1544. René Descartes, in La géometrie, 1637, introduced the concept of the graph of a polynomial equation. He popularized the use of letters from the beginning of the alphabet to denote constants and letters from the end of the alphabet to denote variables, as can be seen above, in the general formula for a polynomial in one variable, where the as denote constants and x denotes a variable. Descartes introduced the use of superscripts to denote exponents as well.[37]

See also

[edit]

Notes

[edit]

References

[edit]
[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
A polynomial is a mathematical expression formed by adding, subtracting, or multiplying constants () and variables raised to non-negative powers, resulting in a finite sum of terms such as anxn+an1xn1++a1x+a0a_n x^n + a_{n-1} x^{n-1} + \cdots + a_1 x + a_0, where the exponents are whole numbers and the coefficients are typically real or complex numbers. This structure ensures polynomials are well-defined functions over the real or complex numbers, with the ** defined as the highest exponent with a non-zero , determining its end behavior and number of possible . Polynomials extend naturally to multiple variables, such as in ax2y+bxy2+ca x^2 y + b x y^2 + c, where the total degree is the sum of exponents in each term. The study of polynomials dates back over 4,000 years to ancient Babylonian mathematicians around 2000 BCE, who developed methods to solve quadratic equations using geometric interpretations and tablet inscriptions for practical problems like land measurement. Progress accelerated during the , with Italian mathematicians like , , and discovering general solutions for cubic equations in the 1540s, followed by Lodovico Ferrari's work on quartics, as detailed in Cardano's 1545 publication Ars Magna. In the 19th century, provided the first rigorous proof of the in 1799, stating that every non-constant polynomial with complex coefficients has at least one complex root, implying it factors completely into linear terms over the complexes. This theorem, reproved multiple times since, underscores polynomials' in the . Key properties of polynomials include closure under and , forming polynomial rings, and the ability to perform operations like division, which yields a and via the . is central, allowing polynomials to be expressed as products of irreducibles, with found using formulas for low degrees (e.g., ) or numerical methods for higher ones, as Abel-Ruffini theorem proves no general algebraic solution exists for degrees five or above. Polynomials also approximate continuous functions via expansions around a point, enabling precise modeling in . Polynomials find extensive applications across mathematics and sciences, serving as foundational tools in algebra for solving equations and in calculus for differentiation and integration, where derivatives and integrals of polynomials remain polynomials of adjusted degrees. In physics and engineering, they model trajectories, electrical circuits, and signal processing through Fourier and polynomial approximations. Economics employs polynomial regression for forecasting trends in data like GDP growth, while combinatorics uses generating functions based on polynomials to count discrete structures. In computer science, polynomial-time algorithms distinguish efficient computations in complexity theory, and multivariate polynomials optimize problems in operations research.

Basic Concepts

Definition

In , a polynomial in one variable xx over a ring RR is formally defined as an expression of the form p(x)=anxn+an1xn1++a1x+a0p(x) = a_n x^n + a_{n-1} x^{n-1} + \cdots + a_1 x + a_0, where nn is a non-negative , the coefficients aia_i belong to RR for each i=0,1,,ni = 0, 1, \dots, n, and an0a_n \neq 0 if n>0n > 0. The coefficients aia_i are elements of the ring RR, which may be a field like numbers or a more general with unity; the degree of the polynomial is the highest power nn with a non-zero ana_n, while the constant term is a0a_0, the of x0x^0. This definition generalizes to multivariate polynomials over RR, which are finite sums of terms of the form ai1i2ikx1i1x2i2xkika_{i_1 i_2 \dots i_k} x_1^{i_1} x_2^{i_2} \dots x_k^{i_k}, where kk is the number of variables, the exponents iji_j are non-negative integers, and only finitely many coefficients are non-zero. For example, in two variables xx and yy, a polynomial might take the form p(x,y)=a20x2+a11xy+a02y2+a10x+a01y+a00p(x,y) = a_{20} x^2 + a_{11} x y + a_{02} y^2 + a_{10} x + a_{01} y + a_{00}. Common examples include constant polynomials, such as p(x)=5p(x) = 5, which have degree 0; linear polynomials, like p(x)=3x+2p(x) = 3x + 2, of degree 1; and quadratic polynomials, such as p(x)=x24x+7p(x) = x^2 - 4x + 7, of degree 2. Unlike , which allow infinitely many non-zero terms in the expansion i=0aixi\sum_{i=0}^\infty a_i x^i, polynomials are distinguished by their finite number of terms, ensuring they are well-defined as elements of the RR.

Notation and Terminology

Polynomials are commonly denoted using lowercase letters such as p(x)p(x), f(x)f(x), or g(x)g(x), where the argument xx represents the indeterminate or variable./05%3A_Polynomials_and_Their_Operations/5.02%3A_Introduction_to_Polynomials) Alternatively, uppercase letters like PP or QQ may be used for polynomials, especially in contexts emphasizing their formal structure. The coefficients of a polynomial are typically subscripted, such as aka_k or cic_i, to indicate the multiplier for each power of the variable; for instance, a general polynomial can be written as p(x)=anxn+an1xn1++a1x+a0p(x) = a_n x^n + a_{n-1} x^{n-1} + \cdots + a_1 x + a_0, where the aka_k are constants from the underlying field or ring./06%3A_Polynomial_Functions/6.02%3A_Zeros_of_Polynomials) In terminology, a is a single term of the form axka x^k, where aa is a nonzero and kk is a nonnegative exponent./2%3A_Polynomials/2.01%3A_The_Anatomy_of_a_Polynomial) A binomial consists of exactly two such terms, while a has precisely three terms; these are special cases of polynomials, which may have any finite number of terms greater than or equal to one./05%3A_Polynomial_and_Polynomial_Functions/5.02%3A_Add_and_Subtract_Polynomials) The zero polynomial, denoted as [0](/page/0)[0](/page/0), has all coefficients equal to zero and is the in the ; its degree is conventionally undefined, though some contexts assign it -\infty to preserve certain algebraic properties./A%3A_Appendices/12.4%3A_Polynomials) Standard conventions include identifying the leading as ana_n, the nonzero multiplier of the highest-degree term xnx^n, which determines the polynomial's degree nn. A constant polynomial has degree 0 and takes the form p(x)=cp(x) = c, where cc is a constant. of a polynomial at a point aa is denoted p(a)p(a), yielding the value obtained by substituting x=ax = a. For example, consider the polynomial 3x2+2x13x^2 + 2x - 1. Here, the terms are the monomials 3x23x^2, 2x2x, and 1-1; the leading is 3, the degree is 2, and at x=1x = 1 gives p(1)=3(1)2+2(1)1=4p(1) = 3(1)^2 + 2(1) - 1 = 4./05%3A_Polynomials_and_Their_Operations/5.02%3A_Introduction_to_Polynomials)

Classification

Degree and Leading Coefficient

In , the degree of a univariate polynomial p(x)=anxn+an1xn1++a1x+a0p(x) = a_n x^n + a_{n-1} x^{n-1} + \dots + a_1 x + a_0, where the aia_i are and an0a_n \neq 0, is defined as the highest exponent nn with a nonzero . The degree, often denoted deg(p)\deg(p), quantifies the polynomial's complexity and determines key behaviors such as the number of it can have. For the zero polynomial, where all are zero, the degree is undefined, as there is no highest power with a nonzero . The leading coefficient of a polynomial is the coefficient ana_n of its highest-degree term. This nonzero scalar influences the polynomial's end behavior; for instance, in the polynomial 2x3x2x^3 - x, the degree is 3 and the leading coefficient is 2, since the x3x^3 term has the highest power and its coefficient is nonzero. If a coefficient of a higher-degree term is zero, it does not contribute to the degree, effectively lowering it to the highest power with a nonzero coefficient; for example, 0x4+3x2+10 \cdot x^4 + 3x^2 + 1 has degree 2, not 4. A fundamental property of degrees arises in addition: for two polynomials pp and qq, deg(p+q)max(deg(p),deg(q))\deg(p + q) \leq \max(\deg(p), \deg(q)), with equality holding if the degrees differ or if the leading coefficients do not cancel upon . This inequality reflects how leading terms may combine or vanish, as in (3x2+x)+(3x2+2x)=3x(3x^2 + x) + (-3x^2 + 2x) = 3x, where the degree drops from 2 to 1 due to cancellation. For multivariate polynomials, such as p(x1,,xk)=ai1ikx1i1xkikp(x_1, \dots, x_k) = \sum a_{i_1 \dots i_k} x_1^{i_1} \dots x_k^{i_k}, the total degree is the maximum sum of exponents i1++iki_1 + \dots + i_k over all terms with nonzero ai1ika_{i_1 \dots i_k}. The leading coefficient in this context is the of a term achieving this maximum total degree, though multiple terms may share it in non-homogeneous cases. For example, in 2x2y+xy22x^2 y + x y^2, the total degree is 3, with leading terms 2x2y2x^2 y and xy2x y^2.

Univariate vs. Multivariate Polynomials

A univariate polynomial is a polynomial in a single indeterminate or variable, expressed as a finite sum of terms where each term consists of a multiplied by a non-negative power of that variable. For example, the polynomial p(x)=x2+3x+2p(x) = x^2 + 3x + 2 is univariate, with coefficients from a field such as the real numbers, and it exhibits a simpler linear along one , making it fundamental in introductory for tasks like root-finding and graphing. The support of a univariate polynomial is the set of exponents with non-zero coefficients, forming a finite of the non-negative integers. In contrast, a multivariate polynomial involves two or more indeterminates, with each term being a times a that is a product of powers of these variables. For instance, p(x,y)=x2y+3x+2p(x,y) = x^2 y + 3x + 2 is a bivariate polynomial, where monomials like x2yx^2 y have exponents distributed across variables. A special case is the , in which all monomials share the same total degree, defined as the sum of the exponents in each term; an example is x3+xyz+y2z+z3x^3 + xyz + y^2 z + z^3, where every term has total degree 3. The support here consists of multi-indices—tuples of non-negative integers representing the exponents for each variable—with non-zero . Key structural differences arise in evaluation, degrees, and representation. To evaluate a univariate polynomial, one substitutes a single value for the variable, yielding a scalar; multivariate evaluation requires assigning values to each , producing a result in higher-dimensional space. While univariate polynomials have a single degree—the highest exponent—multivariate polynomials feature a total degree as the maximum sum of exponents across all monomials, alongside partial degrees for individual variables (as referenced in the discussion of degree in multivariate contexts). The support's dimensionality also differs: univariate supports are one-dimensional subsets of integers, whereas multivariate supports are subsets of Nk\mathbb{N}^k for kk variables. An illustrative example of multivariate polynomials is the bivariate quadratic form, such as ax2+bxy+cy2+dx+ey+fax^2 + bxy + cy^2 + dx + ey + f, which defines conic sections like ellipses or hyperbolas when set to zero in the plane. The zero set of a non-constant multivariate polynomial, known as a , forms a codimension-one in the ambient space, capturing geometric structures beyond simple curves.

Operations

Addition and Subtraction

Addition and subtraction of polynomials are fundamental operations performed by combining like terms, which are terms sharing the same power of the variable. To add two polynomials, align them by powers of the variable (typically xx) and add the coefficients of corresponding terms. For instance, consider the polynomials p(x)=2x2+3x1p(x) = 2x^2 + 3x - 1 and q(x)=x2+4x+5q(x) = -x^2 + 4x + 5; adding them yields p(x)+q(x)=(2x2x2)+(3x+4x)+(1+5)=x2+7x+4p(x) + q(x) = (2x^2 - x^2) + (3x + 4x) + (-1 + 5) = x^2 + 7x + 4. Formally, if p(x)=k=0nakxkp(x) = \sum_{k=0}^n a_k x^k and q(x)=k=0mbkxkq(x) = \sum_{k=0}^m b_k x^k, their sum is given by p(x)+q(x)=k=0max(n,m)(ak+bk)xk,p(x) + q(x) = \sum_{k=0}^{\max(n,m)} (a_k + b_k) x^k, where coefficients are zero if the index exceeds the polynomial's degree. Subtraction follows similarly: p(x)q(x)=p(x)+(q(x))p(x) - q(x) = p(x) + (-q(x)), where q(x)=k=0m(bk)xk-q(x) = \sum_{k=0}^m (-b_k) x^k, distributing the negative sign to each coefficient before combining like terms. These operations inherit properties from the underlying ring structure of polynomials over a field (such as the real or complex numbers); together with , they form a . Addition is commutative, meaning p(x)+q(x)=q(x)+p(x)p(x) + q(x) = q(x) + p(x), and associative, so (p(x)+q(x))+r(x)=p(x)+(q(x)+r(x))(p(x) + q(x)) + r(x) = p(x) + (q(x) + r(x)) for any polynomials p(x)p(x), q(x)q(x), and r(x)r(x). The zero polynomial, 0(x)=00(x) = 0, serves as the , satisfying p(x)+0(x)=p(x)p(x) + 0(x) = p(x). The degree of the sum or difference is the maximum of the degrees of the addends unless the leading coefficients cancel, in which case the degree may be lower. For example, adding x2+1x^2 + 1 and x2+2x-x^2 + 2x results in 2x+12x + 1, reducing the degree from 2 to 1 due to cancellation. Adding the zero polynomial preserves the original degree and polynomial unchanged.

Multiplication

Polynomial multiplication is performed by applying the to each term of one polynomial across every term of the second polynomial, followed by combining according to the rules of . For two polynomials f(x)=k=0dakxkf(x) = \sum_{k=0}^{d} a_k x^k and g(x)=m=0ebmxmg(x) = \sum_{m=0}^{e} b_m x^m, the product is obtained by multiplying each akxka_k x^k by each bmxmb_m x^m, resulting in terms akbmxk+ma_k b_m x^{k+m}, and then grouping terms with the same exponent. The general formula for the product is (f(x))(g(x))=n=0d+ecnxn,(f(x))(g(x)) = \sum_{n=0}^{d+e} c_n x^n, where cn=k+m=nakbm.c_n = \sum_{k+m=n} a_k b_m. This of coefficients ensures the result is a polynomial of degree at most d+ed + e. If the leading coefficients ada_d and beb_e are nonzero, the degree of the product is exactly d+ed + e, as the highest-degree term adbexd+ea_d b_e x^{d+e} does not cancel. Multiplication of polynomials inherits key from the ring structure: it is commutative, meaning f(x)g(x)=g(x)f(x)f(x) g(x) = g(x) f(x); associative, so (f(x)g(x))h(x)=f(x)(g(x)h(x))(f(x) g(x)) h(x) = f(x) (g(x) h(x)); and distributive over , f(x)(g(x)+h(x))=f(x)g(x)+f(x)h(x)f(x) (g(x) + h(x)) = f(x) g(x) + f(x) h(x). These allow polynomials to form a ring under and . A special case is the multiplication of a by a polynomial, which simplifies to distributing the 's and adding its exponent to each term's exponent in the polynomial. For example, 2x(5x3+4)=10x4+8x2x (5x^3 + 4) = 10x^4 + 8x. For binomials, the process yields the difference of squares in certain cases, such as (x+1)(x1)=x21(x + 1)(x - 1) = x^2 - 1, obtained by distributing: xx+x(1)+1x+1(1)=x2x+x1=x21x \cdot x + x \cdot (-1) + 1 \cdot x + 1 \cdot (-1) = x^2 - x + x - 1 = x^2 - 1. Another example is (3x+1)(2x5)=6x215x+2x5=6x213x5(3x + 1)(2x - 5) = 6x^2 - 15x + 2x - 5 = 6x^2 - 13x - 5, where are combined after distribution.

Division and Remainder Theorem

Polynomial division allows for the decomposition of one polynomial into a and when divided by another non-zero polynomial. The process, known as , involves arranging both polynomials in descending order of degrees, dividing the leading term of the by the leading term of the to obtain the first term of the , multiplying this term by the entire , subtracting the result from the , and repeating with the new until the degree of the is less than the degree of the . The division algorithm formalizes this procedure: for polynomials f(T)f(T) and g(T)g(T) in F[T]F[T] over a field FF with g(T)0g(T) \neq 0, there exist unique polynomials q(T)q(T) and r(T)r(T) such that f(T)=g(T)q(T)+r(T)f(T) = g(T) q(T) + r(T) and either r(T)=0r(T) = 0 or degr(T)<degg(T)\deg r(T) < \deg g(T). This uniqueness follows from degree considerations: if two such decompositions exist, their difference implies a contradiction in degrees unless the quotients and remainders match. Existence is established by induction on the degree of f(T)f(T), reducing the degree at each step by subtracting a suitable multiple of g(T)g(T). A key consequence is the remainder theorem, which states that when a polynomial p(x)p(x) is divided by xax - a, the remainder is p(a)p(a). This follows directly from the division algorithm by evaluating at x=ax = a, yielding p(a)=q(a)0+r(a)p(a) = q(a) \cdot 0 + r(a), so r(a)=p(a)r(a) = p(a) since r(x)r(x) is constant in this case. For example, dividing x31x^3 - 1 by x1x - 1 using long division: the leading terms give quotient term x2x^2, multiplying yields x3x2x^3 - x^2, subtracting from x31x^3 - 1 gives x21x^2 - 1, then next term xx, multiplying gives x2xx^2 - x, subtracting yields x1x - 1, then term 1, multiplying gives x1x - 1, subtracting yields 0. Thus, the quotient is x2+x+1x^2 + x + 1 and remainder 0, consistent with the remainder theorem since p(1)=131=0p(1) = 1^3 - 1 = 0. For linear divisors xcx - c, synthetic division provides a streamlined method using only coefficients. To divide p(x)=anxn++a0p(x) = a_n x^n + \cdots + a_0 by xcx - c, list coefficients an,,a0a_n, \dots, a_0 and cc; bring down ana_n, multiply by cc and add to next coefficient, repeating across the row. The final addend is the remainder, and prior values form quotient coefficients. For instance, dividing 5x3x2+65x^3 - x^2 + 6 by x4x - 4 (coefficients 5, -1, 0, 6; note implicit 0 for xx): bring down 5, multiply by 4 to get 20, add to -1 yields 19; multiply 19 by 4 gets 76, add to 0 yields 76; multiply 76 by 4 gets 304, add to 6 yields 310. Quotient is 5x2+19x+765x^2 + 19x + 76, remainder 310, matching p(4)=310p(4) = 310.

Advanced Operations

Composition

In mathematics, the composition of polynomials is a fundamental operation that combines two polynomials by substituting the output of one as the input to the other. For polynomials ff and gg with coefficients in a field such as the real or complex numbers, the composition fgf \circ g is defined by (fg)(x)=f(g(x))(f \circ g)(x) = f(g(x)). This operation produces another polynomial, as substituting a polynomial into another preserves the polynomial structure./05:_Further_Topics_in_Functions/5.02:_Function_Composition) A key property of polynomial composition is the multiplicative rule for degrees: if deg(g)1\deg(g) \geq 1, then deg(fg)=deg(f)deg(g)\deg(f \circ g) = \deg(f) \cdot \deg(g). This follows from the leading term analysis, where the highest-degree term in f(g(x))f(g(x)) arises from the leading term of ff applied to the leading term of gg, yielding exponents that multiply. For instance, if f(x)=amxm+f(x) = a_m x^m + \cdots with m=deg(f)m = \deg(f) and g(x)=bnxn+g(x) = b_n x^n + \cdots with n=deg(g)n = \deg(g), the leading term of f(g(x))f(g(x)) is am(bnxn)m=ambnmxmna_m (b_n x^n)^m = a_m b_n^m x^{m n}. If deg(g)=0\deg(g) = 0, then fgf \circ g has the same degree as ff. Composition of polynomials is associative, meaning (fg)h=f(gh)(f \circ g) \circ h = f \circ (g \circ h) for any compatible polynomials ff, gg, and hh, but it is generally not commutative, as fggff \circ g \neq g \circ f unless the polynomials satisfy specific symmetry conditions. To illustrate, consider f(x)=x2f(x) = x^2 and g(x)=x+1g(x) = x + 1: then f(g(x))=(x+1)2=x2+2x+1f(g(x)) = (x + 1)^2 = x^2 + 2x + 1, which has degree 21=22 \cdot 1 = 2, while g(f(x))=x2+1g(f(x)) = x^2 + 1, which has degree 2 but differs in form. Expanding the result of composition often relies on multiplication of polynomials to distribute terms./05:_Further_Topics_in_Functions/5.02:_Function_Composition) Repeated composition, or iteration, arises naturally in studying dynamical systems defined by polynomials. The nn-fold composition f(n)f^{(n)} is defined recursively as f(1)=ff^{(1)} = f and f(n)=ff(n1)f^{(n)} = f \circ f^{(n-1)} for n2n \geq 2, with degree deg(f(n))=[deg(f)]n\deg(f^{(n)}) = [\deg(f)]^n. Fixed points of a polynomial ff, which are solutions to f(x)=xf(x) = x, play a central role in analyzing the long-term behavior of iterations, as they represent equilibria under repeated application. For example, the fixed points of f(x)=x22f(x) = x^2 - 2 are 2 and -1, influencing the convergence or divergence of orbits under iteration. In the multivariate setting, composition extends to substituting vector-valued polynomials, such as (f(g,h))(x,y)=f(g(x,y),h(x,y))(f \circ (g, h))(x, y) = f(g(x, y), h(x, y)), but introduces additional complexities. The resulting degree remains the product of the individual degrees under appropriate conditions, yet the operation can explode in complexity due to the need to handle multiple substitutions and expansions, leading to high-dimensional terms that challenge efficient computation and symbolic manipulation.

Factoring

Factoring a polynomial involves expressing it as a product of simpler polynomials, typically over a specified field such as the rationals Q\mathbb{Q} or reals R\mathbb{R}. This process decomposes the polynomial into factors that cannot be further simplified within the field, aiding in solving equations, integration, and understanding polynomial structure. Over fields, polynomials admit unique factorization into irreducibles, analogous to the fundamental theorem of arithmetic for integers. The factor theorem states that for a polynomial p(x)p(x) over a field FF, a linear factor (xa)(x - a) with aFa \in F divides p(x)p(x) if and only if p(a)=0p(a) = 0; this is the converse of the remainder theorem from the division algorithm. To apply it, one evaluates p(x)p(x) at potential roots to identify linear factors, then uses polynomial division or synthetic division to extract them. Common methods for factoring include the rational root theorem, which identifies possible rational roots of polynomials with integer coefficients. For a polynomial p(x)=anxn++a0p(x) = a_n x^n + \cdots + a_0 where aiZa_i \in \mathbb{Z}, any rational root r/sr/s in lowest terms has rr dividing a0a_0 and ss dividing ana_n. Testing these candidates via the factor theorem allows extraction of linear factors. Another approach is factoring by grouping, where terms are paired to factor out common binomials, as in x3+x2+2x+2=(x3+x2)+(2x+2)=x2(x+1)+2(x+1)=(x2+2)(x+1)x^3 + x^2 + 2x + 2 = (x^3 + x^2) + (2x + 2) = x^2(x + 1) + 2(x + 1) = (x^2 + 2)(x + 1). Special forms like the difference of squares, x2a2=(xa)(x+a)x^2 - a^2 = (x - a)(x + a), enable direct factorization for binomials. Over the reals R\mathbb{R}, irreducible polynomials are precisely the linear polynomials and quadratic polynomials with negative discriminant (no real roots), as higher-degree polynomials factor into these by the fundamental theorem of algebra. In polynomial rings FF over any field FF, every nonconstant polynomial factors uniquely (up to units and order) into irreducible polynomials, forming a unique factorization domain. For example, the quadratic x25x+6x^2 - 5x + 6 factors over Q\mathbb{Q} as (x2)(x3)(x - 2)(x - 3), verified by roots 2 and 3 via the factor theorem. Cyclotomic polynomials, such as Φ5(x)=x4+x3+x2+x+1\Phi_5(x) = x^4 + x^3 + x^2 + x + 1, are irreducible over Q\mathbb{Q} and arise in factoring xn1=dnΦd(x)x^n - 1 = \prod_{d|n} \Phi_d(x). Factoring also sets up partial fraction decomposition for rational functions p(x)/q(x)p(x)/q(x) where deg(p)<deg(q)\deg(p) < \deg(q). After factoring q(x)q(x) into irreducibles over Q\mathbb{Q}, one writes p(x)q(x)=Aili(x)+Bjx+Cjqj(x)\frac{p(x)}{q(x)} = \sum \frac{A_i}{l_i(x)} + \sum \frac{B_j x + C_j}{q_j(x)}, with linear denominators for linear factors and quadratic for irreducibles, solving for coefficients without performing the full integration.

Differentiation and Integration

Differentiation of a polynomial p(x)=k=0nakxkp(x) = \sum_{k=0}^n a_k x^k yields p(x)=k=1nkakxk1p'(x) = \sum_{k=1}^n k a_k x^{k-1}, resulting in a polynomial of degree n1n-1 unless p(x)p(x) is a non-zero constant, in which case the derivative is zero./03%3A_Derivatives/3.03%3A_Differentiation_Rules) This operation follows from the power rule for differentiation, applied term by term, combined with the linearity of the derivative./03%3A_Derivatives/3.03%3A_Differentiation_Rules) For example, the derivative of x3+2xx^3 + 2x is 3x2+23x^2 + 2./03%3A_Derivatives/3.03%3A_Differentiation_Rules) Higher-order derivatives of a polynomial of degree nn reduce the degree successively until the (n+1)(n+1)-th derivative, which is identically zero. This property arises because each differentiation lowers the degree by one, eventually eliminating all terms. For products of polynomials, the product rule states that if p(x)p(x) and q(x)q(x) are polynomials, then (pq)(x)=p(x)q(x)+p(x)q(x)(p q)'(x) = p'(x) q(x) + p(x) q'(x), producing another polynomial./03%3A_Derivatives/3.03%3A_Differentiation_Rules) In the case of composition, where r(x)=p(q(x))r(x) = p(q(x)), the chain rule gives r(x)=p(q(x))q(x)r'(x) = p'(q(x)) q'(x), again yielding a polynomial derivative./03%3A_Derivatives/3.05%3A_The_Chain_Rule) The indefinite integral of a polynomial p(x)=k=0nakxkp(x) = \sum_{k=0}^n a_k x^k is p(x)dx=k=0nakk+1xk+1+C\int p(x) \, dx = \sum_{k=0}^n \frac{a_k}{k+1} x^{k+1} + C, which is a polynomial of degree n+1n+1 plus the constant of integration. This follows from integrating each term using the power rule for integration. For instance, 3x2dx=x3+C\int 3x^2 \, dx = x^3 + C. Taylor polynomials offer local approximations to functions near a point using the derivatives of the polynomial at that point, though polynomials themselves are exactly represented by their Taylor expansions up to their degree./08%3A_Sequences_and_Series/8.07%3A_Taylor_Polynomials)

Polynomial Functions

Graphing and Behavior

The graph of a polynomial function y=p(x)y = p(x) of degree nn is always continuous and smooth, meaning it can be drawn without lifting the pencil and has no sharp corners, breaks, or cusps. This smoothness arises because polynomials are infinitely differentiable everywhere in their domain, which is all real numbers. Additionally, the graph can have at most n1n - 1 turning points, where the function changes from increasing to decreasing or vice versa, limiting the number of local maxima and minima. The end behavior of the graph as x±x \to \pm \infty is determined by the degree nn and the sign of the leading coefficient ana_n. For even-degree polynomials (nn even), the ends point in the same direction: both up if an>0a_n > 0, or both down if an<0a_n < 0. For odd-degree polynomials (nn odd), the ends point in opposite directions: left down and right up if an>0a_n > 0, or left up and right down if an<0a_n < 0. This behavior reflects the dominance of the leading term anxna_n x^n for large x|x|. Key features of the graph include the y-intercept at p(0)p(0) and x-intercepts at the real roots of p(x)=0p(x) = 0, which indicate where the graph crosses the axes. Unlike rational functions, polynomial graphs have no vertical or horizontal asymptotes, as they remain defined and finite for all real xx. For example, the cubic polynomial p(x)=x3xp(x) = x^3 - x has degree 3 and leading coefficient 1 (positive), so its end behavior shows the graph falling as xx \to -\infty and rising as xx \to \infty; it crosses the x-axis at three points and exhibits one local maximum and one local minimum, consistent with at most 2 turning points for a cubic. Similarly, quadratic polynomials like p(x)=x2p(x) = x^2 form a smooth parabola opening upward, with a single turning point at the vertex, illustrating the even-degree pattern where both ends rise.

Roots and Zeros

In mathematics, a root (or zero) of a polynomial p(x)p(x) is a value rr such that p(r)=0p(r) = 0. This means substituting rr into the polynomial yields zero, indicating that xrx - r is a factor of p(x)p(x). The multiplicity of a root rr is the largest positive integer mm such that (xr)m(x - r)^m divides p(x)p(x) evenly, but (xr)m+1(x - r)^{m+1} does not. Roots with multiplicity greater than 1 are called multiple roots; for instance, a root of multiplicity 2 touches the x-axis at that point without crossing it in the graph of the polynomial function. A simple root has multiplicity 1. Consider the polynomial p(x)=(x1)2(x+2)=x33x+2p(x) = (x - 1)^2 (x + 2) = x^3 - 3x + 2. Here, x=1x = 1 is a root of multiplicity 2, and x=2x = -2 is a root of multiplicity 1, as verified by the factorizations and direct substitution. Descartes' rule of signs provides bounds on the number of positive real roots of a polynomial with real coefficients. The number of positive real roots is either equal to the number of sign changes in the coefficients of p(x)p(x) or less than that by an even integer. For negative real roots, apply the rule to p(x)p(-x). For example, in p(x)=x33x+2p(x) = x^3 - 3x + 2, there are two sign changes, so at most two positive real roots are possible (though here there is one distinct positive real root of multiplicity two). For polynomials with real coefficients, non-real complex roots occur in conjugate pairs. That is, if a+bia + bi (with b0b \neq 0) is a root, then its complex conjugate abia - bi is also a root. This follows from the fact that if a complex number satisfies the equation, its conjugate satisfies the equation with conjugated coefficients, which are real and thus unchanged. Vieta's formulas relate the roots of a polynomial to its coefficients. For a polynomial p(x)=anxn+an1xn1++a1x+a0=an(xr1)(xr2)(xrn)p(x) = a_n x^n + a_{n-1} x^{n-1} + \cdots + a_1 x + a_0 = a_n (x - r_1)(x - r_2) \cdots (x - r_n) with roots r1,r2,,rnr_1, r_2, \dots, r_n (counting multiplicities), the sum of the roots r1+r2++rn=an1/anr_1 + r_2 + \cdots + r_n = -a_{n-1}/a_n, the sum of the products of roots taken two at a time equals an2/ana_{n-2}/a_n, and so on, alternating signs up to the product of all roots (1)na0/an(-1)^n a_0 / a_n. These symmetric relations hold over the complex numbers.

Polynomial Equations

Solving Methods

Solving polynomial equations of the form p(x)=0p(x) = 0 relies on a range of analytical and numerical techniques, depending on the degree of the polynomial and the desired precision. For low-degree polynomials, closed-form solutions using radicals exist, while higher-degree cases often require approximation methods due to theoretical limitations. For quadratic equations ax2+bx+c=0ax^2 + bx + c = 0 with a0a \neq 0, the quadratic formula yields the roots explicitly: x=b±b24ac2ax = \frac{ -b \pm \sqrt{b^2 - 4ac} }{ 2a }
Add your contribution
Related Hubs
User Avatar
No comments yet.