Geometrical Properties Of Polynomial Roots
   HOME

TheInfoList



OR:

In
mathematics Mathematics is a field of study that discovers and organizes methods, Mathematical theory, theories and theorems that are developed and Mathematical proof, proved for the needs of empirical sciences and mathematics itself. There are many ar ...
, a
univariate polynomial In mathematics, a polynomial is a mathematical expression consisting of indeterminates (also called variables) and coefficients, that involves only the operations of addition, subtraction, multiplication and exponentiation to nonnegative integer ...
of degree with real or complex coefficients has complex ''
roots A root is the part of a plant, generally underground, that anchors the plant body, and absorbs and stores water and nutrients. Root or roots may also refer to: Art, entertainment, and media * ''The Root'' (magazine), an online magazine focusin ...
'' (if counted with their
multiplicities In mathematics, the multiplicity of a member of a multiset is the number of times it appears in the multiset. For example, the number of times a given polynomial has a root at a given point is the multiplicity of that root. The notion of multipl ...
). They form a
multiset In mathematics, a multiset (or bag, or mset) is a modification of the concept of a set that, unlike a set, allows for multiple instances for each of its elements. The number of instances given for each element is called the ''multiplicity'' of ...
of points in the
complex plane In mathematics, the complex plane is the plane (geometry), plane formed by the complex numbers, with a Cartesian coordinate system such that the horizontal -axis, called the real axis, is formed by the real numbers, and the vertical -axis, call ...
, whose
geometry Geometry (; ) is a branch of mathematics concerned with properties of space such as the distance, shape, size, and relative position of figures. Geometry is, along with arithmetic, one of the oldest branches of mathematics. A mathematician w ...
can be deduced from the degree and the coefficients of the polynomial. Some of these geometrical properties are related to a single polynomial, such as upper bounds on the absolute values of the roots, which define a disk containing all roots, or lower bounds on the distance between two roots. Such bounds are widely used for
root-finding algorithm In numerical analysis, a root-finding algorithm is an algorithm for finding zeros, also called "roots", of continuous functions. A zero of a function is a number such that . As, generally, the zeros of a function cannot be computed exactly nor ...
s for polynomials, either for tuning them, or for computing their
computational complexity In computer science, the computational complexity or simply complexity of an algorithm is the amount of resources required to run it. Particular focus is given to computation time (generally measured by the number of needed elementary operations ...
. Some other properties are probabilistic, such as the expected number of real roots of a random polynomial of degree with real coefficients, which is less than 1+\frac 2\pi \ln (n) for sufficiently large.


Notation

In this article, a polynomial is always denoted : p(x)=a_0 + a_1 x + \cdots + a_n x^n, where a_0, \dots, a_n are real or
complex number In mathematics, a complex number is an element of a number system that extends the real numbers with a specific element denoted , called the imaginary unit and satisfying the equation i^= -1; every complex number can be expressed in the for ...
s and a_n \neq 0; thus n is the degree of the polynomial.


Continuous dependence on coefficients

The roots of a polynomial of degree depend continuously on the coefficients. For simple roots, this results immediately from the
implicit function theorem In multivariable calculus, the implicit function theorem is a tool that allows relations to be converted to functions of several real variables. It does so by representing the relation as the graph of a function. There may not be a single functi ...
. This is true also for multiple roots, but some care is needed for the proof. A small change of coefficients may induce a dramatic change of the roots, including the change of a real root into a complex root with a rather large imaginary part (see
Wilkinson's polynomial In numerical analysis, Wilkinson's polynomial is a specific polynomial which was used by James H. Wilkinson in 1963 to illustrate a difficulty when finding the roots of a polynomial: the location of the roots can be very sensitive to perturbatio ...
). A consequence is that, for classical numeric
root-finding algorithm In numerical analysis, a root-finding algorithm is an algorithm for finding zeros, also called "roots", of continuous functions. A zero of a function is a number such that . As, generally, the zeros of a function cannot be computed exactly nor ...
s, the problem of approximating the roots given the coefficients can be
ill-conditioned In numerical analysis, the condition number of a function measures how much the output value of the function can change for a small change in the input argument. This is used to measure how sensitive a function is to changes or errors in the inpu ...
for many inputs.


Conjugation

The
complex conjugate root theorem In mathematics, the complex conjugate root theorem states that if ''P'' is a polynomial in one variable with real coefficients, and ''a'' + ''bi'' is a root of ''P'' with ''a'' and ''b'' being real numbers, then its complex conjugate ''a ...
states that if the coefficients of a polynomial are real, then the non-real roots appear in pairs of the form . It follows that the roots of a polynomial with real coefficients are mirror-symmetric with respect to the real axis. This can be extended to algebraic conjugation: the roots of a polynomial with
rational Rationality is the quality of being guided by or based on reason. In this regard, a person acts rationally if they have a good reason for what they do, or a belief is rational if it is based on strong evidence. This quality can apply to an ...
coefficients are ''conjugate'' (that is, invariant) under the action of the
Galois group In mathematics, in the area of abstract algebra known as Galois theory, the Galois group of a certain type of field extension is a specific group associated with the field extension. The study of field extensions and their relationship to the pol ...
of the polynomial. However, this symmetry can rarely be interpreted geometrically.


Bounds on all roots

Upper bounds on the absolute values of polynomial roots are widely used for
root-finding algorithm In numerical analysis, a root-finding algorithm is an algorithm for finding zeros, also called "roots", of continuous functions. A zero of a function is a number such that . As, generally, the zeros of a function cannot be computed exactly nor ...
s, either for limiting the regions where roots should be searched, or for the computation of the
computational complexity In computer science, the computational complexity or simply complexity of an algorithm is the amount of resources required to run it. Particular focus is given to computation time (generally measured by the number of needed elementary operations ...
of these algorithms. Many such bounds have been given, and the sharper one depends generally on the specific sequence of coefficient that are considered. Most bounds are greater or equal to one, and are thus not sharp for a polynomial which have only roots of absolute values lower than one. However, such polynomials are very rare, as shown below. Any upper bound on the absolute values of roots provides a corresponding lower bound. In fact, if a_n\ne 0, and is an upper bound of the absolute values of the roots of : a_0 + a_1 x + \cdots + a_n x^n, then is a lower bound of the absolute values of the roots of : a_n + a_ x + \cdots + a_0 x^n, since the roots of either polynomial are the multiplicative inverses of the roots of the other. ''Therefore, in the remainder of the article lower bounds will not be given explicitly''.


Lagrange's and Cauchy's bounds

Lagrange Joseph-Louis Lagrange (born Giuseppe Luigi LagrangiaCauchy Baron Augustin-Louis Cauchy ( , , ; ; 21 August 1789 – 23 May 1857) was a French mathematician, engineer, and physicist. He was one of the first to rigorously state and prove the key theorems of calculus (thereby creating real a ...
were the first to provide upper bounds on all complex roots. Lagrange's bound is :\max\left\, and Cauchy's bound isCauchy Augustin-Louis (1829). ''Exercices de mathématique''. Œuvres 2 (9) p.122 :1+\max\left\ Lagrange's bound is sharper (smaller) than Cauchy's bound only when 1 is larger than the sum of all \left, \frac\ but the largest. This is relatively rare in practice, and explains why Cauchy's bound is more widely used than Lagrange's. Both bounds result from the
Gershgorin circle theorem In mathematics, the Gershgorin circle theorem may be used to bound the spectrum of a square matrix. It was first published by the Soviet mathematician Semyon Aronovich Gershgorin in 1931. Gershgorin's name has been transliterated in several diffe ...
applied to the
companion matrix In linear algebra, the Frobenius companion matrix of the monic polynomial p(x)=c_0 + c_1 x + \cdots + c_x^ + x^n is the square matrix defined as C(p)=\begin 0 & 0 & \dots & 0 & -c_0 \\ 1 & 0 & \dots & 0 & -c_1 \\ 0 & 1 & \dots & 0 & -c_2 \\ \ ...
of the polynomial and its
transpose In linear algebra, the transpose of a Matrix (mathematics), matrix is an operator which flips a matrix over its diagonal; that is, it switches the row and column indices of the matrix by producing another matrix, often denoted by (among other ...
. They can also be proved by elementary methods. If is a root of the polynomial, and one has :, a_n, , z^n, = \left, \sum_^a_iz^i\ \le \sum_^, a_iz^i, \le \sum_^, a_i, , z, ^. Dividing by , a_n, , z, ^, one gets :, z, \le \sum_^\frac, which is Lagrange's bound when there is at least one root of absolute value larger than 1. Otherwise, 1 is a bound on the roots, and is not larger than Lagrange's bound. Similarly, for Cauchy's bound, one has, if , :, a_n, , z^n, = \left, \sum_^a_iz^i\ \le \sum_^, a_iz^i, \le \max , a_i, \sum_^, z, ^i = \frac\max , a_i, \le \frac\max , a_i, . Thus : , a_n, (, z, -1) \le \max , a_i, . Solving in , one gets Cauchy's bound if there is a root of absolute value larger than 1. Otherwise the bound is also correct, as Cauchy's bound is larger than 1. These bounds are not invariant by scaling. That is, the roots of the polynomial are the quotient by of the root of , and the bounds given for the roots of are not the quotient by of the bounds of . Thus, one may get sharper bounds by minimizing over possible scalings. This gives \min_\left(\max\left\\right), and \min_\left(s+\max_\left(\left, \frac\ s^\right)\right), for Lagrange's and Cauchy's bounds respectively. Another bound, originally given by Lagrange, but attributed to Zassenhaus by
Donald Knuth Donald Ervin Knuth ( ; born January 10, 1938) is an American computer scientist and mathematician. He is a professor emeritus at Stanford University. He is the 1974 recipient of the ACM Turing Award, informally considered the Nobel Prize of comp ...
, is 2\max\left\. This bound is invariant by scaling. Let be the largest \left, \frac\^\frac 1 for . Thus one has :\frac \le A^ for 0\le i If is a root of , one has :-a_nz^n=\sum_^a_i z^i, and thus, after dividing by a_n, :\begin , z, ^n&\le\sum_^A^, z, ^i\\ &=A\frac. \end As we want to prove , we may suppose that (otherwise there is nothing to prove). Thus :, z, ^n \le A\frac, which gives the result, since , z, >A. Lagrange improved this latter bound into the sum of the two largest values (possibly equal) in the sequence \left \frac\, \left, \frac\^, \ldots, \left, \frac\^\right Lagrange also provided the bound \sum_i \left, \frac\, where a_i denotes the th ''nonzero'' coefficient when the terms of the polynomials are sorted by increasing degrees.


Using Hölder's inequality

Hölder's inequality In mathematical analysis, Hölder's inequality, named after Otto Hölder, is a fundamental inequality (mathematics), inequality between Lebesgue integration, integrals and an indispensable tool for the study of Lp space, spaces. The numbers an ...
allows the extension of Lagrange's and Cauchy's bounds to every -norm. The -norm of a sequence :s=(a_0, \ldots, a_n) is :\, s\, _h = \left(\sum_^n , a_i, ^h\right)^, for any real number , and :\, s\, _\infty = \textstyle , a_i, . If \frac 1h+ \frac 1k=1, with , and , an upper bound on the absolute values of the roots of is \frac 1\left\, (, a_n, , \left\, (, a_, , \ldots, , a_0, \right)\, _h\right)\, _k. For and , one gets respectively Cauchy's and Lagrange's bounds. For , one has the bound \frac 1\sqrt. This is not only a bound of the absolute values of the roots, but also a bound of the product of their absolute values larger than 1; see , below. Let be a root of the polynomial :p(x)=a_nx^n+a_x^+\cdots+a_1x +a_0. Setting :A=\left(\frac,\ldots, \frac,\frac\right), we have to prove that every root of satisfies :z\le \left\, (1, \left\, A\right\, _h\right)\, _k. If , z, \le 1, the inequality is true; so, one may suppose , z, > 1 for the remainder of the proof. Writing the equation as :-z^n=\fracz^+\cdots+\fracz+\frac,
Hölder's inequality In mathematical analysis, Hölder's inequality, named after Otto Hölder, is a fundamental inequality (mathematics), inequality between Lebesgue integration, integrals and an indispensable tool for the study of Lp space, spaces. The numbers an ...
implies :, z, ^n\leq \, A\, _h \cdot\left \, (z^,\ldots,z, 1) \right \, _k. If , this is :, z, ^n\leq\, A\, _1\max \left \ =\, A\, _1, z, ^. Thus :, z, \leq \max\. In the case , the summation formula for a
geometric progression A geometric progression, also known as a geometric sequence, is a mathematical sequence of non-zero numbers where each term after the first is found by multiplying the previous one by a fixed number called the ''common ratio''. For example, the s ...
, gives :, z, ^n\leq \, A\, _h \left(, z, ^+\cdots+, z, ^k +1\right)^=\, A\, _h \left(\frac\right)^\leq\, A\, _h \left(\frac\right)^\frac. Thus :, z, ^\leq \left(\, A\, _h\right)^k \frac, which simplifies to :, z, ^k\leq 1+\left(\, A\, _h\right)^k. Thus, in all cases :, z, \leq \left \, \left (1,\, A\, _h \right ) \right \, _k, which finishes the proof.


Other bounds

Many other upper bounds for the magnitudes of all roots have been given. Fujiwara's bound :2\, \max \left\, slightly improves the bound given above by dividing the last argument of the maximum by two. Kojima's bound is :2\,\max \left\, where a_i denotes the th ''nonzero'' coefficient when the terms of the polynomials are sorted by increasing degrees. If all coefficients are nonzero, Fujiwara's bound is sharper, since each element in Fujiwara's bound is the
geometric mean In mathematics, the geometric mean is a mean or average which indicates a central tendency of a finite collection of positive real numbers by using the product of their values (as opposed to the arithmetic mean which uses their sum). The geometri ...
of first elements in Kojima's bound. Sun and Hsieh obtained another improvement on Cauchy's bound. Assume the polynomial is monic with general term . Sun and Hsieh showed that upper bounds and could be obtained from the following equations. :d_1 = \tfrac \left((, a_, - 1) + \sqrt \right), \qquad a = \max \. is the positive root of the cubic equation :Q(x) = x^3 + (2 - , a_, ) x^2 + (1 - , a_, - , a_, ) x - a, \qquad a = \max \ They also noted that .


Landau's inequality

The previous bounds are upper bounds for each root separately. Landau's inequality provides an upper bound for the absolute values of the product of the roots that have an absolute value greater than one. This inequality, discovered in 1905 by
Edmund Landau Edmund Georg Hermann Landau (14 February 1877 – 19 February 1938) was a German mathematician who worked in the fields of number theory and complex analysis. Biography Edmund Landau was born to a Jewish family in Berlin. His father was Leopo ...
, has been forgotten and rediscovered at least three times during the 20th century. This bound of the product of roots is not much greater than the best preceding bounds of each root separately. Let z_1, \ldots, z_n be the roots of the polynomial . If :M(p)=, a_n, \prod_^n \max(1,, z_j, ) is the
Mahler measure In mathematics, the Mahler measure M(p) of a polynomial p(z) with complex coefficients is defined as M(p) = , a, \prod_ , \alpha_i, = , a, \prod_^n \max\, where p(z) factorizes over the complex numbers \mathbb as p(z) = a(z-\alpha_1)(z-\alpha ...
of , then :M(p)\le \sqrt. Surprisingly, this bound of the product of the absolute values larger than 1 of the roots is not much larger than the best bounds of ''one'' root that have been given above for a single root. This bound is even exactly equal to one of the bounds that are obtained using Hölder's inequality. This bound is also useful to bound the coefficients of a divisor of a polynomial with integer coefficients: if :q= \sum_^m b_k x^k is a divisor of , then :, b_m, \le, a_n, , and, by
Vieta's formulas In mathematics, Vieta's formulas relate the coefficients of a polynomial to sums and products of its roots. They are named after François Viète (1540-1603), more commonly referred to by the Latinised form of his name, "Franciscus Vieta." Basi ...
, :\frac\le \binom mi \frac, for , where \binom mi is a
binomial coefficient In mathematics, the binomial coefficients are the positive integers that occur as coefficients in the binomial theorem. Commonly, a binomial coefficient is indexed by a pair of integers and is written \tbinom. It is the coefficient of the t ...
. Thus :, b_i, \le \binom mi M(p)\le \binom mi \sqrt, and :\sum_^m , b_i, \le 2^m M(p) \le 2^m \sqrt.


Discs containing some roots


From Rouché theorem

Rouché's theorem Rouché's theorem, named after Eugène Rouché, states that for any two complex-valued functions and holomorphic inside some region K with closed contour \partial K, if on \partial K, then and have the same number of zeros inside K, where ...
allows defining discs centered at zero and containing a given number of roots. More precisely, if there is a positive real number and an integer such that :, a_k, R^k > , a_0, +\cdots+, a_, R^+, a_, R^+\cdots+, a_n, R^n, then there are exactly roots, counted with multiplicity, of absolute value less than . If , z, =R, then :\begin , a_0 &+\cdots+a_z^+a_ z^+\cdots+a_n z^n, \\ &\le , a_0, +\cdots+, a_, R^+, a_, R^+\cdots+, a_n, R^n\\ &\le , a_k, R^k \le , a_k z^k, . \end By Rouché's theorem, this implies directly that p(z) and z^k have the same number of roots of absolute values less than , counted with multiplicities. As this number is , the result is proved. The above result may be applied if the polynomial :h_k(x)=, a_0, +\cdots+, a_, x^-, a_k, x^k+, a_, x^+\cdots+, a_n, x^n. takes a negative value for some positive real value of . In the remaining of the section, suppose that . If it is not the case, zero is a root, and the localization of the other roots may be studied by dividing the polynomial by a power of the indeterminate, getting a polynomial with a nonzero constant term. For and ,
Descartes' rule of signs In mathematics, Descartes' rule of signs, described by René Descartes in his ''La Géométrie'', counts the roots of a polynomial by examining sign changes in its coefficients. The number of positive real roots is at most the number of sign chang ...
shows that the polynomial has exactly one positive real root. If R_0 and R_n are these roots, the above result shows that all the roots satisfy :R_0\le , z, \le R_1. As these inequalities apply also to h_0 and h_n, these bounds are optimal for polynomials with a given sequence of the absolute values of their coefficients. They are thus sharper than all bounds given in the preceding sections. For , Descartes' rule of signs implies that h_k(x) either has two positive real roots that are not multiple, or is nonnegative for every positive value of . So, the above result may be applied only in the first case. If R_ are these two roots, the above result implies that :, z, \le R_ for roots of , and that :, z, \ge R_ for the other roots. Instead of explicitly computing R_ and R_, it is generally sufficient to compute a value R_k such that h_k(R_k)<0 (necessarily R_). These R_k have the property of separating roots in terms of their absolute values: if, for , both R_h and R_k exist, there are exactly roots such that R_h < , z, < R_k. For computing R_k, one can use the fact that \frac is a
convex function In mathematics, a real-valued function is called convex if the line segment between any two distinct points on the graph of a function, graph of the function lies above or on the graph between the two points. Equivalently, a function is conve ...
(its second derivative is positive). Thus R_k exists if and only if \frac is negative at its unique minimum. For computing this minimum, one can use any
optimization Mathematical optimization (alternatively spelled ''optimisation'') or mathematical programming is the selection of a best element, with regard to some criteria, from some set of available alternatives. It is generally divided into two subfiel ...
method, or, alternatively,
Newton's method In numerical analysis, the Newton–Raphson method, also known simply as Newton's method, named after Isaac Newton and Joseph Raphson, is a root-finding algorithm which produces successively better approximations to the roots (or zeroes) of a ...
for computing the unique positive zero of the derivative of \frac (it converges rapidly, as the derivative is a
monotonic function In mathematics, a monotonic function (or monotone function) is a function between ordered sets that preserves or reverses the given order. This concept first arose in calculus, and was later generalized to the more abstract setting of or ...
). One can increase the number of existing R_k's by applying the root squaring operation of the Dandelin–Graeffe iteration. If the roots have distinct absolute values, one can eventually completely separate the roots in terms of their absolute values, that is, compute positive numbers R_0 < R_1 <\dots such there is exactly one root with an absolute value in the open interval (R_,R_k), for .


From Gershgorin circle theorem

The
Gershgorin circle theorem In mathematics, the Gershgorin circle theorem may be used to bound the spectrum of a square matrix. It was first published by the Soviet mathematician Semyon Aronovich Gershgorin in 1931. Gershgorin's name has been transliterated in several diffe ...
applies the
companion matrix In linear algebra, the Frobenius companion matrix of the monic polynomial p(x)=c_0 + c_1 x + \cdots + c_x^ + x^n is the square matrix defined as C(p)=\begin 0 & 0 & \dots & 0 & -c_0 \\ 1 & 0 & \dots & 0 & -c_1 \\ 0 & 1 & \dots & 0 & -c_2 \\ \ ...
of the polynomial on a basis related to
Lagrange interpolation In numerical analysis, the Lagrange interpolating polynomial is the unique polynomial of lowest degree that interpolates a given set of data. Given a data set of coordinate pairs (x_j, y_j) with 0 \leq j \leq k, the x_j are called ''nodes'' ...
to define discs centered at the interpolation points, each containing a root of the polynomial; see for details. If the interpolation points are close to the roots of the roots of the polynomial, the radii of the discs are small, and this is a key ingredient of Durand–Kerner method for computing polynomial roots.


Bounds of real roots

For polynomials with real coefficients, it is often useful to bound only the real roots. It suffices to bound the positive roots, as the negative roots of are the positive roots of . Clearly, every bound of all roots applies also for real roots. But in some contexts, tighter bounds of real roots are useful. For example, the efficiency of the
method of continued fractions The method of continued fractions is a method developed specifically for solution of integral equations of quantum scattering theory like Lippmann–Schwinger equation or Faddeev equations. It was invented by Jiří Horáček (physicist), Horáček ...
for
real-root isolation In mathematics, and, more specifically in numerical analysis and computer algebra, real-root isolation of a polynomial consist of producing disjoint intervals of the real line, which contain each one (and only one) real root of the polynomial, and ...
strongly depends on tightness of a bound of positive roots. This has led to establishing new bounds that are tighter than the general bounds of all roots. These bounds are generally expressed not only in terms of the absolute values of the coefficients, but also in terms of their signs. Other bounds apply only to polynomials whose all roots are reals (see below).


Bounds of positive real roots

To give a bound of the positive roots, one can assume a_n >0 without loss of generality, as changing the signs of all coefficients does not change the roots. Every upper bound of the positive roots of :q(x)=a_nx^n + \sum_^ \min(0,a_i)x^i is also a bound of the real zeros of :p(x)=\sum_^n a_ix^i. In fact, if is such a bound, for all , one has . Applied to Cauchy's bound, this gives the upper bound :1+ \frac for the real roots of a polynomial with real coefficients. If this bound is not greater than , this means that all nonzero coefficients have the same sign, and that there is no positive root. Similarly, another upper bound of the positive roots is :2\,\left(\frac\right)^. If all nonzero coefficients have the same sign, there is no positive root, and the maximum must be zero. Other bounds have been recently developed, mainly for the
method of continued fractions The method of continued fractions is a method developed specifically for solution of integral equations of quantum scattering theory like Lippmann–Schwinger equation or Faddeev equations. It was invented by Jiří Horáček (physicist), Horáček ...
for
real-root isolation In mathematics, and, more specifically in numerical analysis and computer algebra, real-root isolation of a polynomial consist of producing disjoint intervals of the real line, which contain each one (and only one) real root of the polynomial, and ...
.


Polynomials whose roots are all real

If all roots of a polynomial are real,
Laguerre Edmond Nicolas Laguerre (9 April 1834, Bar-le-Duc – 14 August 1886, Bar-le-Duc) was a French mathematician and a member of the Académie des sciences (1885). His main works were in the areas of geometry and complex analysis. He also investigate ...
proved the following lower and upper bounds of the roots, by using what is now called Samuelson's inequality.. Let \sum_^n a_k x^k be a polynomial with all real roots. Then its roots are located in the interval with endpoints :-\frac \pm \frac\sqrt. For example, the roots of the polynomial x^4+5x^3+5x^2-5x-6=(x+3)(x+2)(x+1)(x-1) satisfy :-3.8118<-\frac - \frac\sqrt\le x\le -\frac + \frac\sqrt<1.3118.


Root separation

The root separation of a polynomial is the minimal distance between two roots, that is the minimum of the absolute values of the difference of two roots: :\operatorname(p) = \min\ The root separation is a fundamental parameter of the
computational complexity In computer science, the computational complexity or simply complexity of an algorithm is the amount of resources required to run it. Particular focus is given to computation time (generally measured by the number of needed elementary operations ...
of
root-finding algorithm In numerical analysis, a root-finding algorithm is an algorithm for finding zeros, also called "roots", of continuous functions. A zero of a function is a number such that . As, generally, the zeros of a function cannot be computed exactly nor ...
s for polynomials. In fact, the root separation determines the precision of number representation that is needed for being certain of distinguishing distinct roots. Also, for
real-root isolation In mathematics, and, more specifically in numerical analysis and computer algebra, real-root isolation of a polynomial consist of producing disjoint intervals of the real line, which contain each one (and only one) real root of the polynomial, and ...
, it allows bounding the number of interval divisions that are needed for isolating all roots. For polynomials with real or complex coefficients, it is not possible to express a lower bound of the root separation in terms of the degree and the absolute values of the coefficients only, because a small change on a single coefficient transforms a polynomial with multiple roots into a
square-free polynomial In mathematics, a square-free polynomial is a univariate polynomial (over a field or an integral domain) that has no multiple root in an algebraically closed field containing its coefficients. In characteristic 0, or over a finite field, a univar ...
with a small root separation, and essentially the same absolute values of the coefficient. However, involving the
discriminant In mathematics, the discriminant of a polynomial is a quantity that depends on the coefficients and allows deducing some properties of the zero of a function, roots without computing them. More precisely, it is a polynomial function of the coef ...
of the polynomial allows a lower bound. For square-free polynomials with integer coefficients, the discriminant is an integer, and has thus an absolute value that is not smaller than . This allows lower bounds for root separation that are independent from the discriminant. Mignotte's separation bound is :\operatorname(p) > \frac , where \Delta(p) is the discriminant, and \textstyle\, p\, _2=\sqrt. For a square free polynomial with integer coefficients, this implies :\operatorname(p) > \frac > \frac 1, where is the
bit The bit is the most basic unit of information in computing and digital communication. The name is a portmanteau of binary digit. The bit represents a logical state with one of two possible values. These values are most commonly represented as ...
size of , that is the sum of the bitsize of its coefficients.


Gauss–Lucas theorem

The Gauss–Lucas theorem states that the
convex hull In geometry, the convex hull, convex envelope or convex closure of a shape is the smallest convex set that contains it. The convex hull may be defined either as the intersection of all convex sets containing a given subset of a Euclidean space, ...
of the roots of a polynomial contains the roots of the
derivative In mathematics, the derivative is a fundamental tool that quantifies the sensitivity to change of a function's output with respect to its input. The derivative of a function of a single variable at a chosen input value, when it exists, is t ...
of the polynomial. A sometimes useful corollary is that, if all roots of a polynomial have positive real part, then so do the roots of all derivatives of the polynomial. A related result is Bernstein's inequality. It states that for a polynomial ''P'' of degree ''n'' with derivative ''P′'' we have :\max_ \big, P'(z)\big, \le n \max_ \big, P(z)\big, .


Statistical distribution of the roots

If the coefficients of a random polynomial are independently and identically distributed with a
mean A mean is a quantity representing the "center" of a collection of numbers and is intermediate to the extreme values of the set of numbers. There are several kinds of means (or "measures of central tendency") in mathematics, especially in statist ...
of zero, most complex roots are on the unit circle or close to it. In particular, the real roots are mostly located near , and, moreover, their expected number is, for a large degree, less than the
natural logarithm The natural logarithm of a number is its logarithm to the base of a logarithm, base of the e (mathematical constant), mathematical constant , which is an Irrational number, irrational and Transcendental number, transcendental number approxima ...
of the degree. If the coefficients are Gaussian distributed with a mean of zero and
variance In probability theory and statistics, variance is the expected value of the squared deviation from the mean of a random variable. The standard deviation (SD) is obtained as the square root of the variance. Variance is a measure of dispersion ...
of ''σ'' then the mean density of real roots is given by the Kac formula : m( x ) = \frac where : \begin A(x) &= \sigma \sum_^ x^ = \sigma \frac, \\ B(x) &= \frac 1 2 \frac A( x ), \\ C(x) &= \frac 1 4 \frac A(x) + \frac 1 \frac d A(x). \end When the coefficients are Gaussian distributed with a non-zero mean and variance of ''σ'', a similar but more complex formula is known.


Real roots

For large , the mean density of real roots near is asymptotically : m( x ) = \frac if x^2-1\ne 0, and : m(\pm 1) = \frac 1 \pi \sqrt It follows that the expected number of real roots is, using big notation : N_n = \frac 2 \pi \ln n + C + \frac 2 +O( n^ ) where is a constant approximately equal to . In other words, ''the expected number of real roots of a random polynomial of high degree is lower than the
natural logarithm The natural logarithm of a number is its logarithm to the base of a logarithm, base of the e (mathematical constant), mathematical constant , which is an Irrational number, irrational and Transcendental number, transcendental number approxima ...
of the degree''. Kac, Erdős and others have shown that these results are insensitive to the distribution of the coefficients, if they are independent and have the same distribution with mean zero. However, if the variance of the th coefficient is equal to \binom ni, the expected number of real roots is \sqrt n.


Geometry of multiple roots

A polynomial p can be written in the form of p(x) = a (x-z_1)^ \cdots (x-z_k)^ with distinct roots z_1,\ldots,z_k and corresponding
multiplicities In mathematics, the multiplicity of a member of a multiset is the number of times it appears in the multiset. For example, the number of times a given polynomial has a root at a given point is the multiplicity of that root. The notion of multipl ...
m_1,\ldots,m_k. A root z_j is a simple root if m_j=1 or a multiple root if m_j\ge 2. Simple roots are
Lipschitz continuous In mathematical analysis, Lipschitz continuity, named after Germany, German mathematician Rudolf Lipschitz, is a strong form of uniform continuity for function (mathematics), functions. Intuitively, a Lipschitz continuous function is limited in h ...
with respect to coefficients but multiple roots are not. In other words, simple roots have bounded sensitivities but multiple roots are infinitely sensitive if the coefficients are perturbed arbitrarily. As a result, most root-finding algorithms suffer substantial loss of accuracy on multiple roots in numerical computation. In 1972,
William Kahan William "Velvel" Morton Kahan (born June 5, 1933) is a Canadian mathematician and computer scientist, who is a professor emeritus at University of California, Berkeley. He received the Turing Award in 1989 for "his fundamental contributions to nu ...
proved that there is an inherent stability of multiple roots. Kahan discovered that polynomials with a particular set of multiplicities form what he called a ''pejorative manifold'' and proved that a multiple root is
Lipschitz continuous In mathematical analysis, Lipschitz continuity, named after Germany, German mathematician Rudolf Lipschitz, is a strong form of uniform continuity for function (mathematics), functions. Intuitively, a Lipschitz continuous function is limited in h ...
if the perturbation maintains its multiplicity. This geometric property of multiple roots is crucial in numerical computation of multiple roots.


See also

* * * * Quadratic function#Upper bound on the magnitude of the roots * * * * * Cohn's theorem relating the roots of a self-inversive polynomial with the roots of the reciprocal polynomial of its derivative.


Notes


References

* * {{cite book , last=Yap , first= Chee-Keng, title=Fundamental problems of algorithmic algebra, publisher=
Oxford University Press Oxford University Press (OUP) is the publishing house of the University of Oxford. It is the largest university press in the world. Its first book was printed in Oxford in 1478, with the Press officially granted the legal right to print books ...
, year = 2000, isbn=978-0195125160, url=http://160592857366.free.fr/joe/ebooks/ShareData/Fundamental%20Problems%20in%20Algorithmic%20Algebra%201993%20Yap.pdf.


External links


The beauty of the roots
a visualization of the distribution of all roots of all polynomials with degree and integer coefficients in some range. Polynomials