HOME

TheInfoList



OR:

In
mathematics Mathematics is an area of knowledge that includes the topics of numbers, formulas and related structures, shapes and the spaces in which they are contained, and quantities and their changes. These topics are represented in modern mathematics ...
, specifically in
commutative algebra Commutative algebra, first known as ideal theory, is the branch of algebra that studies commutative rings, their ideals, and modules over such rings. Both algebraic geometry and algebraic number theory build on commutative algebra. Prom ...
, the elementary symmetric polynomials are one type of basic building block for symmetric polynomials, in the sense that any symmetric polynomial can be expressed as a polynomial in elementary symmetric polynomials. That is, any symmetric polynomial is given by an expression involving only additions and multiplication of constants and elementary symmetric polynomials. There is one elementary symmetric polynomial of degree in variables for each
positive integer In mathematics, the natural numbers are those numbers used for counting (as in "there are ''six'' coins on the table") and ordering (as in "this is the ''third'' largest city in the country"). Numbers used for counting are called '' cardinal ...
, and it is formed by adding together all distinct products of distinct variables.


Definition

The elementary symmetric polynomials in variables , written for , are defined by :\begin e_1 (X_1, X_2, \dots,X_n) &= \sum_ X_j,\\ e_2 (X_1, X_2, \dots,X_n) &= \sum_ X_j X_k,\\ e_3 (X_1, X_2, \dots,X_n) &= \sum_ X_j X_k X_l,\\ \end and so forth, ending with : e_n (X_1, X_2, \dots,X_n) = X_1 X_2 \cdots X_n. In general, for we define : e_k (X_1 , \ldots , X_n )=\sum_ X_ \dotsm X_, so that if . (Sometimes, is included among the elementary symmetric polynomials, but excluding it allows generally simpler formulation of results and properties.) Thus, for each positive integer less than or equal to there exists exactly one elementary symmetric polynomial of degree in variables. To form the one that has degree , we take the sum of all products of -subsets of the variables. (By contrast, if one performs the same operation using ''multisets'' of variables, that is, taking variables with repetition, one arrives at the
complete homogeneous symmetric polynomial In mathematics, specifically in algebraic combinatorics and commutative algebra, the complete homogeneous symmetric polynomials are a specific kind of symmetric polynomials. Every symmetric polynomial can be expressed as a polynomial expression ...
s.) Given an
integer partition In number theory and combinatorics, a partition of a positive integer , also called an integer partition, is a way of writing as a sum of positive integers. Two sums that differ only in the order of their summands are considered the same part ...
(that is, a finite non-increasing sequence of positive integers) , one defines the symmetric polynomial , also called an elementary symmetric polynomial, by : e_\lambda (X_1, \dots,X_n) = e_(X_1, \dots, X_n) \cdot e_(X_1, \dots, X_n) \cdots e_(X_1, \dots, X_n). Sometimes the notation is used instead of .


Examples

The following lists the elementary symmetric polynomials for the first four positive values of . For : :e_1(X_1) = X_1. For : :\begin e_1(X_1,X_2) &= X_1 + X_2,\\ e_2(X_1,X_2) &= X_1X_2.\,\\ \end For : :\begin e_1(X_1,X_2,X_3) &= X_1 + X_2 + X_3,\\ e_2(X_1,X_2,X_3) &= X_1X_2 + X_1X_3 + X_2X_3,\\ e_3(X_1,X_2,X_3) &= X_1X_2X_3.\,\\ \end For : :\begin e_1(X_1,X_2,X_3,X_4) &= X_1 + X_2 + X_3 + X_4,\\ e_2(X_1,X_2,X_3,X_4) &= X_1X_2 + X_1X_3 + X_1X_4 + X_2X_3 + X_2X_4 + X_3X_4,\\ e_3(X_1,X_2,X_3,X_4) &= X_1X_2X_3 + X_1X_2X_4 + X_1X_3X_4 + X_2X_3X_4,\\ e_4(X_1,X_2,X_3,X_4) &= X_1X_2X_3X_4.\,\\ \end


Properties

The elementary symmetric polynomials appear when we expand a linear factorization of a
monic polynomial In algebra, a monic polynomial is a single-variable polynomial (that is, a univariate polynomial) in which the leading coefficient (the nonzero coefficient of highest degree) is equal to 1. Therefore, a monic polynomial has the form: :x^n+c_x^+\ ...
: we have the identity :\prod_^n ( \lambda - X_j)=\lambda^n - e_1(X_1,\ldots,X_n)\lambda^ + e_2(X_1,\ldots,X_n)\lambda^ + \cdots +(-1)^n e_n(X_1,\ldots,X_n). That is, when we substitute numerical values for the variables , we obtain the monic
univariate polynomial In mathematics, a polynomial is an expression consisting of indeterminates (also called variables) and coefficients, that involves only the operations of addition, subtraction, multiplication, and positive-integer powers of variables. An ex ...
(with variable ) whose
roots A root is the part of a plant, generally underground, that anchors the plant body, and absorbs and stores water and nutrients. Root or roots may also refer to: Art, entertainment, and media * ''The Root'' (magazine), an online magazine focusing ...
are the values substituted for and whose
coefficient In mathematics, a coefficient is a multiplicative factor in some term of a polynomial, a series, or an expression; it is usually a number, but may be any expression (including variables such as , and ). When the coefficients are themselves ...
s are –
up to Two mathematical objects ''a'' and ''b'' are called equal up to an equivalence relation ''R'' * if ''a'' and ''b'' are related by ''R'', that is, * if ''aRb'' holds, that is, * if the equivalence classes of ''a'' and ''b'' with respect to ''R'' ...
their sign – the elementary symmetric polynomials. These relations between the roots and the coefficients of a polynomial are called
Vieta's formulas In mathematics, Vieta's formulas relate the coefficients of a polynomial to sums and products of its roots. They are named after François Viète (more commonly referred to by the Latinised form of his name, "Franciscus Vieta"). Basic formula ...
. The
characteristic polynomial In linear algebra, the characteristic polynomial of a square matrix is a polynomial which is invariant under matrix similarity and has the eigenvalues as roots. It has the determinant and the trace of the matrix among its coefficients. The c ...
of a
square matrix In mathematics, a square matrix is a matrix with the same number of rows and columns. An ''n''-by-''n'' matrix is known as a square matrix of order Any two square matrices of the same order can be added and multiplied. Square matrices are often ...
is an example of application of Vieta's formulas. The roots of this polynomial are the
eigenvalue In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denote ...
s of the
matrix Matrix most commonly refers to: * ''The Matrix'' (franchise), an American media franchise ** '' The Matrix'', a 1999 science-fiction action film ** "The Matrix", a fictional setting, a virtual reality environment, within ''The Matrix'' (franchi ...
. When we substitute these eigenvalues into the elementary symmetric polynomials, we obtain – up to their sign – the coefficients of the characteristic polynomial, which are invariants of the matrix. In particular, the trace (the sum of the elements of the diagonal) is the value of , and thus the sum of the eigenvalues. Similarly, the
determinant In mathematics, the determinant is a scalar value that is a function of the entries of a square matrix. It characterizes some properties of the matrix and the linear map represented by the matrix. In particular, the determinant is nonzero if a ...
is – up to the sign – the constant term of the characteristic polynomial, i.e. the value of . Thus the determinant of a square matrix is the product of the eigenvalues. The set of elementary symmetric polynomials in variables generates the ring of symmetric polynomials in variables. More specifically, the ring of symmetric polynomials with integer coefficients equals the integral
polynomial ring In mathematics, especially in the field of algebra, a polynomial ring or polynomial algebra is a ring (which is also a commutative algebra) formed from the set of polynomials in one or more indeterminates (traditionally also called variables ...
. (See below for a more general statement and
proof Proof most often refers to: * Proof (truth), argument or sufficient evidence for the truth of a proposition * Alcohol proof, a measure of an alcoholic drink's strength Proof may also refer to: Mathematics and formal logic * Formal proof, a c ...
.) This fact is one of the foundations of invariant theory. For another system of symmetric polynomials with the same property see
Complete homogeneous symmetric polynomial In mathematics, specifically in algebraic combinatorics and commutative algebra, the complete homogeneous symmetric polynomials are a specific kind of symmetric polynomials. Every symmetric polynomial can be expressed as a polynomial expression ...
s, and for a system with a similar, but slightly weaker, property see
Power sum symmetric polynomial In mathematics, specifically in commutative algebra, the power sum symmetric polynomials are a type of basic building block for symmetric polynomials, in the sense that every symmetric polynomial with rational coefficients can be expressed as a su ...
.


Fundamental theorem of symmetric polynomials

For any
commutative ring In mathematics, a commutative ring is a ring in which the multiplication operation is commutative. The study of commutative rings is called commutative algebra. Complementarily, noncommutative algebra is the study of ring properties that are not ...
, denote the ring of symmetric polynomials in the variables with coefficients in by . This is a polynomial ring in the ''n'' elementary symmetric polynomials for . This means that every symmetric polynomial has a unique representation : P(X_1,\ldots, X_n)=Q\big(e_1(X_1 , \ldots ,X_n), \ldots, e_n(X_1 , \ldots ,X_n)\big) for some polynomial . Another way of saying the same thing is that the ring homomorphism that sends to for defines an
isomorphism In mathematics, an isomorphism is a structure-preserving mapping between two structures of the same type that can be reversed by an inverse mapping. Two mathematical structures are isomorphic if an isomorphism exists between them. The word i ...
between and .


Proof sketch

The theorem may be proved for symmetric
homogeneous polynomial In mathematics, a homogeneous polynomial, sometimes called quantic in older texts, is a polynomial whose nonzero terms all have the same degree. For example, x^5 + 2 x^3 y^2 + 9 x y^4 is a homogeneous polynomial of degree 5, in two variables; ...
s by a double induction with respect to the number of variables and, for fixed , with respect to the degree of the homogeneous polynomial. The general case then follows by splitting an arbitrary symmetric polynomial into its homogeneous components (which are again symmetric). In the case the result is trivial because every polynomial in one variable is automatically symmetric. Assume now that the theorem has been proved for all polynomials for variables and all symmetric polynomials in variables with degree . Every homogeneous symmetric polynomial in can be decomposed as a sum of homogeneous symmetric polynomials : P(X_1,\ldots,X_n)= P_ (X_1,\ldots,X_n) + X_1 \cdots X_n \cdot Q(X_1,\ldots,X_n). Here the "lacunary part" is defined as the sum of all monomials in which contain only a proper subset of the variables , i.e., where at least one variable is missing. Because is symmetric, the lacunary part is determined by its terms containing only the variables , i.e., which do not contain . More precisely: If and are two homogeneous symmetric polynomials in having the same degree, and if the coefficient of before each monomial which contains only the variables equals the corresponding coefficient of , then and have equal lacunary parts. (This is because every monomial which can appear in a lacunary part must lack at least one variable, and thus can be transformed by a permutation of the variables into a monomial which contains only the variables .) But the terms of which contain only the variables are precisely the terms that survive the operation of setting to 0, so their sum equals , which is a symmetric polynomial in the variables that we shall denote by . By the inductive hypothesis, this polynomial can be written as : \tilde(X_1, \ldots, X_)=\tilde(\sigma_, \ldots, \sigma_) for some . Here the doubly indexed denote the elementary symmetric polynomials in variables. Consider now the polynomial :R(X_1, \ldots, X_):= \tilde(\sigma_, \ldots, \sigma_) . Then is a symmetric polynomial in , of the same degree as , which satisfies :R(X_1, \ldots, X_,0) = \tilde(\sigma_, \ldots, \sigma_) = P(X_1, \ldots,X_,0) (the first equality holds because setting to 0 in gives , for all ). In other words, the coefficient of before each monomial which contains only the variables equals the corresponding coefficient of . As we know, this shows that the lacunary part of coincides with that of the original polynomial . Therefore the difference has no lacunary part, and is therefore divisible by the product of all variables, which equals the elementary symmetric polynomial . Then writing , the quotient is a homogeneous symmetric polynomial of degree less than (in fact degree at most ) which by the inductive hypothesis can be expressed as a polynomial in the elementary symmetric functions. Combining the representations for and one finds a polynomial representation for . The uniqueness of the representation can be proved inductively in a similar way. (It is equivalent to the fact that the polynomials are algebraically independent over the ring .) The fact that the polynomial representation is unique implies that is isomorphic to .


Alternative proof

The following proof is also inductive, but does not involve other polynomials than those symmetric in , and also leads to a fairly direct procedure to effectively write a symmetric polynomial as a polynomial in the elementary symmetric ones. Assume the symmetric polynomial to be homogeneous of degree ; different homogeneous components can be decomposed separately. Order the
monomial In mathematics, a monomial is, roughly speaking, a polynomial which has only one term. Two definitions of a monomial may be encountered: # A monomial, also called power product, is a product of powers of variables with nonnegative integer expon ...
s in the variables lexicographically, where the individual variables are ordered , in other words the dominant term of a polynomial is one with the highest occurring power of , and among those the one with the highest power of , etc. Furthermore parametrize all products of elementary symmetric polynomials that have degree (they are in fact homogeneous) as follows by
partitions Partition may refer to: Computing Hardware * Disk partitioning, the division of a hard disk drive * Memory partition, a subdivision of a computer's memory, usually for use by a single job Software * Partition (database), the division of ...
of . Order the individual elementary symmetric polynomials in the product so that those with larger indices come first, then build for each such factor a column of boxes, and arrange those columns from left to right to form a
Young diagram In mathematics, a Young tableau (; plural: tableaux) is a combinatorial object useful in representation theory and Schubert calculus. It provides a convenient way to describe the group representations of the symmetric and general linear groups ...
containing boxes in all. The shape of this diagram is a partition of , and each partition of arises for exactly one product of elementary symmetric polynomials, which we shall denote by ) (the is present only because traditionally this product is associated to the transpose partition of ). The essential ingredient of the proof is the following simple property, which uses multi-index notation for monomials in the variables . Lemma. The leading term of is . :''Proof''. The leading term of the product is the product of the leading terms of each factor (this is true whenever one uses a
monomial order In mathematics, a monomial order (sometimes called a term order or an admissible order) is a total order on the set of all ( monic) monomials in a given polynomial ring, satisfying the property of respecting multiplication, i.e., * If u \leq v and ...
, like the lexicographic order used here), and the leading term of the factor is clearly . To count the occurrences of the individual variables in the resulting monomial, fill the column of the Young diagram corresponding to the factor concerned with the numbers of the variables, then all boxes in the first row contain 1, those in the second row 2, and so forth, which means the leading term is . Now one proves by induction on the leading monomial in lexicographic order, that any nonzero homogeneous symmetric polynomial of degree can be written as polynomial in the elementary symmetric polynomials. Since is symmetric, its leading monomial has weakly decreasing exponents, so it is some with a partition of . Let the coefficient of this term be , then is either zero or a symmetric polynomial with a strictly smaller leading monomial. Writing this difference inductively as a polynomial in the elementary symmetric polynomials, and adding back to it, one obtains the sought for polynomial expression for . The fact that this expression is unique, or equivalently that all the products (monomials) of elementary symmetric polynomials are linearly independent, is also easily proved. The lemma shows that all these products have different leading monomials, and this suffices: if a nontrivial linear combination of the were zero, one focuses on the contribution in the linear combination with nonzero coefficient and with (as polynomial in the variables ) the largest leading monomial; the leading term of this contribution cannot be cancelled by any other contribution of the linear combination, which gives a contradiction.


See also

* Symmetric polynomial *
Complete homogeneous symmetric polynomial In mathematics, specifically in algebraic combinatorics and commutative algebra, the complete homogeneous symmetric polynomials are a specific kind of symmetric polynomials. Every symmetric polynomial can be expressed as a polynomial expression ...
* Schur polynomial *
Newton's identities In mathematics, Newton's identities, also known as the Girard–Newton formulae, give relations between two types of symmetric polynomials, namely between power sums and elementary symmetric polynomials. Evaluated at the roots of a monic polynom ...
*
Newton's inequalities In mathematics, the Newton inequalities are named after Isaac Newton. Suppose ''a''1, ''a''2, ..., ''a'n'' are real numbers and let e_k denote the ''k''th elementary symmetric polynomial in ''a''1, ''a''2, ..., ' ...
* Maclaurin's inequality *
MacMahon Master theorem In mathematics, MacMahon's master theorem (MMT) is a result in enumerative combinatorics and linear algebra. It was discovered by Percy MacMahon and proved in his monograph ''Combinatory analysis'' (1916). It is often used to derive binomial i ...
* Symmetric function *
Representation theory Representation theory is a branch of mathematics that studies abstract algebraic structures by ''representing'' their elements as linear transformations of vector spaces, and studies modules over these abstract algebraic structures. In essen ...


References

* * {{cite book, authorlink=Richard P. Stanley, last=Stanley, first=Richard P., date=1999, title=Enumerative Combinatorics, Vol. 2, location=Cambridge, publisher=Cambridge University Press, ISBN=0-521-56069-1 Homogeneous polynomials Symmetric functions Articles containing proofs