Let be a commutative ring and the polynomial ring in indeterminates over
A polynomial in it is said to be a symmetric polynomial (or symmetric function) of if it is unchanged by any permutation of the indeterminates. Consider the polynomials:
All of and are symmetric functions of .
The 's are called elementary symmetric polynomials of .
Sums, products and scalar multiples of symmetric polynomials are again symmetric polynomials.
A polynomial of becomes a symmetric polynomial when 's are written in terms of 's.
The fundamental theorem on symmetric functions is the converse of this statement, that every symmetric polynomial can be expressed uniquely in this way,
which is in this article stated and proved.
Introduce one more indeterminate and consider the polynomial
by which
Note that symmetric polynomials give general expressions for coefficients of a polynomial in terms of its roots.
The fundamental theorem (once proved) can then be thought of as being able to express any symmetric function of the roots of a polynomial, in terms of the coefficients without needing to explicitly compute the roots.
Note that the discussion so far is valid for any ring by e.g., first doing in the ring and tensoring with .
A more sophisticated way to state context is to construct the ring of symmetric functions as a "power series ring" constructed as the inverse limit as varies, of the ring of symmetric polynomials in variables over .
Details of such a construction is available eg in MacDonald (1979) [MD1979].
The exposition here is of van der Waerden (1930) [vdW1930] and Lang (1993) [La1993].
Every polynomial is a homogeneous polynomial of degree in .
Consider the homomorphism from to itself sending .
Given a polynomial , one gets the symmetric polymonial considered as a polynomial in .
The term of where is a scalar, becomes a homogeneous polynomial w. r. t. of degree in
The sum is said to be the weight of the monomial .
Weight of a polynomial may be defined as the highest among the weights of monomial summands.
A polynomial of weight as above thus yields a symmetric polynomial in of weight at most .
Fundamental Theorem of Symmetric Functions
Let be a commutative ring and the polynomial ring in indeterminates over .
A symmetric polynomial of degree can be expressed as a polynomial of the elementary symmetric polynomials.
This representation is unique.
For proof, we follow the presentation of van der Waerden (1930) [vdW1930].
First we establish that any symmetric polynomial can be expressed as a polynomial of the elementary symmetric polynomials.
The proof is constructive and uses a lexicographically ordering of the monomials, in which a term
is greater than if the first nonvanishing difference is positive.
The component monomials of are written in decreasing lexicographical order, where, given a particular monomial,
all other resulting monomials result from permuting the in it are considered together with it and written implicitly in short
using the lexicographically greatest term only inside a summation, like:
Let be the degree of and let the first term in the lexicographic ordering be .
Thus .
Now the product
has weight and when expanded as a function of the
and lexicographically ordered, has initial term .
The weight of this product is less than or equal to
This product may be subtracted from and the resulting polynomial, in lexicographical ordering,
has the term corresponding to weight and indices , equal to zero.
On this difference, the procudure may be inductively applied.
This subtraction does not increase the degree of the polynomial.
For each given degree, there are only finitely many choices of possible indices.
At each step we move further along the lexicographical ordering as some term vanishes.
Therefore the inductive procedure terminates after finitely many steps
and the polynomial is expressed as a polynomial of the elementary symmetric polynomials.
To show uniqueness of the representation, if two different polynomials functions and represent the same symmetric function,
then their (nonzero) difference evaluated on the would yield .
Therefore it is enough to show that if , then .
Given nonzero , after lexicographical ordering, each monomial summand can be written in the form
(together with its permutations).
Among all systems of indices for nonzero coefficients ,
there is one that is first in the lexicographic ordering.
Replacing by in the above and expanding in terms of ,
then in the above algorithm the first term of the lexicographical ordering is
and this term does not cancelled by any ones that are lesser in the lexicographical ordering,
and thus is nonzero.
Therefore uniqueness of the representation is proved.
QED
Weber (1895) [We1895] cites proofs of this theorem by Waring [1], Gauss [2] and Cauchy [3].
Let the symmetric group act on sending where is a permutation.
Then the fundamental theorem says that the subring fixed under this action is the subring generated by
and that there are no relations between the generators ; that is to say, the homomorphism generated by
is an isomorphism from to itself.
As the ring of symmetric functions in variables and the polynomial ring in variables are isomorphic,
one may substitute particular values in or another approprite rings for the and any relations between the the
symmetric polynomials are preserved for the evaluations.
In particular, considering as in eqref{eq:elemsymmpoly}, symmetric functions of the roots of a polynomial
are expressible as polynomials of the coefficients of .
Discriminant of a polynomial
Consider the polynomial with roots .
Consider the difference product
which is an alternating function of that is to say, it changes sign upon an odd permutation of the
Its vanishing indicates that has at least one root of multiplicity greater than .
Its square is called the discriminant:
and is a symmetric function of the .
By the preceding, the discriminant is expressible as a polynomial function of the coefficents of .
Given a more general polynomial where is not necessarily ,
with roots , the discriminant may be defined as:
The product is a vandermonde determinant:
This agrees with definition of descriminant given in the quadratic case where
it is equal to .
For the case of cubics considered for e. g., for elliptic curves such as ,
the discriminant is ,
which may be simplified for the polynomial into .
A method for computing discriminants from the coefficents of will be given proved in §
These statements may be verified either by computation by hand using that methord,
or a software computer algebra system.
The discriminant is a homogeneous polynomial of the roots; therefore it is also a homogenous w. r. t. the coefficients.
The discriminanant is a determinantal expression on polynomials in ;
therefore, it may also be thought of as an element living in the exterior algebra on .
Resultant of two polynomials
Consider two polynomials in at once, viz:
We wish to find conditions for and to have a common divisor.
We follow the approach of van der Waerden (1930) [vdW1930].
In the above, in general, it may be that either or vanish;
we call or the respective formal degree of the polynomial
and or the formal leading coefficent.
For the present we assume that and do not simultaneously vanish.
[U. F. D.]: Unique Factorization Domain
Lemma 1. and have a nonconstant common divisor only if
an equation of the form
exists, where is of degree at most and
is of degree at most , and do not simultaneously equal .
If is a field , the above condition is also sufficient.
Proof.
If is a common factor of and ,
then one may decompose and ;
these and will satisfy eqref{eq:polycommfactor}.
Now we show sufficiency and start with
If is a field , then we have unique factorization into irreducibles also in the polynomial ring .
Without loss of generality let it be that .
Every prime factor of must also be a factor of with at least the same exponent;
but does not divide as is of degree and of degree .
Therefore at least one factor of divides ; therefore is nonvanishing.
Q. E. D.
Now put
where the highest degree coefficent might also be zero.
Evaluating and expanding we get homogeneous linear equations in quantities:
For a solution to exist, the determinant of the Sylvester matrix should vanish:
Here, to avoid minus signs in the determinant, and were regarded as unknowns, and the terms invoving in eqref{eq:sylveqn}
were moved to the left, thus yielding a matrix equation in unknowns that equals zero, for which the matrix should
have zero determinant to have solutions.
[gcd]: greatest common divisor
The quantity is called the resultant of and .
By the preceding, its vanishing is necessary and sufficient condition for and to have a nonconstant common factor.
Generally, for more efficient computation of a determinant of a polynomial matrix with coefficents in a field,
one strategy that may be employed is to use the
row operations mentioned in the [article on determinants]({filename}determinants.md);
one may multiply one row with a scalar and add or subtract it from another,
or even apply a variant of the Euclidean algorithm to compute the “gcd” of the rows involved, and repeat the procedure with more rows,
and finally use the Leibniz expansion.
For ease of computation, software for computer algebra such as PARI/GP have built-in functions to compute resultants and these may be used.
The determinant contains the “principal term”
and this is the highest degree term in considered as a polynomial of and .
It vanishes when .
There are precisely rows with entries and rows with entries
The resultant is a homogeneous polynomial of degree in the and degree in the .
In the above, it was assumed that at least one of was nonzero.
To deal with the case when both are zero, consider the homogeneous polynomials
Any factorization of yield a factorization
and vice versa;
factors of may be obtained from those of by setting .
Same argument may be made for and .
Every common factor of and gives a common factor of and and vice versa.
The whole process above of taking determinant of the the Sylvester matrix may be repeated.
The case corresponds to when a nonconstant common factor of is a pure power of .
Thus the resultant criterion may be reformulated, using homogeneous polynomials, as: if the resultant is zero,
then and have a nonconstant, homogeneous common factor and conversely.
Resultant as a symmetric function
Suppose that the two polynomials above can be factored completely into linear factors:
Then the coefficents of are products by of the elementary symmetric functions of the roots
Likewise the coefficents of are products by of the elementary symmetric functions of the roots
The resultant is homogenous of degree in the and degree in the
where it is unchanged by a permutation of the or .
Applying the fundamental theorem twice in succession,
can be written as times a polynomial function of the coefficents and , that are respectively the
values of the elementary symmetric functions on and .
The following more explicit argument also may be made given and .
Suppose the and are indeterminates.
For some and let .
Then as has a common factor.
Therefore divides and likewise for each pair of and .
Thus the product of all divides .
Consider
and form the product
Similarly, from
form the product
It is convenient to define
Now like , is also homogenous of degree in the and degree in the
and is unchanged by a permutation of the or .
Therefore and can differ only by a scalar factor.
But the highest powers of either or in both and corresponds to the “principal term”
and therefore the scalar factor must be and .
The representation of in terms of and is unique by the uniqueness part of the elementary theorem for symmetric functions.
Therefore the foregoing is valid even if and do not split into linear factors.
Moreover, the computation of via determinant of Sylvester matrix is valid over any ring, as determinant computation involves
only addition, subtraction and scalar multiplication.
Regarding arbitrary field of definition, we may additionally prove the absolute irreducibility of .
That is to say, is an irreducible polynomial of the and over any field of definition.
To see this, assume that factors as where is nonconstant;
that is, for some , the monomial divides .
But as and are polynomial expressions in and , they are also symmetric functions in the roots and .
Thus the product divides .
Therefore must divide .
But looking at the determinant of the Sylvester matrix, one sees that is not divisible by or .
Therefore and a nontrivial factorization of is impossible.
Resultant and discriminant
Consider the polynomial
and its derivative .
The resultant is
By the product rule for differentiation:
Using this in eqref{eq:polyderiresu},
and if denotes the discriminant,
Thus, the discriminant may be computed as using the determinant of Sylvester Matrix for