Resultant
In mathematics, the resultant of two polynomials is a polynomial expression of their coefficients, which is equal to zero if and only if the polynomials have a common root, or, equivalently, a common factor. In some older texts, the resultant is also called the eliminant.
The resultant is widely used in number theory, either directly or through the discriminant, which is essentially the resultant of a polynomial and its derivative. The resultant of two polynomials with rational or polynomial coefficients may be computed efficiently on a computer. It is a basic tool of computer algebra, and is a built-in function of most computer algebra systems. It is used, among others, for cylindrical algebraic decomposition, integration of rational functions and drawing of curves defined by a bivariate polynomial equation.
The resultant of n homogeneous polynomials in n variables is a generalization, introduced by Macaulay, of the usual resultant. It is, with Gröbner bases, one of the main tools of effective elimination theory.
Notation
The resultant of two univariate polynomials and is commonly denoted orIn many applications of the resultant, the polynomials depend on several indeterminates and may be considered as univariate polynomials in one of their indeterminates, with polynomials in the other indeterminates as coefficients. In this case, the indeterminate that is selected for defining and computing the resultant is indicated as a subscript: or
The degrees of the polynomials are used in the definition of the resultant. However, a polynomial of degree may also be considered as a polynomial of higher degree where the leading coefficients are zero. If such a higher degree is used for the resultant, it is usually indicated as a subscript or a superscript, such as or
Definition
The resultant of two univariate polynomials over a field or over a commutative ring is commonly defined as the determinant of their Sylvester matrix. More precisely, letand
be nonzero polynomials of degrees and respectively. Let us denote by the vector space of dimension i whose elements are the polynomials of degree strictly less than i. The map
such that
is a linear map between two spaces of the same dimension. Over the basis of the powers of , this map is represented by a square matrix of dimension, which is called the Sylvester matrix of and .
The resultant of and is thus the determinant
which has columns of and columns of .
For instance, taking and we get
If the coefficients of the polynomials belong to an integral domain, then
where and are respectively the roots, counted with their multiplicities, of and in any algebraically closed field containing the integral domain.
This is a straightforward consequence of the characterizing properties of the resultant that appear below. In the common case of integer coefficients, the algebraically closed field is generally chosen as the field of complex numbers.
Properties
In this section and its subsections, and are two polynomials in of respective degrees and, and their resultant is denotedCharacterizing properties
The resultant of two polynomials and of respective degrees and, with coefficients ina commutative ring, has the following properties that characterize the resultant, if is a field or, more generally, an integral domain
- If is a subring of another ring, then That is and have the same resultant when considered as polynomials over or.
- If then Similarly, if, then
Zeros
- The resultant of two polynomials with coefficients in an integral domain is zero if and only if they have a common divisor of positive degree.
- The resultant of two polynomials with coefficients in an integral domain is zero if and only if they have a common root in an algebraically closed field containing the coefficients.
- There exists a polynomial of degree less than and a polynomial of degree less than such that This is a generalization of Bézout's identity to polynomials over an arbitrary commutative ring. In other words, the resultant of two polynomials belongs to the ideal generated by these polynomials.
Invariance by ring homomorphisms
- If preserves the degrees of and , then
- If and then
- If and and the leading coefficient of is then
- If and and the leading coefficient of is then
Invariance under change of variable
- If and are the reciprocal polynomials of and, respectively, then
Invariance under change of polynomials
- If and are nonzero constants, and and are as above, then
- If and are as above, and is another polynomial such that the degree of is, then
- In particular, if either is monic, or, then
Generic properties
In this section, we consider two polynomialsand
whose coefficients are distinct indeterminates. Let
be the polynomial ring over the integers defined by these indeterminates.
The resultant is often called the generic resultant for the degrees and. It has the following properties.
- is an absolutely irreducible polynomial.
- If is the ideal of generated by and, then is the principal ideal generated by.
Homogeneity
- It is homogeneous of degree in
- It is homogeneous of degree in
- It is homogeneous of degree in all the variables and
- If and are given the weight , then it is quasi-homogeneous of total weight.
- If and are homogeneous multivariate polynomials of respective degrees and, then their resultant in degrees and with respect to an indeterminate, denoted in, is homogeneous of degree in the other indeterminates.
Elimination properties
- is a principal ideal generated by for some
- There exists a positive integer such that
Computation
Theoretically, the resultant could be computed by using the formula expressing it as a product of roots differences. However, as the roots may generally not be computed exactly, such an algorithm would be inefficient and numerically unstable. As the resultant is a symmetric function of the roots of each polynomial, it could also be computed by using the fundamental theorem of symmetric polynomials, but this would be highly inefficient.As the resultant is the determinant of the Sylvester matrix, it may be computed by using any algorithm for computing determinants. This needs arithmetic operations. As algorithms are known with a better complexity, this method is not used in practice.
It follows from that the computation of a resultant is strongly related to the Euclidean algorithm for polynomials. This shows that the computation of the resultant of two polynomials of degrees and may be done in arithmetic operations in the field of coefficients.
However, when the coefficients are integers, rational numbers or polynomials, these arithmetic operations imply a number of GCD computations of coefficients which is of the same order and make the algorithm inefficient.
The subresultant pseudo-remainder sequences were introduced to solve this problem and avoid any fraction and any GCD computation of coefficients. A more efficient algorithm is obtained by using the good behavior of the resultant under a ring homomorphism on the coefficients: to compute a resultant of two polynomials with integer coefficients, one computes their resultants modulo sufficiently many prime numbers and then reconstructs the result with the Chinese remainder theorem.
The use of fast multiplication of integers and polynomials allows algorithms for resultants and greatest common divisors that have a better time complexity, which is of the order of the complexity of the multiplication, multiplied by the logarithm of the size of the input.
Application to polynomial systems
Resultants were introduced for solving systems of polynomial equations and provide the oldest proof that there exist algorithms for solving such systems. These are primarily intended for systems of two equations in two unknowns, but also allow solving general systems.Case of two equations in two unknowns
Consider the system of two polynomial equationswhere and are polynomials of respective total degrees and. Then is a polynomial in, which is generically of degree . A value of is a root of if and only if either there exist in an algebraically closed field containing the coefficients, such that, or and .
Therefore, solutions to the system are obtained by computing the roots of, and for each root computing the common root of and
Bézout's theorem results from the value of, the product of the degrees of and. In fact, after a linear change of variables, one may suppose that, for each root of the resultant, there is exactly one value of such that is a common zero of and. This shows that the number of common zeros is at most the degree of the resultant, that is at most the product of the degrees of and. With some technicalities, this proof may be extended to show that, counting multiplicities and zeros at infinity, the number of zeros is exactly the product of the degrees.
General case
At first glance, it seems that resultants may be applied to a general polynomial system of equationsby computing the resultants of every pair with respect to for eliminating one unknown, and repeating the process until getting univariate polynomials. Unfortunately, this introduces many spurious solutions, which are difficult to remove.
A method, introduced at the end of the 19th century, works as follows: introduce new indeterminates and compute
This is a polynomial in whose coefficients are polynomials in which have the property that is a common zero of these polynomial coefficients, if and only if the univariate polynomials have a common zero, possibly at infinity. This process may be iterated until finding univariate polynomials.
To get a correct algorithm two complements have to be added to the method. Firstly, at each step, a linear change of variable may be needed in order that the degrees of the polynomials in the last variable are the same as their total degree. Secondly, if, at any step, the resultant is zero, this means that the polynomials have a common factor and that the solutions split in two components: one where the common factor is zero, and the other which is obtained by factoring out this common factor before continuing.
This algorithm is very complicated and has a huge time complexity. Therefore, its interest is mainly historical.
Other applications
Number theory
The discriminant of a polynomial, which is a fundamental tool in number theory is the quotient by its leading coefficient of the resultant of the polynomial and its derivative.If and are algebraic numbers such that, then is a root of the resultant and is a root of, where is the degree of. Combined with the fact that is a root of, this shows that the set of algebraic numbers is a field.
Let be an algebraic field extension generated by an element which has as minimal polynomial. Every element of may be written as where is a polynomial. Then is a root of and this resultant is a power of the minimal polynomial of
Algebraic geometry
Given two plane algebraic curves defined as the zeros of the polynomials and, the resultant allows the computation of their intersection. More precisely, the roots of are the x-coordinates of the intersection points and of the common vertical asymptotes, and the roots of are the y-coordinates of the intersection points and of the common horizontal asymptotes.A rational plane curve may be defined by a parametric equation
where, and are polynomials. An implicit equation of the curve is given by
The degree of this curve is the highest degree of, and, which is equal to the total degree of the resultant.
Symbolic integration
In symbolic integration, for computing the antiderivative of a rational fraction, one uses partial fraction decomposition for decomposing the integral into a "rational part", which is a sum of rational fractions whose antiprimitives are rational fractions, and a "logarithmic part" which is a sum of rational fractions of the formwhere is a square-free polynomial and is a polynomial of lower degree than. The antiderivative of such a function involves necessarily logarithms, and generally algebraic numbers. In fact, the antiderivative is
where the sum runs over all complex roots of.
The number of algebraic numbers involved by this expression is generally equal to the degree of, but it occurs frequently that an expression with less algebraic numbers may be computed. The Lazard–Rioboo–Trager method produced an expression, where the number of algebraic numbers is minimal, without any computation with algebraic numbers.
Let
be the square-free factorization of the resultant which appears on the right. Trager proved that the antiderivative is
where the internal sums run over the roots of the , and is a polynomial of degree in. The Lazard-Rioboo contribution is the proof that is the subresultant of degree of and It is thus obtained for free if the resultant is computed by the subresultant pseudo-remainder sequence.
Computer algebra
All preceding applications, and many others, show that the resultant is a fundamental tool in computer algebra. In fact most computer algebra systems include an efficient implementation of the computation of resultants.Homogeneous resultant
The resultant is also defined for two homogeneous polynomial in two indeterminates. Given two homogeneous polynomials and of respective total degrees and, their homogeneous resultant is the determinant of the matrix over the monomial basis of the linear mapwhere runs over the bivariate homogeneous polynomials of degree, and runs over the homogeneous polynomials of degree. In other words, the homogeneous resultant of and is the resultant of
and when they are considered as polynomials of degree and :
.
The homogeneous resultant has essentially the same properties as the usual resultant, with essentially two differences: instead of polynomial roots, one considers zeros in the projective line, and the degree of a polynomial may not change under a ring homomorphism.
That is:
- The resultant of two homogeneous polynomials over an integral domain is zero if and only if they have a non-zero common zero over an algebraically closed field containing the coefficients.
- If and are two bivariate homogeneous polynomials with coefficients in a commutative ring, and a ring homomorphism of into another commutative ring, then extending to polynomials over, ones has
- The property of an homogeneous resultant to be zero is invariant under any projective change of variables.
Macaulay's resultant
Macaulay's resultant, named after Francis Sowerby Macaulay, also called the multivariate resultant, or the multipolynomial resultant, is a generalization of the homogeneous resultant to homogeneous polynomials in indeterminates. Macaulay's resultant is a polynomial in the coefficients of these homogeneous polynomials that vanishes if and only if the polynomials have a common non-zero solution in an algebraically closed field containing the coefficients, or, equivalently, if the hyper surfaces defined by the polynomials have a common zero in the dimensional projective space. The multivariate resultant is, with Gröbner bases, one of the main tools of effective elimination theory.Like the homogeneous resultant, Macaulay's may be defined with determinants, and thus behaves well under ring homomorphisms. However, it cannot be defined by a single determinant. It follows that it is easier to define it first on generic polynomials.
Resultant of generic homogeneous polynomials
A homogeneous polynomial of degree in variables may have up tocoefficients; it is said to be generic, if these coefficients are distinct indeterminates.
Let be generic homogeneous polynomials in indeterminates, of respective degrees Together, they involve
indeterminate coefficients.
Let be the polynomial ring over the integers, in all these
indeterminate coefficients. The polynomials belong thus to and their resultant belongs to.
The Macaulay degree is the integer which is fundamental in Macaulay's theory. For defining the resultant, one considers the Macaulay matrix, which is the matrix over the monomial basis of the -linear map
in which each runs over the homogeneous polynomials of degree and the codomain is the -module of the homogeneous polynomials of degree.
If, the Macaulay matrix is the Sylvester matrix, and is a square matrix, but this is no longer true for. Thus, instead of considering the determinant, one considers all the maximal minors, that is the determinants of the square submatrices that have as many rows as the Macaulay matrix. Macaulay proved that the -ideal generated by these principal minors is a principal ideal, which is generated by the greatest common divisor of these minors. As one is working with polynomials with integer coefficients, this greatest common divisor is defined up its sign. The generic Macaulay resultant is the greatest common divisor which becomes, when, for each, zero is substituted for all coefficients of except the coefficient of for which one is substituted.
Properties of the generic Macaulay resultant
- The generic Macaulay resultant is an irreducible polynomial.
- It is homogeneous of degree in the coefficients of where is the Bézout bound.
- The product with the resultant of every monomial of degree in belongs to the ideal of generated by
Resultant of polynomials over a field
The main property of the resultant is that it is zero if and only if have a nonzero common zero in an algebraically closed extension of.
The "only if" part of this theorem results from the last property of the preceding paragraph, and is an effective version of Projective Nullstellensatz: If the resultant is nonzero, then
where is the Macaulay degree, and is the maximal homogeneous ideal. This implies that have no other common zero than the unique common zero,, of
Computability
As the computation of a resultant may be reduced to computing determinants and polynomial greatest common divisors, there are algorithms for computing resultants in a finite number of steps.However, the generic resultant is a polynomial of very high degree depending on a huge number of indeterminates. It follows that, except for very small and very small degrees of input polynomials, the generic resultant is, in practice, impossible to compute, even with modern computers. Moreover, the number of monomials of the generic resultant is so high, that, if it would be computable, the result could not be stored on available memory devices, even for rather small values of and of the degrees of the input polynomials.
Therefore, computing the resultant makes sense only for polynomials whose coefficients belong to a field or are polynomials in few indeterminates over a field.
In the case of input polynomials with coefficients in a field, the exact value of the resultant is rarely important, only its equality to zero matters. As the resultant is zero if and only if the rank of the Macaulay matrix is lower than its number of its rows, this equality to zero may by tested by applying Gaussian elimination to the Macaulay matrix. This provides a computational complexity where is the maximum degree of input polynomials.
Another case where the computation of the resultant may provide useful information is when the coefficients of the input polynomials are polynomials in a small number of indeterminates, often called parameters. In this case, the resultant, if not zero, defines a hypersurface in the parameter space. A point belongs to this hyper surface, if and only if there are values of which, together with the coordinates of the point are a zero of the input polynomials. In other words, the resultant is the result of the "elimination" of from the input polynomials.
''U''-resultant
Macaulay's resultant provides a method, called "U-resultant" by Macaulay, for solving systems of polynomial equations.Given homogeneous polynomials of degrees in indeterminates over a field, their U-resultant is the resultant of the polynomials where
is the generic linear form whose coefficients are new indeterminates Notation or for these generic coefficients is traditional, and is the origin of the term U-resultant.
The U-resultant is a homogeneous polynomial in It is zero if and only if the common zeros of form a projective algebraic set of positive dimension. If the U-resultant is not zero, its degree is the Bézout bound
The U-resultant factorizes over an algebraically closed extension of into a product of linear forms. If is such a linear factor, then are the homogeneous coordinates of a common zero of Moreover, every common zero may be obtained from one of these linear factors, and the multiplicity as a factor is equal to the intersection multiplicity of the at this zero. In other words, the U-resultant provides a completely explicit version of Bézout's theorem.
Extension to more polynomials and computation
The U-resultant as defined by Macaulay requires the number of homogeneous polynomials in the system of equations to be , where is the number of indeterminates. In 1981, Daniel Lazard extended the notion to the case where the number of polynomials may differ from, and the resulting computation can be performed via a specialized Gaussian elimination procedure followed by symbolic determinant computation.Let be homogeneous polynomials in of degrees over a field. Without loss of generality, one may suppose that Setting for, the Macaulay bound is
Let be new indeterninates and define In this case, the Macaulay matrix is defined to be the matrix, over the basis of the monomials in of the linear map
where, for each, runs over the linear space consisting of zero and the homogeneous polynomials of degree.
Reducing the Macaulay matrix by a variant of Gaussian elimination, one obtains a square matrix of linear forms in The determinant of this matrix is the U-resultant. As with the original U-resultant, it is zero if and only if have infinitely many common projective zeros. Again as with the original U-resultant, when this U-resultant is not zero, it factorizes into linear factors over any algebraically closed extension of. The coefficients of these linear factors are the homogeneous coordinates of the common zeros of and the multiplicity of a common zero equals the multiplicity of the corresponding linear factor.
The number of rows of the Macaulay matrix is less than where is the usual mathematical constant, and is the arithmetic mean of the degrees of the It follows that all solutions of a system of polynomial equations with a finite number of projective zeros can be determined in time Although this bound is large, it is nearly optimal in the following sense: if all input degrees are equal, then the time complexity of the procedure is polynomial in the expected number of solutions. This computation may be practically viable when , and are not large.