There are a number of (families of) identities that, while they should be distinguished from Newton's identities, are very closely related to them.
A variant using complete homogeneous symmetric polynomials Denoting by
hk the
complete homogeneous symmetric polynomial (that is, the sum of all
monomials of degree
k), the power sum polynomials also satisfy identities similar to Newton's identities, but not involving any minus signs. Expressed as identities of in the
ring of symmetric functions, they read :kh_k = \sum_{i=1}^kh_{k-i}p_i, valid for all n ≥
k ≥ 1. Contrary to Newton's identities, the left-hand sides do not become zero for large
k, and the right-hand sides contain ever more non-zero terms. For the first few values of
k, one has :\begin{align} h_1 &= p_1,\\ 2h_2 &= h_1p_1 + p_2,\\ 3h_3 &= h_2p_1 + h_1p_2 + p_3.\\ \end{align} These relations can be justified by an argument analogous to the one by comparing coefficients in
power series given above, based in this case on the generating function identity :\sum_{k=0}^\infty h_k(x_1, \ldots, x_n)t^k = \prod_{i=1}^n\frac1{1 - x_it}. Proofs of Newton's identities, like these given below, cannot be easily adapted to prove these variants of those identities.
Expressing elementary symmetric polynomials in terms of power sums As mentioned, Newton's identities can be used to recursively express elementary symmetric polynomials in terms of power sums. Doing so requires the introduction of integer denominators, so it can be done in the ring Λ
Q of symmetric functions with rational coefficients: :\begin{align} e_1 &= p_1,\\ e_2 &= \textstyle\frac12p_1^2 - \frac12p_2 &&= \textstyle\frac12 ( p_1^2 - p_2 ),\\ e_3 &= \textstyle\frac16p_1^3 - \frac12p_1 p_2 + \frac13p_3 &&= \textstyle\frac{1}{6} ( p_1^3 - 3 p_1 p_2 + 2 p_3 ),\\ e_4 &= \textstyle\frac1{24}p_1^4 - \frac14p_1^2 p_2 + \frac18p_2^2 + \frac13p_1 p_3 - \frac14p_4 &&= \textstyle\frac1{24} ( p_1^4 - 6 p_1^2 p_2 + 3 p_2^2 + 8 p_1 p_3 - 6 p_4 ),\\ &~~\vdots\\ e_n &= (-1)^n \sum_{m_1 + 2m_2 + \cdots + nm_n = n \atop m_1 \ge 0, \ldots, m_n \ge 0} \prod_{i=1}^n \frac{(-p_i)^{m_i}}{m_i ! \, i^{m_i}} \\ \end{align} and so forth. The general formula can be conveniently expressed as :e_k = \frac{(-1)^k}{k!} B_{k}(- p_1, -1! \, p_2, - 2! \, p_3, \ldots, - (k-1)! \, p_k ), where the
Bn is the complete exponential
Bell polynomial. This expression also leads to the following identity for generating functions: : \sum_{k=0}^\infty e_k \,t^k = \exp\left(\sum_{k=1}^\infty \frac{(-1)^{k+1}}{k} p_k \,t^k \right). Applied to a monic polynomial, these formulae express the coefficients in terms of the power sums of the roots: replace each
ei by
ai and each
pk by
sk.
Expressing complete homogeneous symmetric polynomials in terms of power sums The analogous relations involving complete homogeneous symmetric polynomials can be similarly developed, giving equations :\begin{align} h_1 &= p_1,\\ h_2 &= \textstyle\frac12p_1^2 + \frac12p_2 &&= \textstyle\frac12 ( p_1^2 + p_2 ),\\ h_3 &= \textstyle\frac16p_1^3 + \frac12p_1 p_2 + \frac13p_3 &&= \textstyle\frac{1}{6} ( p_1^3 + 3 p_1 p_2 + 2 p_3 ),\\ h_4 &= \textstyle\frac1{24}p_1^4 + \frac14p_1^2 p_2 + \frac18p_2^2 + \frac13p_1 p_3 + \frac14p_4 &&= \textstyle\frac1{24} ( p_1^4 + 6 p_1^2 p_2 + 3 p_2^2 + 8 p_1 p_3 + 6 p_4 ),\\ &~~\vdots\\ h_k &= \sum_{m_1 + 2m_2 + \cdots + km_k = k \atop m_1\ge 0, \ldots, m_k\ge 0} \prod_{i=1}^k \frac{p_i^{m_i}}{m_i ! \, i^{m_i}} \end{align} and so forth, in which there are only plus signs. In terms of the complete Bell polynomial, :h_k = \frac{1}{k!} B_{k}(p_1, 0! \, p_2, 1! \, p_3, \ldots, (k-1)! \, p_k ). These expressions correspond exactly to the
cycle index polynomials of the
symmetric groups, if one interprets the power sums
pi as indeterminates: the coefficient in the expression for
hk of any monomial
p1
m1
p2
m2...''p'
l'm''
l is equal to the fraction of all permutations of
k that have
m1 fixed points,
m2 cycles of length 2, ..., and
ml cycles of length
l. Explicitly, this coefficient can be written as 1/N where N = \prod_{i=1}^l (m_i!\,i^{m_i}); this
N is the number permutations commuting with any given permutation of the given cycle type. The expressions for the elementary symmetric functions have coefficients with the same absolute value, but a sign equal to the sign of , namely (−1)
m2+
m4+.... It can be proved by considering the following inductive step: :\begin{align} mf(m; m_1, \ldots, m_n) &= f(m-1; m_1 - 1, \ldots, m_n) + \cdots + f(m - n; m_1, \ldots, m_n - 1) \\ m_1\prod_{i=1}^n \frac{1}{i^{m_i} m_i!} + \cdots + nm_n \prod_{i=1}^n \frac{1}{i^{m_i} m_i!} &= m\prod_{i=1}^n \frac{1}{i^{m_i} m_i!} \end{align} By analogy with the derivation of the generating function of the e_n, we can also obtain the generating function of the h_n, in terms of the power sums, as: : \sum_{k=0}^\infty h_k \,t^k = \exp\left(\sum_{k=1}^\infty \frac{p_k}{k} \,t^k \right). This generating function is thus the
plethystic exponential of p_1 t = (x_1 + \cdots + x_n)t.
Expressing power sums in terms of elementary symmetric polynomials One may also use Newton's identities to express power sums in terms of elementary symmetric polynomials, which does not introduce denominators: :\begin{align} p_1 &= e_1,\\ p_2 &= e_1^2 - 2 e_2,\\ p_3 &= e_1^3 - 3 e_2 e_1 + 3 e_3,\\ p_4 &= e_1^4 - 4 e_2 e_1^2 + 4 e_3 e_1 + 2 e_2^2 - 4 e_4,\\ p_5 &= e_1^5 - 5 e_2 e_1^3 + 5 e_3 e_1^2 + 5 e_2^2 e_1 - 5 e_4 e_1 - 5 e_3e_2 + 5 e_5,\\ p_6 &= e_1^6 - 6 e_2 e_1^4 + 6 e_3 e_1^3 + 9 e_2^2 e_1^2 - 6 e_4 e_1^2 - 12 e_3 e_2 e_1 + 6 e_5 e_1 - 2 e_2^3 + 3 e_3^2 + 6 e_4 e_2 - 6e_6. \end{align} The first four formulas were obtained by
Albert Girard in 1629 (thus before Newton). The general formula (for all positive integers
m) is: :p_m = (-1)^m m \sum_{r_1 + 2r_2 + \cdots + mr_m = m \atop r_1\ge 0, \ldots, r_m\ge 0} \frac{(r_1 + r_2 + \cdots + r_m - 1)!}{r_1!\,r_2! \cdots r_m!} \prod_{i=1}^m (-e_i)^{r_i}. This can be conveniently stated in terms of
ordinary Bell polynomials as :p_m = (-1)^m m \sum_{k=1}^m \frac{1}{k} \hat{B}_{m,k}(-e_1,\ldots,-e_{m-k+1}), or equivalently as the
generating function: : \begin{align} \sum_{k=1}^{\infty} (-1)^{k-1} p_k \frac{t^k}{k} & = \ln\left(1+e_1t+e_2t^2+e_3t^3+\cdots\right) \\ & = e_1 t - \frac{1}{2}\left(e_1^2-2e_2\right) t^2 + \frac{1}{3}\left(e_1^3-3e_1e_2+3e_3\right) t^3 + \cdots, \end{align} which is analogous to the
Bell polynomial exponential generating function given in the
previous subsection. The multiple summation formula above can be proved by considering the following inductive step: :\begin{align} f(m;\; r_1,\ldots,r_n) = {} & f(m - 1;\; r_1 - 1, \cdots, r_n) + \cdots + f(m - n;\; r_1, \ldots, r_n - 1)\\[8pt] = {} & \frac{1}{(r_1 - 1)!\cdots r_n!}(m - 1)(r_1 + \cdots + r_n - 2)! + \cdots \\ & \cdots + \frac{1}{r_1!\cdots(r_n - 1)!}(m - n)(r_1 + \cdots + r_n - 2)!\\[8pt] = {} & \frac{1}{r_1!\cdots r_n!}\left[r_1(m - 1) + \cdots + r_n(m-n)\right]\left[r_1 + \cdots + r_n - 2\right]!\\[8pt] = {} & \frac{1}{r_1!\cdots r_n!}\left[m(r_1 + \cdots + r_n) - m\right]\left[r_1 + \cdots + r_n - 2\right]!\\[8pt] = {} & \frac{m(r_1 + \cdots + r_n - 1)!}{r_1!\cdots r_n!} \end{align}
Expressing power sums in terms of complete homogeneous symmetric polynomials Finally one may use the variant identities involving complete homogeneous symmetric polynomials similarly to express power sums in term of them: :\begin{align} p_1 &= + h_1,\\ p_2 &= - h_1^2 + 2 h_2,\\ p_3 &= + h_1^3 - 3 h_2 h_1 + 3 h_3,\\ p_4 &= - h_1^4 + 4 h_2 h_1^2 - 4 h_3 h_1 - 2 h_2^2 + 4 h_4,\\ p_5 &= + h_1^5 - 5 h_2 h_1^3 + 5 h_2^2 h_1 + 5 h_3 h_1^2 - 5 h_3h_2 - 5 h_4 h_1 + 5 h_5,\\ p_6 &= - h_1^6 + 6 h_2 h_1^4 - 9 h_2^2 h_1^2 - 6 h_3 h_1^3 + 2 h_2^3 + 12 h_3 h_2 h_1 + 6 h_4 h_1^2 - 3 h_3^2 - 6 h_4 h_2 - 6 h_1 h_5 + 6h_6,\\ \end{align} and so on. Apart from the replacement of each
ei by the corresponding
hi, the only change with respect to the previous family of identities is in the signs of the terms, which in this case depend just on the number of factors present: the sign of the monomial \prod_{i=1}^l h_i^{m_i} is −(−1)
m1+
m2+
m3+.... In particular the above description of the absolute value of the coefficients applies here as well. The general formula (for all non-negative integers
m) is: :p_m = -\sum_{r_1 + 2r_2 + \cdots + mr_m = m \atop r_1\ge 0, \ldots, r_m \ge 0} \frac{m(r_1 + r_2 + \cdots + r_m - 1)!}{r_1!\,r_2!\cdots r_m!} \prod_{i=1}^m (-h_i)^{r_i}
Expressions as determinants One can obtain explicit formulas for the above expressions in the form of determinants, by considering the first
n of Newton's identities (or it counterparts for the complete homogeneous polynomials) as linear equations in which the elementary symmetric functions are known and the power sums are unknowns (or vice versa), and apply
Cramer's rule to find the solution for the final unknown. For instance taking Newton's identities in the form :\begin{align} e_1 &= 1p_1,\\ 2e_2 &= e_1p_1 - 1p_2,\\ 3e_3 &= e_2p_1 - e_1p_2 + 1p_3,\\ & \,\,\,\vdots \\ ne_n &= e_{n-1}p_1 - e_{n-2} p_2 + \cdots + (-1)^ne_1p_{n - 1} + (-1)^{n - 1}p_n \end{align} we consider p_1, -p_2, p_3, \ldots, (-1)^np_{n-1} and p_n as unknowns, and solve for the final one, giving :\begin{align} (-1)^{n-1}p_n ={} &\begin{vmatrix} 1 & 0 & \cdots & & e_1 \\ e_1 & 1 & 0 & \cdots & 2e_2 \\ e_2 & e_1 & 1 & & 3e_3 \\ \vdots & & \ddots & \ddots & \vdots\\ e_{n - 1} & \cdots & e_2 & e_1 & ne_n \end{vmatrix}\begin{vmatrix} 1 & 0 & \cdots & \\ e_1 & 1 & 0 & \cdots \\ e_2 & e_1 & 1 & \\ \vdots & & \ddots & \ddots \\ e_{n - 1} & \cdots & e_2 & e_1 & 1 \end{vmatrix}^{-1} \\[7pt] p_n = {(-1)^{n-1}} &\begin{vmatrix} 1 & 0 & \cdots & & e_1 \\ e_1 & 1 & 0 & \cdots & 2e_2 \\ e_2 & e_1 & 1 & & 3e_3 \\ \vdots & & \ddots &\ddots & \vdots \\ e_{n - 1} & \cdots & e_2 & e_1 & ne_n \end{vmatrix} \\[7pt] ={} &\begin{vmatrix} e_1 & 1 & 0 & \cdots \\ 2e_2 & e_1 & 1 & 0 & \cdots \\ 3e_3 & e_2 & e_1 & 1 \\ \vdots & & & \ddots & \ddots \\ ne_n & e_{n - 1} & \cdots & & e_1 \end{vmatrix}. \end{align} Solving for e_n instead of for p_n is similar, as the analogous computations for the complete homogeneous symmetric polynomials; in each case the details are slightly messier than the final results, which are (Macdonald 1979, p. 20): :\begin{align} e_n = \frac1{n!}&\begin{vmatrix} p_1 & 1 & 0 & \cdots \\ p_2 & p_1 & 2 & 0 & \cdots \\ \vdots & & \ddots & \ddots \\ p_{n-1} & p_{n-2} & \cdots & p_1 & n-1 \\ p_n & p_{n-1} & \cdots & p_2 & p_1 \end{vmatrix} \\[7pt] p_n = (-1)^{n-1}&\begin{vmatrix} h_1 & 1 & 0 & \cdots \\ 2h_2 & h_1 & 1 & 0 & \cdots \\ 3h_3 & h_2 & h_1 & 1 \\ \vdots & & & \ddots & \ddots \\ nh_n & h_{n-1} & \cdots & & h_1 \end{vmatrix} \\[7pt] h_n = \frac1{n!}&\begin{vmatrix} p_1 & -1 & 0 & \cdots \\ p_2 & p_1 & -2 & 0 & \cdots \\ \vdots & & \ddots & \ddots \\ p_{n - 1} & p_{n-2} & \cdots & p_1 & 1 - n \\ p_n & p_{n-1} & \cdots & p_2 & p_1 \end{vmatrix}. \end{align} Note that the use of determinants makes that the formula for h_n has additional minus signs compared to the one for e_n, while the situation for the expanded form given earlier is opposite. As remarked in (Littlewood 1950, p. 84) one can alternatively obtain the formula for h_n by taking the
permanent of the matrix for e_n instead of the determinant, and more generally an expression for any
Schur polynomial can be obtained by taking the corresponding
immanant of this matrix. == Derivation of the identities ==