[net.math] Integration by Elementary Functions

weemba@brahms.BERKELEY.EDU (Matthew P. Wiener) (03/09/86)

>Thanks to all of you who answered my earlier query. Does anyone
>know how to prove that the integral of e**x*sec(x) is not
>expressible in terms of elementary functions?

We give a fairly complete sketch of the proof that certain functions,
including the asked for one, are not integrable in elementary terms.  The
central theorem is due to Liouville in 1835.  His proof was analytic.  The
sketch below is mostly algebraic and is due to Maxwell Rosenlicht.  See his
papers in the _Pacific Journal of Mathematics_, 54, (1968) 153-161 and 65,
(1976), 485-492.

WARNING: Prerequisites for understanding the proof is a first year graduate
course in algebra, and a little complex analysis.  No deep results are used,
but I cannot take the time to explain standard notions or all the deductions.

Notation: a^n is "a power n", a_n is "a sub n".  C is the complex numbers,
for fields F, F[x] is the ring of polynomials in x OR an algebraic extension
of F, F(x) is the field of rational functions in a transcendental x, M is
the field of meromorphic functions in one variable.  If f is a complex
function, I(f) will denote an antiderivative of f.

A differential field is a field F of characteristic 0 with a derivation.
Thus, in addition to the field operations + and *, there is a derivative
mapping ':F->F such that (a+b)'=a'+b' and (ab)'=a'b+ab'.  Two standard
examples are C(z) and M with the usual derivative map.  Notice a basic
identity (logarithmic differentiation) holds:

   [(a_1 ^ k_1) * ... * (a_n ^ k_n)]'      a_1'             a_n'
   --------------------------------- = k_1 ---  + ... + k_n ---
    (a_1 ^ k_1) * ... * (a_n ^ k_n)        a_1              a_n

The usual rules like the quotient rule also hold.  If a in F satisfies
a'=0, we call a a constant of F.  The set of constants of F is called
Con(F), and forms a subfield of F.

The basic idea in showing something has no elementary integral is to
reduce the problem to a sequence of differential fields F_0, F_1, etc.,
where F_0 = C(z), and F_(i+1) is obtained from F_i by adjoining one
new element t.  t is obtained either algebraically, because t satisfies
some polynomial equation p(t)=0, or exponentially, because t'/t=s' for
some s in F_i, or logarithmically, because t'=s'/s is in F_i.  Notice
that we don't actually take exponentials or logarithms, but only attach
abstract elements that have the appropriate derivatives.  Thus a function
f is integrable in elementary terms iff such a sequence exists starting
with C(z).

Just so there is no confusion, there is no notion of "composition" involved
here.  If you want to take log s, you adjoin a transcendental t with the
relation t'=s'/s.  There is no log function running around, for example,
except as motivation, until we reach actual examples.

We need some easy lemmas.  Throughout the lemmas F is a differential field,
and t is transcendental over F.

Lemma 1: If K is an algebraic extension field of F, then there exists a
unique way to extend the derivation map from F to K so as to make K into
a differential field.

Lemma 2: If K=F(t) is a differential field with derivation extending F's,
and t' is in F, then for any polynomial f(t) in F[t], f(t)' is a polynomial
in F[t] of the same degree (if the leading coefficient is not in Con(F))
or of degree one less (if the leading coefficient is in Con(F)).

Lemma 3: If K=F(t) is a differential field with derivation extending F's,
and t'/t is in F, then for any a in F, n a positive integer, there exists
h in F such that (a*t^n)'=h*t^n.  More generally, if f(t) is any polynomial
in F[t], then f(t)' is of the same degree as f(t), and is a multiple of
f(t) iff f(t) is a monomial.

These are all fairly elementary.  For example, (a*t^n)'=(a'+at'/t)*t^n
in lemma 3.  The final 'iff' in lemma 3 is where transcendence of t comes
in.  Lemma 1 in the usual case of subfields of M can be proven analytically
using the implicit function theorem.
--------------------------------------------------------------------------
MAIN THEOREM.  Let F,G be differential fields, let a be in F, let y be in G,
and suppose y'=a and G is an elementary differential extension field of F,
and Con(F)=Con(G).  Then there exist c_1,...,c_n in Con(F), u_1,...,u_n, v
in F such that
                        u_1'            u_n'
               a =  c_1 --- + ... + c_n --- + v'.
                        u_1             u_n

In other words, the only functions that have elementary anti-derivatives
are the ones that have this very specific form.

This is a very useful theorem for proving non-integrability.  In the usual
case, F,G are subfields of M, so Con(F)=Con(G)=C always holds.

Proof:
By assumption there exists a finite chain of fields connecting F to G
such that the extension from one field to the next is given by performing
an algebraic, logarithmic, or exponential extension.  We show that if the
form (*) can be satisfied with values in F2, and F2 is one of the three
kinds of allowable extensions of F1, then the form (*) can be satisfied
in F1.  The form (*) is obviously satisfied in G: let all the c's be 0, the
u's be 1, and let v be the original y for which y'=a.  Thus, if the form
(*) can be pulled down one field, we will be able to pull it down to F,
and the theorem holds.

So we may assume without loss of generality that G=F(t).

Case 1: t is algebraic over F.  Say t is of degree k.  Then there are
polynomials U_i and V such that U_i(t)=u_i and V(t)=v.  So we have

                        U_1(t)'            U_n(t)'
               a =  c_1 ------ + ... + c_n ------ + V(t)'.
                        U_1(t)             U_n(t)

Now, by the uniqueness of extensions of derivatives in the algebraic case,
we may replace t by any of its conjugates t_1,..., t_k, and the same equation
holds.  In other words, because a is in F, it is fixed under the Galois
automorphisms.  Summing up over the conjugates, and converting the U'/U
terms into products using logarithmic differentiation, we have

            [U_1(t_1)*...*U_1(t_k)]'
 k a =  c_1 ----------------------- + ...  + [V(t_1)+...+V(t_k)]'.
             U_1(t_1)*...*U_n(t_k)

But the expressions in [...] are symmetric polynomials in t_i, and as
they are polynomials with coefficients in F, the resulting expressions
are in F.  So dividing by k gives us (*) holding in F.

Case 2: t is logarithmic over F.  Because of logarithmic differentiation
we may assume that the u's are monic and irreducible in t and distinct.
Furthermore, we may assume v has been decomposed into partial fractions.
The fractions can only be of the form f/g^j, where deg(f)<def(g) and g
is monic irreducible.  The fact that no terms outside of F appear on the
left hand side of (*), namely just a appears, means a lot of cancellation
must be occuring.

Let t'=s'/s, for some s in F.  If f(t) is monic in F[t], then f(t)' is also
in F[t], of one less degree.  Thus f(t) does not divide f(t)'.  In particular,
all the u'/u terms are in lowest terms already.  In the f/g^j terms in v,
we have a g^(j+1) denominator contribution in v' of the form -jfg'/g^(j+1).
But g doesn't divide fg', so no cancellation occurs.  But no u'/u term can
cancel, as the u's are irreducible, and no (xx)/g^(j+1) term appears in
a, because a is a member of F.  Thus no f/g^j term occurs at all in v.  But
then none of the u's can be outside of F, since nothing can cancel them.
(Remember the u's are distinct, monic, and irreducible.)  Thus each of the
u's is in F already, and v is a polynomial.  But v' = a - expression in u's,
so v' is in F also.  Thus v = b t + c for some b in con(F), c in F, by lemma
2.  Then

                   u_1'            u_n'     s'
          a =  c_1 --- + ... + c_n --- + b --- + c'
                   u_1             u_n      s

is the desired form.  So case 2 holds.

Case 3: t is exponential over F.  So let t'/t=s' for some s in F.  As in
case 2 above, we may assume all the u's are monic, irreducible, and distinct
and put v in partial fraction decomposition form.  Indeed the argument is
identical as in case 2 until we try to conclude what form v is.  Here lemma
3 tells us that v is a finite sum of terms b*t^j where each coefficient is
in F.  Each of the u's is also in F, with the possible exception that one
of them may be t.  Thus every u'/u term is in F, so again we conclude v'
is in F.  By lemma 3, v is in F.  So if every u is in F, a is in the desired
form.  Otherwise, one of the u's, say u_n, is actually t, then 

                        u_1'
               a =  c_1 --- + ... + (c_n s + v)'
                        u_1

is the desired form.  So case 3 holds.
------------------------------------------------------------------QED------
This proof, by the way, is a LOT easier than it looks.  Just work out some
examples, and you'll see what's going on.  (If this were a real expository
paper, such examples would be provided.  Maybe it's better this way.  Indeed,
if anybody out there takes the time to work some out and post them, I would
be much obliged.)

So how to you actually go about using this theorem?  Suppose you want to
integrate f*exp(g) for f,g in C(z), g non zero.  [This isn't yet the asked
for problem.]  Let t=exp(g), so t'/t=g'.  Let F=C(z)(t), G=any differential
extension field containing an antiderivative of f*t.  [Note that t is in
fact transcendental over C(z): g is rational and non-zero, so it has a
pole (possibly at infinity) and so t has an essential singularity and can't
be algebraic over C(z).]  Is G an elementary extension?  If so, then

                        u_1'            u_n'
             f*t =  c_1 --- + ... + c_n --- + v'
                        u_1             u_n

where the c_i, u_i, and v are in F.  Now the left hand side can be viewed
as a polynomial in C(z)[t] with exactly one term.  We must identify the
coefficient of t in the right hand side and get an equation for f.  But
the first n terms can be factored until the u_i's are linear (using the
logarithmic differentiation identity to preserve the abstract from).  As
for the v' term, long divide and use partial fractions to conclude v is a
sum of monomials: if v had a linear denominator other than t, raised to
some power in its partial fraction decomposition, its derivative would be
one higher power, and so cannot be cancelled with anything from the u_i
terms.  (As in the proof.)  If w is the coefficient of t in v, we have
f=w'+wg' with w in C(z).  Solving this first order ODE, we find that
w=exp(-g)*I(f*exp(g)).  In other words, if an elementary antiderivative
can be found for f*exp(g), where f,g are rational functions, then it is of
the form w*exp(g) for some rational function w.  [Notice that the conclusion
would fail for g equal to 0!]

For example, consider f=1 and g=-z^2.  Now exp(z^2)*I(exp(-z^2)) has no
poles (except perhaps at infinity), so if it is a rational function, it
must be a polynomial.  So let (p(z)*exp(-z^2))'=exp(-z^2).  One quickly
verifies that p'-2zp=1.  But the only solution to that ODE is the error
function I(exp(-z^2)) itself (within an additive constant somewhere)!
And the error function is NOT a polynomial!  (Proof?  OK, for one thing,
its Taylor series obtained by termwise integration is infinite.  For
another, its derivative is an exponential.)

As an exercise, prove that I(exp(z)/z) is not elementary.  Conclude that
neither is I(exp(exp(z))) and I(1/log(z)).

For a slightly harder exercise, prove that I(sin(z)/z) is not elementary.
Conclude that neither is I(sin(exp(z))).

Finally, we settle the case I(exp(z)*sec(z)).

We let s=exp(iz) and t=exp(z).  We have

	 2 t           u_1'            u_n'
       -------  =  c_1 --- + ... + c_n --- + v'
       s + 1/s         u_1             u_n
                  
where the u's and v are in C(z,s,t).  By reasoning as in the application
above, we equate t coefficients and get 

                    2
                 -------  =  w' + w,
                 s + 1/s   

where w is in C(z,s).  [To see that t is transcendental over C(z,s), it
suffices to show p(cos(z)) is not rational in z and exp(z) for any non-
trivial polynomial p.  But p(cos(z)) oscillates as z approaches infinity
along the real line, whereas any member of C(z,t) approaches a definite,
possibly infinite, limit at infinity along the real line.]

To see that this is impossible, expand w as a Laurent series in C(z)(s).
That is, express w as a sum of terms w_j*s^j, where j ranges through the
integers, and w_j is in C(z).  But 2/(s+1/s)=2*s/(1+s^2)=2-2*s^3+2*s^5-....
So we can equate coefficients.  In particular, we get the differential
equations w_j'+(1+i*n)*w_j=0 or 2 or -2.  But such ODE's don't have
rational function solutions!
---------------------------------------------------------------------------
Liouville's theorem can be generalized.  For example, I(J_0) is not
expressible using elementary functions and any of the Bessel functions.
The hard part is finding the appropriate normal form.  See Watson _A
Treatise on Bessel Functions_.

In G H Hardy _The Integration of Functions in Finite Terms_ the question
of how useful Liouville's theorem in finding antiderivatives was studied.
The general decision problem raised some subtle algebraic geometry questions
which were not settled until 1970 by R Risch, based on earlier work of J F
Ritt.  See J H Davenport _On the Integration of Algebraic Functions_ for a
discussion and further references.

ucbvax!brahms!weemba	Matthew P Wiener/UCB Math Dept/Berkeley CA 94720