SlideShare a Scribd company logo
Chapter 2

The Euler-Lagrange equation

In this chapter, we will give necessary conditions for an extremum of a function of the type
                                                   b
                                   I(x) =              F (x(t), x (t), t) dt,
                                               a

with various types of boundary conditions. The necessary condition is in the form of a differential
equation that the extremal curve should satisfy, and this differential equation is called the Euler-
Lagrange equation.

   We begin with the simplest type of boundary conditions, where the curves are allowed to vary
between two fixed points.



2.1     The simplest optimisation problem

The simplest optimisation problem can be formulated as follows:

Let F (α, β, γ) be a function with continuous first and second partial derivatives with respect to
(α, β, γ). Then find x ∈ C 1 [a, b] such that x(a) = ya and x(b) = yb , and which is an extremum for
the function
                                                   b
                                   I(x) =              F (x(t), x (t), t) dt.                        (2.1)
                                               a
In other words, the simplest optimisation problem consists of finding an extremum of a function
of the form (2.5), where the class of admissible curves comprises all smooth curves joining two
fixed points; see Figure 2.1. We will apply the necessary condition for an extremum (established
in Theorem 1.4.2) to the solve the simplest optimisation problem described above.

Theorem 2.1.1 Let S = {x ∈ C 1 [a, b] | x(a) = ya and x(b) = yb }, and let I : S → R be a
function of the form
                                                   b
                                   I(x) =              F (x(t), x (t), t) dt.
                                               a
If I has an extremum at x0 ∈ S, then x0 satisfies the Euler-Lagrange equation:
                 ∂F                       d        ∂F
                    (x0 (t), x0 (t), t) −             (x0 (t), x0 (t), t)       = 0,   t ∈ [a, b].   (2.2)
                 ∂α                       dt       ∂β

                                                          27
28                                                                                                   Chapter 2. The Euler-Lagrange equation



                                                    yb

             PSfrag replacements


                                                 ya
                                                                                                         t
                                                                a                                b

              Figure 2.1: Possible paths joining the two fixed points (a, ya ) and (b, yb ).

Proof The proof is long and so we divide it into several steps.

Step 1. First of all we note that the set S is not a vector space (unless ya = 0 = yb )! So Theorem
1.4.2 is not applicable directly. Hence we introduce a new linear space X, and consider a new
          ˜
function I : X → R which is defined in terms of the old function I.

     Introduce the linear space
                                            X = {x ∈ C 1 [a, b] | x(a) = x(b) = 0},
with the induced norm from C 1 [a, b]. Then for all h ∈ X, x0 + h satisfies (x0 + h)(a) = ya and
                             ˜                                           ˜
(x0 + h)(b) = yb . Defining I(h) = I(x0 + h), for h ∈ X, we note that I : X → R has a local
                                                   1
extremum at 0. It follows from Theorem 1.4.2 that DI(0)˜ = 0.

                           ˜
Step 2. We now calculate D I(0). We have
                                            b                                                                       b
           ˜      ˜
           I(h) − I(0) =                        F ((x0 + h)(t), (x0 + h) (t), t) dt −                                   F (x0 (t), x0 (t), t) dt
                                        a                                                                       a
                                            b
                             =                  [F (x0 (t) + h(t), x0 (t) + h (t), t) dt − F (x0 (t), x0 (t), t)] dt.
                                        a

Recall that from Taylor’s theorem, if F possesses partial derivatives of order 2 in a ball B of radius
r around the point (α0 , β0 , γ0 ) in R3 , then for all (α, β, γ) ∈ B, there exists a Θ ∈ [0, 1] such that
                                                                             ∂              ∂              ∂
F (α, β, γ) = F (α0 , β0 , γ0 ) +                        (α − α0 )             + (β − β0 )    + (γ − γ0 )                        F                   +
                                                                            ∂α             ∂β             ∂γ                         (α0 ,β0 ,γ0 )
                                                                                                        2
                   1                         ∂              ∂             ∂
                            (α − α0 )          + (β − β0 )    + (γ − γ0 )                                   F                                                .
                   2!                       ∂α             ∂β             ∂γ
                                                                                                                    (α0 ,β0 ,γ0 )+Θ((α,β,γ)−(α0 ,β0 ,γ0 ))

Hence for h ∈ X such that h is small enough,
                                        b
         ˜      ˜                               ∂F                            ∂F
         I(h) − I(0) =                             (x0 (t), x0 (t), t) h(t) +    (x0 (t), x0 (t), t) h (t) dt +
                                    a           ∂α                            ∂β
                                                b                                        2
                                   1                             ∂          ∂
                                                         h(t)      + h (t)                   F                                                 dt.
                                   2!       a                   ∂α         ∂β                    (x0 (t)+Θ(t)h(t),x0 (t)+Θ(t)h (t),t)

It can be checked that there exists a M > 0 such that
                       b                                            2
              1                    ∂          ∂
                           h(t)      + h (t)                            F                                                   dt ≤ M h 2 ,
              2!   a              ∂α         ∂β                              (x0 (t)+Θ(t)h(t),x0 (t)+Θ(t)h (t),t)

  1 Note that by the 0 in the right hand side of the equality, we mean the zero map, namely the continuous linear

map from X to R, which is defined by h → 0 for all h ∈ X.
2.1. The simplest optimisation problem                                                                             29

        ˜
and so DI(0) is the map
                                 b
                                         ∂F                            ∂F
                 h→                         (x0 (t), x0 (t), t) h(t) +    (x0 (t), x0 (t), t) h (t) dt.           (2.3)
                             a           ∂α                            ∂β


Step 3. Next we show that if the map in (2.3) is the zero map, then this implies that (2.2) holds.
Define
                                       t
                                         ∂F
                              A(t) =         (x0 (τ ), x0 (τ ), τ ) dτ.
                                      a ∂α

Integrating by parts, we find that
                                         b                                           b
                                             ∂F
                                                (x0 (t), x0 (t), t) h(t)dt = −           A(t)h (t)dt,
                                     a       ∂α                                  a

                                     ˜
and so from (2.3), it follows that D I(0) = 0 implies that
                         b
                                                  ∂F
                             −A(t) +                 (x0 (t), x0 (t), t) h (t)dt = 0 for all h ∈ X.
                     a                            ∂β


Step 4. Finally, using Lemma 1.4.5, we obtain

                                                   ∂F
                                 −A(t) +              (x0 (t), x0 (t), t) = k for all t ∈ [a, b].
                                                   ∂β

Differentiating with respect to t, we obtain (2.3). This completes the proof of Theorem 2.1.1.

    Note that the Euler-Lagrange equation is only a necessary condition for the existence of an
extremum (see the remark following Theorem 1.4.2). However, in many cases, the Euler-Lagrange
equation by itself is enough to give a complete solution of the problem. In fact, the existence of
an extremum is sometimes clear from the context of the problem. If in such scenarios, there exists
only one solution to the Euler-Lagrange equation, then this solution must a fortiori be the point
for which the extremum is achieved.

Example. Let S = {x ∈ C 1 [0, 1] | x(0) = 0 and x(1) = 1}. Consider the function I : S → R
given by
                                          1             2
                                             d
                               I(x) =          x(t) − 1 dt.
                                        0   dt
We wish to find x0 ∈ S that minimizes I. We proceed as follows:

                                                                    ∂F         ∂F
Step 1. We have F (α, β, γ) = (β − 1)2 , and so                        = 0 and    = 2(β − 1).
                                                                    ∂α         ∂β

Step 2. The Euler-Lagrange equation (2.2) is now given by

                                                d
                                         0−        (2(x0 (t) − 1)) = 0     for all t ∈ [0, 1].
                                                dt


                                                                                                          C
Step 3. Integrating , we obtain 2(x0 (t) − 1) = C, for some constant C, and so x0 =                       2   + 1 =: A.
Integrating again, we have x0 (t) = At + B, where A and B are suitable constants.
30                                                                                   Chapter 2. The Euler-Lagrange equation


Step 4. The constants A and B can be determined by using that fact that x0 ∈ S, and so
x0 (0) = 0 and x0 (a) = 1. Thus we have

                                                      A0 + B         = 0,
                                                      A1 + B         = 1,

which yield A = 1 and B = 0.

    So the unique solution x0 of the Euler-Lagrange equation in S is x0 (t) = t, t ∈ [0, 1]; see
Figure 2.2.

                                                  1
                   PSfrag replacements
                                                         x0


                                                                                 t
                                                  0                         1

                                             Figure 2.2: Minimizer for I.


     Now we argue that the solution x0 indeed minimizes I. Since (x (t) − 1)2 ≥ 0 for all t ∈ [0, 1],
it follows that I(x) ≥ 0 for all x ∈ C 1 [0, 1]. But
                                       1                                1                         1
                      I(x0 ) =             (x0 (t) − 1)2 dt =               (1 − 1)2 dt =             0dt = 0.
                                   0                                0                         0

As I(x) ≥ 0 = I(x0 ) for all x ∈ S, it follows that x0 minimizes I.                                                       ♦

Definition. The solutions of the Euler-Lagrange equation (2.3) are called critical curves.

    The Euler-Lagrange equation is in general a second order differential equation, but in some
special cases, it can be reduced to a first order differential equation or where its solution can be
obtained entirely by evaluating integrals. We indicate some special cases in Exercise 3 on page 31,
where in each instance, F is independent of one of its arguments.

Exercises.

  1. Let S = {x ∈ C 1 [0, 1] | x(0) = 0 = x(1)}. Consider the map I : S → R given by
                                                              1
                                                I(x) =            (x(t))3 dt,        x ∈ S.
                                                          0

     Using Theorem 2.1.1, find the critical curve x0 ∈ S for I. Does I have a local extremum at
     x0 ?
  2. Write the Euler-Lagrange equation when F is given by
      (a) F (α, β, γ) = sin β,
      (b) F (α, β, γ) = α3 β 3 ,
       (c) F (α, β, γ) = α2 − β 2 ,
      (d) F (α, β, γ) = 2γβ − β 2 + 3βα2 .
2.1. The simplest optimisation problem                                                            31


 3. Prove that:

     (a) If F (α, β, γ) does not depend on α, then the Euler-Lagrange equation becomes

                                                      ∂F
                                                         (x(t), x (t), t) = c,
                                                      ∂β

         where C is a constant.
     (b) If F does not depend on β, then the Euler-Lagrange equation becomes

                                                      ∂F
                                                         (x(t), x (t), t) = 0.
                                                      ∂α

     (c) If F does not depend on γ and if x0 is twice-differentiable in [a, b], then the Euler-
         Lagrange equation becomes

                                                                       ∂F
                                     F (x(t), x (t), t) − x (t)           (x(t), x (t), t) = C,
                                                                       ∂β

         where C is a constant.
                        d
         Hint: What is dt F (x(t), x (t), t) − x (t) ∂F (x(t), x (t), t) ?
                                                     ∂β

 4. Find the curve which has minimum length between (0, 0) and (1, 1).

 5. Let S = {x ∈ C 1 [0, 1] | x(0) = 0 and x(1) = 1}. Find critical curves in S for the functions
    I : S → R, where I is given by:
                       1
     (a) I(x) =            x (t)dt
                   0
                       1
     (b) I(x) =            x(t)x (t)dt
                   0
                       1
     (c) I(x) =            (x(t) + tx (t))dt
                   0

    for x ∈ S.

 6. Find critical curves for the function
                                                                2
                                                   I(x) =           t3 (x (t))2 dt
                                                            1

    where x ∈ C 1 [1, 2] with x(1) = 5 and x(2) = 2.

 7. Find critical curves for the function
                                                                2
                                                                     (x (t))3
                                                   I(x) =                     dt
                                                            1           t2

    where x ∈ C 1 [1, 2] with x(1) = 1 and x(2) = 7.

 8. Find critical curves for the function
                                               1
                                  I(x) =           2tx(t) − (x (t))2 + 3x (t)(x(t))2 dt
                                           0

    where x ∈ C 1 [0, 1] with x(0) = 0 and x(1) = −1.
32                                                                    Chapter 2. The Euler-Lagrange equation


  9. Find critical curves for the function
                                                   1
                                    I(x) =             2(x(t))3 + 3t2 x (t) dt
                                               0

      where x ∈ C 1 [0, 1] with x(0) = 0 and x(1) = 1. What if x(0) = 0 and x(1) = 2?

 10. Consider the copper mining company mentioned in the introduction. If future money is
     discounted continuously at a constant rate r, then we can assess the present value of profits
     from this mining operation by introducing a factor of e−rt in the integrand of (1.37). Suppose
     that α = 4, β = 1, r = 1 and P = 2. Find a critical mining operation x0 such that x0 (0) = 0
     and x0 (T ) = Q.



2.2     Calculus of variations: some classical problems

Optimisation problems of the type considered in the previous section were studied in various
special cases by many leading mathematicians in the past. These were often solved by various
techniques, and these gave rise to the branch of mathematics known as the ‘calculus of variations’.
The name comes from the fact that often the procedure involved the calculation of the ‘variation’
in the function I when its argument (which was typically a curve) was changed, and then passing
limits. In this section, we mention two classical problems, and indicate how these can be solved
using the Euler-Lagrange equation.


2.2.1    The brachistochrone problem

The calculus of variations originated from a problem posed by the Swiss mathematician Johann
Bernoulli (1667-1748). He required the form of the curve joining two fixed points A and B in a
vertical plane such that a body sliding down the curve (under gravity and no friction) travels from
A to B in minimum time. This problem does not have a trivial solution; the straight line from A
to B is not the solution (this is also intuitively clear, since if the slope is high at the beginning,
the body picks up a high velocity and so its plausible that the travel time could be reduced) and
it can be verified experimentally by sliding beads down wires in various shapes.
               PSfrag replacements (0, 0)
                                 A                               x0
                                                                      x

                                  gravity
                                                ¡£¢
                                                 ¢¤

                                      y0
                                           y                  B (x0 , y0 )

                             Figure 2.3: The brachistochrone problem.


    To pose the problem in mathematical terms, we introduce coordinates as shown in Figure 2.3,
so that A is the point (0, 0), and B corresponds to (x0 , y0 ). Assuming that the particle is released
                                               1
from rest at A, conservation of energy gives 2 mv 2 − mgy = 0, where we have taken the zero
potential energy level at y = 0, and where v denotes the speed of the particle. Thus the speed is
                     √
given by v = ds = 2gy, where s denotes arc length along the curve. From Figure 2.4, we see
               dt
                                                                              1
that an element of arc length, δs is given approximately by ((δx)2 + (δy)2 ) 2 .
2.2. Calculus of variations: some classical problems                                                                         33

                                    PSfrag replacements
                                                                         δs
                                                              δy

                                                                       δx

                                              Figure 2.4: Element of arc length.

          Hence the time of descent is given by
                                                                                                        2
                                                                                                  dx
                                                        ds     1                y0     1+         dy
                                        T =            √     =√                                             dy.
                                               curve     2gy    2g          0                 y
          Our problem is to find the path {x(y), y ∈ [0, y0 ]}, satisfying x(0) = 0 and x(y0 ) = x0 , which
          minimizes T , that is, to determine the minimizer for the function I : S → R, where
                                                                                        1
                                                             y0
                                               1                   1 + (x (y))2         2

                                       I(x) = √                                             dy,        x ∈ S,
                                                2g       0              y
                                                                                                                  1+β 2
          and S = {x ∈ C 1 [0, y0 ] | x(0) = 0 and x(y0 ) = x0 }. Here2 F (α, β, γ) =                               γ     is independent of
          α, and so the Euler-Lagrange equation becomes

                                                  d             x (y)        1
                                                                            √               = 0.
                                                 dy           1 + (x (y)) 2   y

          Integrating with respect to y, we obtain
                                                              x (y)          1
                                                                            √ = C,
                                                         1 + (x     (y))2     y
          where C is a constant. It can be shown that the general solution of this differential equation is
          given by
                                         1                 ˜            1
                                 x(Θ) =      (Θ − sin Θ) + C, y(Θ) =       (1 − cos Θ),
                                        2C 2                          2C 2
                  ˜
          where C is another constant. The constants are chosen so that the curve passes through the points
          (0, 0) and (x0 , y0 ).

                           (0, 0)
                                                                                                                                x



PSfrag replacements                                                                  (x0 , y0 )



                              y

                                     Figure 2.5: The cycloid through (0, 0) and (x0 , y0 ).


               This curve is known as a cycloid, and in fact it is the curve described by a point P in a circle
          that rolls without slipping on the x axis, in such a way that P passes through (x0 , y0 ); see Figure
          2.5.
             2 Strictly speaking, the F here does not satisfy the demands made in Theorem 2.1.1. Notwithstanding this fact,

          with some additional argument, the solution given here can be fully justified.
34                                                                    Chapter 2. The Euler-Lagrange equation


2.2.2     Minimum surface area of revolution

The problem of minimum surface area of revolution is to find among all the curves joining two
given points (x0 , y0 ) and (x1 , y1 ), the one which generates the surface of minimum area when
rotated about the x axis.

     The area of the surface of revolution generated by rotating the curve y about the x axis is
                                              x1
                                S(y) = 2π          y(x)    1 + (y (x))2 dx.
                                             x0

Since the integrand does not depend explicitly on x, the Euler-Lagrange equation is
                                                       ∂F
                          F (y(x), y (x), x) − y (x)      (y(x), y (x), x) = C,
                                                       ∂β
where C is a constant, that is,
                                                          (y )2
                                  y    1 + (y )2 − y                  = C.
                                                          1 + (y )2

Thus y = C     1 + (y )2 , and it can be shown that this differential equation has the general solution

                                                           x + C1
                                       y(x) = C cosh                   .                                (2.4)
                                                             C
This curve is called a catenary. The values of the arbitrary constants C and C 1 are determined
by the conditions y(x0 ) = y0 and y(x1 ) = y1 . It can be shown that the following three cases are
possible, depending on the positions of the points (x0 , y0 ) and (x1 , y1 ):

  1. If a single curve of the form (2.4) passes through the points (x0 , y0 ) and (x1 , y1 ), then this
     curve is the solution of the problem; see Figure 2.6.


                                  y1
                               y0
             PSfrag replacements



                                             x0              x1

                      Figure 2.6: The catenary through (x0 , y0 ) and (x1 , y1 ).


  2. If two critical curves can be drawn through the points (x0 , y0 ) and (x1 , y1 ), then one of the
     curves actually corresponds to the surface of revolution if minimum area, and the other does
     not.
  3. If there does not exist a curve of the form (2.4) passing through the points (x0 , y0 ) and
     (x1 , y1 ), then there is no surface in the class of smooth surfaces of revolution which achieves
     the minimum area. In fact, if the location of the two points is such that the distance between
     them is sufficiently large compared to their distances from the x axis, then the area of the
     surface consisting of two circles of radius y0 and y1 will be less than the area of any surface
     of revolution generated by a smooth curve passing through the points; see Figure 2.7.
2.3. Free boundary conditions                                                                                          35



            PSfrag replacements
                                   y1
                                   y0

                                             x0                                      x1

              Figure 2.7: The polygonal curve (x0 , y0 ) − (x0 , 0) − (x1 , 0) − (x1 , y1 ).


      This is intuitively expected: imagine a soap bubble between concentric rings which are
      being pulled apart. Initially we get a soap bubble between these rings, but if the distance
      separating the rings becomes too large, then the soap bubble breaks, leaving soap films on
      each of the two rings. This example shows that a critical curve need not always exist in the
      class of curves under consideration.



2.3     Free boundary conditions

Besides the simplest optimisation problem considered in the previous section, we now consider the
optimisation problem with free boundary conditions (see Figure 2.8).




                                                  a                                 b
                                                                                             t
            PSfrag replacements




                                 Figure 2.8: Free boundary conditions.


    Let F (α, β, γ) be a function with continuous first and second partial derivatives with respect
to (α, β, γ). Then find x ∈ C 1 [a, b] which is an extremum for the function
                                                          b
                                                                            dx
                                    I(x) =                    F     x(t),      (t), t dt.                             (2.5)
                                                      a                     dt


Theorem 2.3.1 Let I : C 1 [a, b] → R be a function of the form
                                         b
                                                                  dx
                          I(x) =             F    x(t),              (t), t dt,      x ∈ C 1 [a, b],
                                     a                            dt

where F (α, β, γ) is a function with continuous first and second partial derivatives with respect to
(α, β, γ). If I has a local extremum at x0 , then x0 satisfies the Euler-Lagrange equation:

              ∂F              dx0          d                  ∂F                dx0
                    x0 (t),       (t), t −                            x0 (t),       (t), t       = 0,   t ∈ [a, b],   (2.6)
              ∂α               dt          dt                 ∂β                 dt
36                                                                                Chapter 2. The Euler-Lagrange equation


together with the transversality conditions

              ∂F              dx0                                          ∂F                dx0
                    x0 (t),       (t), t              =0       and                 x0 (t),       (t), t         = 0.           (2.7)
              ∂β               dt               t=a                        ∂β                 dt          t=b


Proof

Step 1. We take X = C 1 [a, b] and compute DI(x0 ). Proceeding as in the proof of Theorem 2.1.1,
it is easy to see that
                                    b
                                        ∂F                            ∂F
             DI(x0 )(h) =                  (x0 (t), x0 (t), t) h(t) +    (x0 (t), x0 (t), t) h (t) dt,
                                a       ∂α                            ∂β

h ∈ C 1 [a, b]. Theorem 1.4.2 implies that this linear functional must be the zero map, that is,
(DI(x0 ))(h) = 0 for all h ∈ C 1 [a, b]. In particular, it is also zero for all h in C 1 [a, b] such that
h(a) = h(b) = 0. But recall that in Step 3 and Step 4 of the proof of Theorem 2.1.1, we proved
that if
                     b
                       ∂F                             ∂F
                           (x0 (t), x0 (t), t) h(t) +    (x0 (t), x0 (t), t) h (t) dt = 0           (2.8)
                   a   ∂α                             ∂β
for all h in C 1 [a, b] such that h(a) = h(b) = 0, then this implies that he Euler-Lagrange equation
(2.6) holds.

Step 2. Integration by parts in (2.8) now gives
                                    b
                                        ∂F                       d              ∂F
         DI(x0 )(h)    =                   (x0 (t), x0 (t), t) −                   (x0 (t), x0 (t), t)      h(t)dt +           (2.9)
                                a       ∂α                       dt             ∂β
                                                                     t=b
                         ∂F
                            (x0 (t), x0 (t), t) h(t)                                                                          (2.10)
                         ∂β                           t=a
                            ∂F                                  ∂F
                       = 0+      (x0 (t), x0 (t), t)     h(b) −    (x0 (t), x0 (t), t)                                h(a).
                            ∂β                       t=b        ∂β                                              t=a

The integral in (2.9) vanishes since we have shown in Step 1 above that (2.6) holds. Thus the
condition DI(x0 ) = 0 now takes the form

                   ∂F                                              ∂F
                      (x0 (t), x0 (t), t)                 h(b) −      (x0 (t), x0 (t), t)           h(a) = 0,
                   ∂β                               t=b            ∂β                         t=a

from which (2.7) follows, since h is arbitrary. This completes the proof.


Exercises.

   1. Find all curves y = y(x) which have minimum length between the lines x = 0 and the line
      x = 1.

   2. Find critical curves for the following function, when the values of x are free at the endpoints:
                                                1
                                                    1
                               I(x) =                 (x (t))2 + x(t)x (t) + x (t) + x(t) dt.
                                            0       2
2.3. Free boundary conditions                                                                                                  37


        Similarly, we can also consider the mixed case (see Figure 2.9), when one end of the curve is
    fixed, say x(a) = ya , and the other end is free. Then it can be shown that the curve x satisfies
    the Euler-Lagrange equation, the transversality condition
                                                ∂F
                                                   (x0 (t), x0 (t), t)         h(a) = 0
                                                ∂β                       t=a
    at the free end point, and x(a) = ya serves as the other boundary condition.



                                                                         yb                          ¢£




                      ya             ¡




PSfrag replacements                                          t                                                t
                                a                      b                         a                        b




                                                    Figure 2.9: Mixed cases.

       We can summarize the results by the following: critical curves for (2.5) satisfy the Euler-
    Lagrange equation (2.6) and moreover there holds
                                         ∂F
                                            (x0 (t), x0 (t), t) = 0 at the free end point.
                                         ∂β


    Exercises.

      1. Find the curve y = y(x) which has minimum length between (0, 0) and the line x = 1.
      2. The cost of a manufacturing process in an industry is described by the function I given by
                                                                1
                                                                    1
                                                  I(x) =              (x (t))2 + x(t) dt,
                                                            0       2
         for x ∈ C 1 [0, 1] with x(0) = 1.
          (a) If also x(1) = 0, find a critical curve x∗ for I, and find I(x∗ ).
          (b) If x(1) is not specified, find a critical curve x∗∗ for I, and find I(x∗∗ ).
          (c) Which of the values I(x∗ ) and I(x∗∗ ) found in parts above is larger? Explain why
              you would expect this, assuming that x∗ and x∗∗ in fact minimize I on the respective
              domains specified above.
      3. Find critical curves for the following functions:
                            π
                            2
          (a) I(x) =            (x(t))2 − (x (t))2 dt, x(0) = 0 and x                π
                                                                                     2    is free.
                        0
                            π
                            2
          (b) I(x) =            (x(t))2 − (x (t))2 dt, x(0) = 1 and x                π
                                                                                     2    is free.
                        0
                                                                                                                      1
      4. Determine the curves that maximize the function I : S → R, where I(x) =                                          cos(x (t))dt,
                                                                                                                  0
                                           1
         x ∈ S and S = {x ∈ C [0, 1] | x(0) = 0}. What are the curves that minimize I?
38                                                                                          Chapter 2. The Euler-Lagrange equation


2.4        Generalization

The results in this chapter can be generalized to the case when the integrand F is a function of
more than one independent variable: if we wish to find extremum values of the function
                                                  b
                                                                                        dx1              dxn
                     I(x1 , . . . , xn ) =            F       x1 (t), . . . , xn (t),       (t), . . . ,     (t), t dt,
                                              a                                          dt               dt
where F (α1 , . . . , αn , β1 , . . . , βn , γ) is a function with continuous partial derivatives of order ≤ 2,
and x1 , . . . , xn are continuously differentiable functions of the variable t, then following a similar
analysis as before, we obtain n Euler-Lagrange equations to be satisfied by the optimal curve, that
is,
∂F                                                           d ∂F
    (x1∗ (t), . . . , xn∗ (t), x1∗ (t), . . . , xn∗ (t), t)−        (x1∗ (t), . . . , xn∗ (t), x1∗ (t), . . . , xn∗ (t), t) = 0,
∂αk                                                          dt ∂βk
for t ∈ [a, b], k ∈ {1, . . . , n}. Also at any end point where xk is free,
                             ∂F                                            dx1∗              dxn∗
                                      x1∗ (t), . . . , xn∗ (t),                 (t), . . . ,      (t), t     = 0.
                             ∂βk                                            dt                dt


Exercise. Find critical curves of the function
                                                          π
                                                          2
                             I(x1 , x2 ) =                    (x1 (t))2 + (x2 (t))2 + 2x1 (t)x2 (t) dt
                                                      0
                                 π                                           π
such that x1 (0) = 0, x1         2   = 1, x2 (0) = 0, x2                     2   = 1.

Remark. Note that with the above result, we can also solve the problem of finding extremal
curves for a function of the type
                                                      b
                                                                           dx             dn x
                                     I(x) :               F    x(t),          (t), . . . , n (t), t dt,
                                                  a                        dt             dt
for over all (sufficiently differentiable) curves x defined on an interval [a, b], taking values in R.
Indeed, we may introduce the auxiliary functions
                                                              dx                      dn x
                      x1 (t) = x(t), x2 (t) =                    (t), . . . , xn (t) = n (t),                t ∈ [a, b],
                                                              dt                      dt
                                                                        ˜
and consider the problem of finding extremal curves for the new function I defined by
                                                                    b
                               ˜
                               I(x1 , . . . , xn ) =                    F (x1 (t), x2 (t), . . . , xn (t), t)dt.
                                                                a

Using the result mentioned in this section, we can then solve this problem. Note that we eliminated
high order derivatives at the price of converting the scalar function into a vector-valued function.
Since we can always do this, this is one of the reasons in fact for considering functions of the type
(1.28) where no high order derivatives occur.



2.5        Optimisation in function spaces versus that in Rn

To understand the basic ‘infinite-dimensionality’ in the problems of optimisation in function spaces,
it is interesting to see how they are related to the problems of the study of functions of n real
2.5. Optimisation in function spaces versus that in Rn                                                  39


variables. Thus, consider a function of the form
                                    b
                                                    dx
                     I(x) =             F   x(t),      (t), t dt,       x(a) = ya ,     x(b) = yb .
                                a                   dt

Here each curve x is assigned a certain number. To find a related function of the sort considered
in classical analysis, we may proceed as follows. Using the points

                                            a = t0 , t1 , . . . , tn , tn+1 = b,

we divide the interval [a, b] into n + 1 equal parts. Then we replace the curve {x(t), t ∈ [a, b]} by
the polygonal line joining the points

                              (t0 , ya ), (t1 , x(t1 )), . . . , (tn , x(tn )), (tn+1 , yb ),

and we approximate the function I at x by the sum
                                                       n
                                                                       xk − xk−1
                           In (x1 , . . . , xn ) =          F   xk ,             , t k hk ,           (2.11)
                                                                          hk
                                                      k=1

where xk = x(tk ) and hk = tk − tk−1 . Each polygonal line is uniquely determined by the ordinates
x1 , . . . , xn of its vertices (recall that x0 = ya and xn+1 = yb are fixed), and the sum (2.11) is
therefore a function of the n variables x1 , . . . , xn . Thus as an approximation, we can regard the
optimisation problem as the problem of finding the extrema of the function In (x1 , . . . , xn ).

    In solving optimisation problems in function spaces, Euler made extensive use of this ‘method
of finite differences’. By replacing smooth curves by polygonal lines, he reduced the problem of
finding extrema of a function to the problem of finding extrema of a function of n variables, and
then he obtained exact solutions by passing to the limit as n → ∞. In this sense, functions can
be regarded as ‘functions of infinitely many variables’ (that is, the infinitely many values of x(t)
at different points), and the calculus of variations can be regarded as the corresponding analog of
differential calculus of functions of n real variables.
40   Chapter 2. The Euler-Lagrange equation

More Related Content

PDF
2nd order ode applications
PPTX
Over view of Laplace Transform and its Properties
PPTX
Math*4 Laplace and Inverse laplace transform
PPTX
Linear dependence(ld) &linear independence(li)
PDF
Eigenvalues and eigenvectors
PDF
Beta gamma functions
PDF
Jacobians new
PPT
Fourier transform
2nd order ode applications
Over view of Laplace Transform and its Properties
Math*4 Laplace and Inverse laplace transform
Linear dependence(ld) &linear independence(li)
Eigenvalues and eigenvectors
Beta gamma functions
Jacobians new
Fourier transform

What's hot (20)

PPTX
Application of interpolation and finite difference
PPTX
Laplace transform: UNIT STEP FUNCTION, SECOND SHIFTING THEOREM, DIRAC DELTA F...
PPT
Vector calculus
PPTX
Rank nullity theorem
PPTX
Differential equations
PPTX
Laplace transform and its application
PPTX
Partial differential equations
PPT
Linear transformation.ppt
PPTX
Applications of differential equations
PPTX
Partial differential equation & its application.
PPT
Application of Gauss,Green and Stokes Theorem
PDF
Numerical Solution of Ordinary Differential Equations
PDF
applications of second order differential equations
PDF
Partial Differential Equation - Notes
PPTX
Chapter 2 laplace transform
PDF
FINITE DIFFERENCE MODELLING FOR HEAT TRANSFER PROBLEMS
PPT
1532 fourier series
PPT
5 introduction to quantum mechanics
PDF
tensor-decomposition
PPT
First order non-linear partial differential equation & its applications
Application of interpolation and finite difference
Laplace transform: UNIT STEP FUNCTION, SECOND SHIFTING THEOREM, DIRAC DELTA F...
Vector calculus
Rank nullity theorem
Differential equations
Laplace transform and its application
Partial differential equations
Linear transformation.ppt
Applications of differential equations
Partial differential equation & its application.
Application of Gauss,Green and Stokes Theorem
Numerical Solution of Ordinary Differential Equations
applications of second order differential equations
Partial Differential Equation - Notes
Chapter 2 laplace transform
FINITE DIFFERENCE MODELLING FOR HEAT TRANSFER PROBLEMS
1532 fourier series
5 introduction to quantum mechanics
tensor-decomposition
First order non-linear partial differential equation & its applications
Ad

Viewers also liked (20)

PPT
Lagrange's Theorem
PDF
Goldstein Solution chapter 6
PPT
15695 20080506135543
PDF
Lagrange
PPT
Basic fluids
PDF
Optimal control systems
PPT
Introduction to flight 1
PDF
Lesson 1
PPTX
Lagrange's theorem
PPT
Mekanika fluida 2 ok
PDF
Classical mechanics
PPT
Pe 3032 wk 1 introduction to control system march 04e
PDF
Solution of skill Assessment Control Systems Engineering By Norman S.Nise 6t...
PPTX
Computational fluid dynamics
PDF
Control Systems notes
PDF
8178001772 control
PPTX
control system
PDF
Control system(smarajit ghosh) By sol
PDF
CFD analysis of an Airfoil
PDF
[Goldstein herbert] classical_mechanics_solution_m(book_see.org)
Lagrange's Theorem
Goldstein Solution chapter 6
15695 20080506135543
Lagrange
Basic fluids
Optimal control systems
Introduction to flight 1
Lesson 1
Lagrange's theorem
Mekanika fluida 2 ok
Classical mechanics
Pe 3032 wk 1 introduction to control system march 04e
Solution of skill Assessment Control Systems Engineering By Norman S.Nise 6t...
Computational fluid dynamics
Control Systems notes
8178001772 control
control system
Control system(smarajit ghosh) By sol
CFD analysis of an Airfoil
[Goldstein herbert] classical_mechanics_solution_m(book_see.org)
Ad

Similar to Euler lagrange equation (20)

PDF
13.Partial_derivative_total_differential.pdf
PDF
Special second order non symmetric fitted method for singular
PDF
11.[95 103]solution of telegraph equation by modified of double sumudu transf...
PDF
Ism et chapter_3
PDF
Ism et chapter_3
PDF
Ism et chapter_3
PDF
Introduction to inverse problems
PDF
11.solution of linear and nonlinear partial differential equations using mixt...
PDF
Solution of linear and nonlinear partial differential equations using mixture...
PDF
Notes 17
PDF
Axisymmetric finite element convergence results
PDF
Fourier series 1
PDF
Local Volatility 1
PDF
Den5200 ps1
PDF
11.[104 111]analytical solution for telegraph equation by modified of sumudu ...
PDF
L. Jonke - A Twisted Look on Kappa-Minkowski: U(1) Gauge Theory
DOC
Chapter 9(laplace transform)
PDF
Fourier series
PDF
Lesson 32: The Fundamental Theorem Of Calculus
PPT
Chain rule
13.Partial_derivative_total_differential.pdf
Special second order non symmetric fitted method for singular
11.[95 103]solution of telegraph equation by modified of double sumudu transf...
Ism et chapter_3
Ism et chapter_3
Ism et chapter_3
Introduction to inverse problems
11.solution of linear and nonlinear partial differential equations using mixt...
Solution of linear and nonlinear partial differential equations using mixture...
Notes 17
Axisymmetric finite element convergence results
Fourier series 1
Local Volatility 1
Den5200 ps1
11.[104 111]analytical solution for telegraph equation by modified of sumudu ...
L. Jonke - A Twisted Look on Kappa-Minkowski: U(1) Gauge Theory
Chapter 9(laplace transform)
Fourier series
Lesson 32: The Fundamental Theorem Of Calculus
Chain rule

Recently uploaded (20)

PDF
ANTIBIOTICS.pptx.pdf………………… xxxxxxxxxxxxx
PDF
Module 4: Burden of Disease Tutorial Slides S2 2025
PDF
Abdominal Access Techniques with Prof. Dr. R K Mishra
PDF
Insiders guide to clinical Medicine.pdf
PDF
2.FourierTransform-ShortQuestionswithAnswers.pdf
PPTX
IMMUNITY IMMUNITY refers to protection against infection, and the immune syst...
PDF
O7-L3 Supply Chain Operations - ICLT Program
PPTX
Renaissance Architecture: A Journey from Faith to Humanism
PDF
Black Hat USA 2025 - Micro ICS Summit - ICS/OT Threat Landscape
PDF
Computing-Curriculum for Schools in Ghana
PDF
The Lost Whites of Pakistan by Jahanzaib Mughal.pdf
PDF
Pre independence Education in Inndia.pdf
PDF
STATICS OF THE RIGID BODIES Hibbelers.pdf
PPTX
Institutional Correction lecture only . . .
PDF
01-Introduction-to-Information-Management.pdf
PDF
Physiotherapy_for_Respiratory_and_Cardiac_Problems WEBBER.pdf
PDF
FourierSeries-QuestionsWithAnswers(Part-A).pdf
PDF
Chapter 2 Heredity, Prenatal Development, and Birth.pdf
PDF
TR - Agricultural Crops Production NC III.pdf
PDF
3rd Neelam Sanjeevareddy Memorial Lecture.pdf
ANTIBIOTICS.pptx.pdf………………… xxxxxxxxxxxxx
Module 4: Burden of Disease Tutorial Slides S2 2025
Abdominal Access Techniques with Prof. Dr. R K Mishra
Insiders guide to clinical Medicine.pdf
2.FourierTransform-ShortQuestionswithAnswers.pdf
IMMUNITY IMMUNITY refers to protection against infection, and the immune syst...
O7-L3 Supply Chain Operations - ICLT Program
Renaissance Architecture: A Journey from Faith to Humanism
Black Hat USA 2025 - Micro ICS Summit - ICS/OT Threat Landscape
Computing-Curriculum for Schools in Ghana
The Lost Whites of Pakistan by Jahanzaib Mughal.pdf
Pre independence Education in Inndia.pdf
STATICS OF THE RIGID BODIES Hibbelers.pdf
Institutional Correction lecture only . . .
01-Introduction-to-Information-Management.pdf
Physiotherapy_for_Respiratory_and_Cardiac_Problems WEBBER.pdf
FourierSeries-QuestionsWithAnswers(Part-A).pdf
Chapter 2 Heredity, Prenatal Development, and Birth.pdf
TR - Agricultural Crops Production NC III.pdf
3rd Neelam Sanjeevareddy Memorial Lecture.pdf

Euler lagrange equation

  • 1. Chapter 2 The Euler-Lagrange equation In this chapter, we will give necessary conditions for an extremum of a function of the type b I(x) = F (x(t), x (t), t) dt, a with various types of boundary conditions. The necessary condition is in the form of a differential equation that the extremal curve should satisfy, and this differential equation is called the Euler- Lagrange equation. We begin with the simplest type of boundary conditions, where the curves are allowed to vary between two fixed points. 2.1 The simplest optimisation problem The simplest optimisation problem can be formulated as follows: Let F (α, β, γ) be a function with continuous first and second partial derivatives with respect to (α, β, γ). Then find x ∈ C 1 [a, b] such that x(a) = ya and x(b) = yb , and which is an extremum for the function b I(x) = F (x(t), x (t), t) dt. (2.1) a In other words, the simplest optimisation problem consists of finding an extremum of a function of the form (2.5), where the class of admissible curves comprises all smooth curves joining two fixed points; see Figure 2.1. We will apply the necessary condition for an extremum (established in Theorem 1.4.2) to the solve the simplest optimisation problem described above. Theorem 2.1.1 Let S = {x ∈ C 1 [a, b] | x(a) = ya and x(b) = yb }, and let I : S → R be a function of the form b I(x) = F (x(t), x (t), t) dt. a If I has an extremum at x0 ∈ S, then x0 satisfies the Euler-Lagrange equation: ∂F d ∂F (x0 (t), x0 (t), t) − (x0 (t), x0 (t), t) = 0, t ∈ [a, b]. (2.2) ∂α dt ∂β 27
  • 2. 28 Chapter 2. The Euler-Lagrange equation yb PSfrag replacements ya t a b Figure 2.1: Possible paths joining the two fixed points (a, ya ) and (b, yb ). Proof The proof is long and so we divide it into several steps. Step 1. First of all we note that the set S is not a vector space (unless ya = 0 = yb )! So Theorem 1.4.2 is not applicable directly. Hence we introduce a new linear space X, and consider a new ˜ function I : X → R which is defined in terms of the old function I. Introduce the linear space X = {x ∈ C 1 [a, b] | x(a) = x(b) = 0}, with the induced norm from C 1 [a, b]. Then for all h ∈ X, x0 + h satisfies (x0 + h)(a) = ya and ˜ ˜ (x0 + h)(b) = yb . Defining I(h) = I(x0 + h), for h ∈ X, we note that I : X → R has a local 1 extremum at 0. It follows from Theorem 1.4.2 that DI(0)˜ = 0. ˜ Step 2. We now calculate D I(0). We have b b ˜ ˜ I(h) − I(0) = F ((x0 + h)(t), (x0 + h) (t), t) dt − F (x0 (t), x0 (t), t) dt a a b = [F (x0 (t) + h(t), x0 (t) + h (t), t) dt − F (x0 (t), x0 (t), t)] dt. a Recall that from Taylor’s theorem, if F possesses partial derivatives of order 2 in a ball B of radius r around the point (α0 , β0 , γ0 ) in R3 , then for all (α, β, γ) ∈ B, there exists a Θ ∈ [0, 1] such that ∂ ∂ ∂ F (α, β, γ) = F (α0 , β0 , γ0 ) + (α − α0 ) + (β − β0 ) + (γ − γ0 ) F + ∂α ∂β ∂γ (α0 ,β0 ,γ0 ) 2 1 ∂ ∂ ∂ (α − α0 ) + (β − β0 ) + (γ − γ0 ) F . 2! ∂α ∂β ∂γ (α0 ,β0 ,γ0 )+Θ((α,β,γ)−(α0 ,β0 ,γ0 )) Hence for h ∈ X such that h is small enough, b ˜ ˜ ∂F ∂F I(h) − I(0) = (x0 (t), x0 (t), t) h(t) + (x0 (t), x0 (t), t) h (t) dt + a ∂α ∂β b 2 1 ∂ ∂ h(t) + h (t) F dt. 2! a ∂α ∂β (x0 (t)+Θ(t)h(t),x0 (t)+Θ(t)h (t),t) It can be checked that there exists a M > 0 such that b 2 1 ∂ ∂ h(t) + h (t) F dt ≤ M h 2 , 2! a ∂α ∂β (x0 (t)+Θ(t)h(t),x0 (t)+Θ(t)h (t),t) 1 Note that by the 0 in the right hand side of the equality, we mean the zero map, namely the continuous linear map from X to R, which is defined by h → 0 for all h ∈ X.
  • 3. 2.1. The simplest optimisation problem 29 ˜ and so DI(0) is the map b ∂F ∂F h→ (x0 (t), x0 (t), t) h(t) + (x0 (t), x0 (t), t) h (t) dt. (2.3) a ∂α ∂β Step 3. Next we show that if the map in (2.3) is the zero map, then this implies that (2.2) holds. Define t ∂F A(t) = (x0 (τ ), x0 (τ ), τ ) dτ. a ∂α Integrating by parts, we find that b b ∂F (x0 (t), x0 (t), t) h(t)dt = − A(t)h (t)dt, a ∂α a ˜ and so from (2.3), it follows that D I(0) = 0 implies that b ∂F −A(t) + (x0 (t), x0 (t), t) h (t)dt = 0 for all h ∈ X. a ∂β Step 4. Finally, using Lemma 1.4.5, we obtain ∂F −A(t) + (x0 (t), x0 (t), t) = k for all t ∈ [a, b]. ∂β Differentiating with respect to t, we obtain (2.3). This completes the proof of Theorem 2.1.1. Note that the Euler-Lagrange equation is only a necessary condition for the existence of an extremum (see the remark following Theorem 1.4.2). However, in many cases, the Euler-Lagrange equation by itself is enough to give a complete solution of the problem. In fact, the existence of an extremum is sometimes clear from the context of the problem. If in such scenarios, there exists only one solution to the Euler-Lagrange equation, then this solution must a fortiori be the point for which the extremum is achieved. Example. Let S = {x ∈ C 1 [0, 1] | x(0) = 0 and x(1) = 1}. Consider the function I : S → R given by 1 2 d I(x) = x(t) − 1 dt. 0 dt We wish to find x0 ∈ S that minimizes I. We proceed as follows: ∂F ∂F Step 1. We have F (α, β, γ) = (β − 1)2 , and so = 0 and = 2(β − 1). ∂α ∂β Step 2. The Euler-Lagrange equation (2.2) is now given by d 0− (2(x0 (t) − 1)) = 0 for all t ∈ [0, 1]. dt C Step 3. Integrating , we obtain 2(x0 (t) − 1) = C, for some constant C, and so x0 = 2 + 1 =: A. Integrating again, we have x0 (t) = At + B, where A and B are suitable constants.
  • 4. 30 Chapter 2. The Euler-Lagrange equation Step 4. The constants A and B can be determined by using that fact that x0 ∈ S, and so x0 (0) = 0 and x0 (a) = 1. Thus we have A0 + B = 0, A1 + B = 1, which yield A = 1 and B = 0. So the unique solution x0 of the Euler-Lagrange equation in S is x0 (t) = t, t ∈ [0, 1]; see Figure 2.2. 1 PSfrag replacements x0 t 0 1 Figure 2.2: Minimizer for I. Now we argue that the solution x0 indeed minimizes I. Since (x (t) − 1)2 ≥ 0 for all t ∈ [0, 1], it follows that I(x) ≥ 0 for all x ∈ C 1 [0, 1]. But 1 1 1 I(x0 ) = (x0 (t) − 1)2 dt = (1 − 1)2 dt = 0dt = 0. 0 0 0 As I(x) ≥ 0 = I(x0 ) for all x ∈ S, it follows that x0 minimizes I. ♦ Definition. The solutions of the Euler-Lagrange equation (2.3) are called critical curves. The Euler-Lagrange equation is in general a second order differential equation, but in some special cases, it can be reduced to a first order differential equation or where its solution can be obtained entirely by evaluating integrals. We indicate some special cases in Exercise 3 on page 31, where in each instance, F is independent of one of its arguments. Exercises. 1. Let S = {x ∈ C 1 [0, 1] | x(0) = 0 = x(1)}. Consider the map I : S → R given by 1 I(x) = (x(t))3 dt, x ∈ S. 0 Using Theorem 2.1.1, find the critical curve x0 ∈ S for I. Does I have a local extremum at x0 ? 2. Write the Euler-Lagrange equation when F is given by (a) F (α, β, γ) = sin β, (b) F (α, β, γ) = α3 β 3 , (c) F (α, β, γ) = α2 − β 2 , (d) F (α, β, γ) = 2γβ − β 2 + 3βα2 .
  • 5. 2.1. The simplest optimisation problem 31 3. Prove that: (a) If F (α, β, γ) does not depend on α, then the Euler-Lagrange equation becomes ∂F (x(t), x (t), t) = c, ∂β where C is a constant. (b) If F does not depend on β, then the Euler-Lagrange equation becomes ∂F (x(t), x (t), t) = 0. ∂α (c) If F does not depend on γ and if x0 is twice-differentiable in [a, b], then the Euler- Lagrange equation becomes ∂F F (x(t), x (t), t) − x (t) (x(t), x (t), t) = C, ∂β where C is a constant. d Hint: What is dt F (x(t), x (t), t) − x (t) ∂F (x(t), x (t), t) ? ∂β 4. Find the curve which has minimum length between (0, 0) and (1, 1). 5. Let S = {x ∈ C 1 [0, 1] | x(0) = 0 and x(1) = 1}. Find critical curves in S for the functions I : S → R, where I is given by: 1 (a) I(x) = x (t)dt 0 1 (b) I(x) = x(t)x (t)dt 0 1 (c) I(x) = (x(t) + tx (t))dt 0 for x ∈ S. 6. Find critical curves for the function 2 I(x) = t3 (x (t))2 dt 1 where x ∈ C 1 [1, 2] with x(1) = 5 and x(2) = 2. 7. Find critical curves for the function 2 (x (t))3 I(x) = dt 1 t2 where x ∈ C 1 [1, 2] with x(1) = 1 and x(2) = 7. 8. Find critical curves for the function 1 I(x) = 2tx(t) − (x (t))2 + 3x (t)(x(t))2 dt 0 where x ∈ C 1 [0, 1] with x(0) = 0 and x(1) = −1.
  • 6. 32 Chapter 2. The Euler-Lagrange equation 9. Find critical curves for the function 1 I(x) = 2(x(t))3 + 3t2 x (t) dt 0 where x ∈ C 1 [0, 1] with x(0) = 0 and x(1) = 1. What if x(0) = 0 and x(1) = 2? 10. Consider the copper mining company mentioned in the introduction. If future money is discounted continuously at a constant rate r, then we can assess the present value of profits from this mining operation by introducing a factor of e−rt in the integrand of (1.37). Suppose that α = 4, β = 1, r = 1 and P = 2. Find a critical mining operation x0 such that x0 (0) = 0 and x0 (T ) = Q. 2.2 Calculus of variations: some classical problems Optimisation problems of the type considered in the previous section were studied in various special cases by many leading mathematicians in the past. These were often solved by various techniques, and these gave rise to the branch of mathematics known as the ‘calculus of variations’. The name comes from the fact that often the procedure involved the calculation of the ‘variation’ in the function I when its argument (which was typically a curve) was changed, and then passing limits. In this section, we mention two classical problems, and indicate how these can be solved using the Euler-Lagrange equation. 2.2.1 The brachistochrone problem The calculus of variations originated from a problem posed by the Swiss mathematician Johann Bernoulli (1667-1748). He required the form of the curve joining two fixed points A and B in a vertical plane such that a body sliding down the curve (under gravity and no friction) travels from A to B in minimum time. This problem does not have a trivial solution; the straight line from A to B is not the solution (this is also intuitively clear, since if the slope is high at the beginning, the body picks up a high velocity and so its plausible that the travel time could be reduced) and it can be verified experimentally by sliding beads down wires in various shapes. PSfrag replacements (0, 0) A x0 x gravity  ¡£¢ ¢¤ y0 y B (x0 , y0 ) Figure 2.3: The brachistochrone problem. To pose the problem in mathematical terms, we introduce coordinates as shown in Figure 2.3, so that A is the point (0, 0), and B corresponds to (x0 , y0 ). Assuming that the particle is released 1 from rest at A, conservation of energy gives 2 mv 2 − mgy = 0, where we have taken the zero potential energy level at y = 0, and where v denotes the speed of the particle. Thus the speed is √ given by v = ds = 2gy, where s denotes arc length along the curve. From Figure 2.4, we see dt 1 that an element of arc length, δs is given approximately by ((δx)2 + (δy)2 ) 2 .
  • 7. 2.2. Calculus of variations: some classical problems 33 PSfrag replacements δs δy δx Figure 2.4: Element of arc length. Hence the time of descent is given by 2 dx ds 1 y0 1+ dy T = √ =√ dy. curve 2gy 2g 0 y Our problem is to find the path {x(y), y ∈ [0, y0 ]}, satisfying x(0) = 0 and x(y0 ) = x0 , which minimizes T , that is, to determine the minimizer for the function I : S → R, where 1 y0 1 1 + (x (y))2 2 I(x) = √ dy, x ∈ S, 2g 0 y 1+β 2 and S = {x ∈ C 1 [0, y0 ] | x(0) = 0 and x(y0 ) = x0 }. Here2 F (α, β, γ) = γ is independent of α, and so the Euler-Lagrange equation becomes d x (y) 1 √ = 0. dy 1 + (x (y)) 2 y Integrating with respect to y, we obtain x (y) 1 √ = C, 1 + (x (y))2 y where C is a constant. It can be shown that the general solution of this differential equation is given by 1 ˜ 1 x(Θ) = (Θ − sin Θ) + C, y(Θ) = (1 − cos Θ), 2C 2 2C 2 ˜ where C is another constant. The constants are chosen so that the curve passes through the points (0, 0) and (x0 , y0 ). (0, 0) x PSfrag replacements (x0 , y0 ) y Figure 2.5: The cycloid through (0, 0) and (x0 , y0 ). This curve is known as a cycloid, and in fact it is the curve described by a point P in a circle that rolls without slipping on the x axis, in such a way that P passes through (x0 , y0 ); see Figure 2.5. 2 Strictly speaking, the F here does not satisfy the demands made in Theorem 2.1.1. Notwithstanding this fact, with some additional argument, the solution given here can be fully justified.
  • 8. 34 Chapter 2. The Euler-Lagrange equation 2.2.2 Minimum surface area of revolution The problem of minimum surface area of revolution is to find among all the curves joining two given points (x0 , y0 ) and (x1 , y1 ), the one which generates the surface of minimum area when rotated about the x axis. The area of the surface of revolution generated by rotating the curve y about the x axis is x1 S(y) = 2π y(x) 1 + (y (x))2 dx. x0 Since the integrand does not depend explicitly on x, the Euler-Lagrange equation is ∂F F (y(x), y (x), x) − y (x) (y(x), y (x), x) = C, ∂β where C is a constant, that is, (y )2 y 1 + (y )2 − y = C. 1 + (y )2 Thus y = C 1 + (y )2 , and it can be shown that this differential equation has the general solution x + C1 y(x) = C cosh . (2.4) C This curve is called a catenary. The values of the arbitrary constants C and C 1 are determined by the conditions y(x0 ) = y0 and y(x1 ) = y1 . It can be shown that the following three cases are possible, depending on the positions of the points (x0 , y0 ) and (x1 , y1 ): 1. If a single curve of the form (2.4) passes through the points (x0 , y0 ) and (x1 , y1 ), then this curve is the solution of the problem; see Figure 2.6. y1 y0 PSfrag replacements x0 x1 Figure 2.6: The catenary through (x0 , y0 ) and (x1 , y1 ). 2. If two critical curves can be drawn through the points (x0 , y0 ) and (x1 , y1 ), then one of the curves actually corresponds to the surface of revolution if minimum area, and the other does not. 3. If there does not exist a curve of the form (2.4) passing through the points (x0 , y0 ) and (x1 , y1 ), then there is no surface in the class of smooth surfaces of revolution which achieves the minimum area. In fact, if the location of the two points is such that the distance between them is sufficiently large compared to their distances from the x axis, then the area of the surface consisting of two circles of radius y0 and y1 will be less than the area of any surface of revolution generated by a smooth curve passing through the points; see Figure 2.7.
  • 9. 2.3. Free boundary conditions 35 PSfrag replacements y1 y0 x0 x1 Figure 2.7: The polygonal curve (x0 , y0 ) − (x0 , 0) − (x1 , 0) − (x1 , y1 ). This is intuitively expected: imagine a soap bubble between concentric rings which are being pulled apart. Initially we get a soap bubble between these rings, but if the distance separating the rings becomes too large, then the soap bubble breaks, leaving soap films on each of the two rings. This example shows that a critical curve need not always exist in the class of curves under consideration. 2.3 Free boundary conditions Besides the simplest optimisation problem considered in the previous section, we now consider the optimisation problem with free boundary conditions (see Figure 2.8). a b t PSfrag replacements Figure 2.8: Free boundary conditions. Let F (α, β, γ) be a function with continuous first and second partial derivatives with respect to (α, β, γ). Then find x ∈ C 1 [a, b] which is an extremum for the function b dx I(x) = F x(t), (t), t dt. (2.5) a dt Theorem 2.3.1 Let I : C 1 [a, b] → R be a function of the form b dx I(x) = F x(t), (t), t dt, x ∈ C 1 [a, b], a dt where F (α, β, γ) is a function with continuous first and second partial derivatives with respect to (α, β, γ). If I has a local extremum at x0 , then x0 satisfies the Euler-Lagrange equation: ∂F dx0 d ∂F dx0 x0 (t), (t), t − x0 (t), (t), t = 0, t ∈ [a, b], (2.6) ∂α dt dt ∂β dt
  • 10. 36 Chapter 2. The Euler-Lagrange equation together with the transversality conditions ∂F dx0 ∂F dx0 x0 (t), (t), t =0 and x0 (t), (t), t = 0. (2.7) ∂β dt t=a ∂β dt t=b Proof Step 1. We take X = C 1 [a, b] and compute DI(x0 ). Proceeding as in the proof of Theorem 2.1.1, it is easy to see that b ∂F ∂F DI(x0 )(h) = (x0 (t), x0 (t), t) h(t) + (x0 (t), x0 (t), t) h (t) dt, a ∂α ∂β h ∈ C 1 [a, b]. Theorem 1.4.2 implies that this linear functional must be the zero map, that is, (DI(x0 ))(h) = 0 for all h ∈ C 1 [a, b]. In particular, it is also zero for all h in C 1 [a, b] such that h(a) = h(b) = 0. But recall that in Step 3 and Step 4 of the proof of Theorem 2.1.1, we proved that if b ∂F ∂F (x0 (t), x0 (t), t) h(t) + (x0 (t), x0 (t), t) h (t) dt = 0 (2.8) a ∂α ∂β for all h in C 1 [a, b] such that h(a) = h(b) = 0, then this implies that he Euler-Lagrange equation (2.6) holds. Step 2. Integration by parts in (2.8) now gives b ∂F d ∂F DI(x0 )(h) = (x0 (t), x0 (t), t) − (x0 (t), x0 (t), t) h(t)dt + (2.9) a ∂α dt ∂β t=b ∂F (x0 (t), x0 (t), t) h(t) (2.10) ∂β t=a ∂F ∂F = 0+ (x0 (t), x0 (t), t) h(b) − (x0 (t), x0 (t), t) h(a). ∂β t=b ∂β t=a The integral in (2.9) vanishes since we have shown in Step 1 above that (2.6) holds. Thus the condition DI(x0 ) = 0 now takes the form ∂F ∂F (x0 (t), x0 (t), t) h(b) − (x0 (t), x0 (t), t) h(a) = 0, ∂β t=b ∂β t=a from which (2.7) follows, since h is arbitrary. This completes the proof. Exercises. 1. Find all curves y = y(x) which have minimum length between the lines x = 0 and the line x = 1. 2. Find critical curves for the following function, when the values of x are free at the endpoints: 1 1 I(x) = (x (t))2 + x(t)x (t) + x (t) + x(t) dt. 0 2
  • 11. 2.3. Free boundary conditions 37 Similarly, we can also consider the mixed case (see Figure 2.9), when one end of the curve is fixed, say x(a) = ya , and the other end is free. Then it can be shown that the curve x satisfies the Euler-Lagrange equation, the transversality condition ∂F (x0 (t), x0 (t), t) h(a) = 0 ∂β t=a at the free end point, and x(a) = ya serves as the other boundary condition. yb ¢£ ya  ¡ PSfrag replacements t t a b a b Figure 2.9: Mixed cases. We can summarize the results by the following: critical curves for (2.5) satisfy the Euler- Lagrange equation (2.6) and moreover there holds ∂F (x0 (t), x0 (t), t) = 0 at the free end point. ∂β Exercises. 1. Find the curve y = y(x) which has minimum length between (0, 0) and the line x = 1. 2. The cost of a manufacturing process in an industry is described by the function I given by 1 1 I(x) = (x (t))2 + x(t) dt, 0 2 for x ∈ C 1 [0, 1] with x(0) = 1. (a) If also x(1) = 0, find a critical curve x∗ for I, and find I(x∗ ). (b) If x(1) is not specified, find a critical curve x∗∗ for I, and find I(x∗∗ ). (c) Which of the values I(x∗ ) and I(x∗∗ ) found in parts above is larger? Explain why you would expect this, assuming that x∗ and x∗∗ in fact minimize I on the respective domains specified above. 3. Find critical curves for the following functions: π 2 (a) I(x) = (x(t))2 − (x (t))2 dt, x(0) = 0 and x π 2 is free. 0 π 2 (b) I(x) = (x(t))2 − (x (t))2 dt, x(0) = 1 and x π 2 is free. 0 1 4. Determine the curves that maximize the function I : S → R, where I(x) = cos(x (t))dt, 0 1 x ∈ S and S = {x ∈ C [0, 1] | x(0) = 0}. What are the curves that minimize I?
  • 12. 38 Chapter 2. The Euler-Lagrange equation 2.4 Generalization The results in this chapter can be generalized to the case when the integrand F is a function of more than one independent variable: if we wish to find extremum values of the function b dx1 dxn I(x1 , . . . , xn ) = F x1 (t), . . . , xn (t), (t), . . . , (t), t dt, a dt dt where F (α1 , . . . , αn , β1 , . . . , βn , γ) is a function with continuous partial derivatives of order ≤ 2, and x1 , . . . , xn are continuously differentiable functions of the variable t, then following a similar analysis as before, we obtain n Euler-Lagrange equations to be satisfied by the optimal curve, that is, ∂F d ∂F (x1∗ (t), . . . , xn∗ (t), x1∗ (t), . . . , xn∗ (t), t)− (x1∗ (t), . . . , xn∗ (t), x1∗ (t), . . . , xn∗ (t), t) = 0, ∂αk dt ∂βk for t ∈ [a, b], k ∈ {1, . . . , n}. Also at any end point where xk is free, ∂F dx1∗ dxn∗ x1∗ (t), . . . , xn∗ (t), (t), . . . , (t), t = 0. ∂βk dt dt Exercise. Find critical curves of the function π 2 I(x1 , x2 ) = (x1 (t))2 + (x2 (t))2 + 2x1 (t)x2 (t) dt 0 π π such that x1 (0) = 0, x1 2 = 1, x2 (0) = 0, x2 2 = 1. Remark. Note that with the above result, we can also solve the problem of finding extremal curves for a function of the type b dx dn x I(x) : F x(t), (t), . . . , n (t), t dt, a dt dt for over all (sufficiently differentiable) curves x defined on an interval [a, b], taking values in R. Indeed, we may introduce the auxiliary functions dx dn x x1 (t) = x(t), x2 (t) = (t), . . . , xn (t) = n (t), t ∈ [a, b], dt dt ˜ and consider the problem of finding extremal curves for the new function I defined by b ˜ I(x1 , . . . , xn ) = F (x1 (t), x2 (t), . . . , xn (t), t)dt. a Using the result mentioned in this section, we can then solve this problem. Note that we eliminated high order derivatives at the price of converting the scalar function into a vector-valued function. Since we can always do this, this is one of the reasons in fact for considering functions of the type (1.28) where no high order derivatives occur. 2.5 Optimisation in function spaces versus that in Rn To understand the basic ‘infinite-dimensionality’ in the problems of optimisation in function spaces, it is interesting to see how they are related to the problems of the study of functions of n real
  • 13. 2.5. Optimisation in function spaces versus that in Rn 39 variables. Thus, consider a function of the form b dx I(x) = F x(t), (t), t dt, x(a) = ya , x(b) = yb . a dt Here each curve x is assigned a certain number. To find a related function of the sort considered in classical analysis, we may proceed as follows. Using the points a = t0 , t1 , . . . , tn , tn+1 = b, we divide the interval [a, b] into n + 1 equal parts. Then we replace the curve {x(t), t ∈ [a, b]} by the polygonal line joining the points (t0 , ya ), (t1 , x(t1 )), . . . , (tn , x(tn )), (tn+1 , yb ), and we approximate the function I at x by the sum n xk − xk−1 In (x1 , . . . , xn ) = F xk , , t k hk , (2.11) hk k=1 where xk = x(tk ) and hk = tk − tk−1 . Each polygonal line is uniquely determined by the ordinates x1 , . . . , xn of its vertices (recall that x0 = ya and xn+1 = yb are fixed), and the sum (2.11) is therefore a function of the n variables x1 , . . . , xn . Thus as an approximation, we can regard the optimisation problem as the problem of finding the extrema of the function In (x1 , . . . , xn ). In solving optimisation problems in function spaces, Euler made extensive use of this ‘method of finite differences’. By replacing smooth curves by polygonal lines, he reduced the problem of finding extrema of a function to the problem of finding extrema of a function of n variables, and then he obtained exact solutions by passing to the limit as n → ∞. In this sense, functions can be regarded as ‘functions of infinitely many variables’ (that is, the infinitely many values of x(t) at different points), and the calculus of variations can be regarded as the corresponding analog of differential calculus of functions of n real variables.
  • 14. 40 Chapter 2. The Euler-Lagrange equation