# Calculus of Variations/CHAPTER XII

CHAPTER XII: A FOURTH AND FINAL CONDITION FOR THE EXISTENCE OF A MAXIMUM OR A MINIMUM, AND A PROOF THAT THE CONDITIONS WHICH HAVE BEEN GIVEN ARE SUFFICIENT.

• 153 The notion of a field continued from the preceding Chapter.
• 154 The function ${\displaystyle {\mathcal {E}}(x,y,p,q,{\bar {p}},{\bar {q}})}$.
• 155 The function ${\displaystyle {\mathcal {E}}}$ must have the same sign for every point of the curve.
• 156 The sufficiency of the above condition.
• 157 Another form of the function ${\displaystyle {\mathcal {E}}}$.
• 158 Still another form.
• 159,160 The signs of the function ${\displaystyle {\mathcal {E}}}$ and ${\displaystyle F_{1}}$.
• 161 Another proof of the sufficiency of the condition as given in Article 156.
• 162 The function ${\displaystyle {\mathcal {E}}}$ cannot be zero along an entire curve in the given field.
• 163 The envelope of conjugate points.
• 164 The curve may be composed of a finite number of regular traces.
• 165 Cases where the traces are not regular.
• 166 Generalizations in the Integral Calculus.
• 167,168,169,170,171,172 Applications to the four problems already considered.
• 173 When ${\displaystyle F(x,y,x',y')}$ is a rational function of ${\displaystyle x'}$ and ${\displaystyle y'}$, there can exist neither a maximum nor a minimum value of the integral.
• 174 General summary.
• 175 Extensions and generalizations: Instead of the determination of a structure of the first kind in the domain of two variables, it may be required to determine a structure of the first kind in the domain of ${\displaystyle n}$ quantities.
• 176 When equations of condition exist among the variables.
• 177 When the second and higher derivatives appear.
• 178 The Calculus of variations applied to the determination of structures of a higher kind. The minimal surfaces.

Article 153.
In the preceding Chapter we considered the family of curves that have the same initial point ${\displaystyle A}$ and satisfy the differential equation ${\displaystyle G=0}$. These deviate very little from one another in their initial direction. We saw that the curves again intersect only in the neighborhood of points that are the conjugates of ${\displaystyle A}$, the conjugate point along any curve being the limiting position of the point of intersection of this curve and a neighboring curve when the angle between their initial directions becomes infinitesimally small. All points that lie on these curves before the points that are conjugate to ${\displaystyle A}$ form a connected portion of surface ; that is, if ${\displaystyle P_{1}}$ is a point belonging to this collectivity of points, a boundary may be described about ${\displaystyle P_{1}}$ so that all points within this boundary also belong to the collectivity of points.

For, let

${\displaystyle x=\phi (t,\alpha ,\beta )\qquad y=\psi (t,\alpha ,\beta )}$

be the equations of a given curve which satisfies ${\displaystyle G=0}$, and let the coordinates of a point on this curve be

${\displaystyle x_{1}=\phi (t_{1},\alpha ,\beta )\qquad y_{1}=\psi (t_{1},\alpha ,\beta )}$

Further, let ${\displaystyle x_{1}+\xi ,y_{1}+\eta }$ be the coordinates of another point ${\displaystyle P_{2}}$ that lies in the neighborhood of ${\displaystyle P_{1},}$ so that ${\displaystyle \xi ,\eta }$ are quantities arbitrarily small.

We may then (Art. 151) draw a curve between ${\displaystyle A}$ and ${\displaystyle P_{2}}$ which satisfies the differential equation ${\displaystyle G=0}$, if we can determine four quantities ${\displaystyle \tau ,\tau ',\alpha ',\beta '}$ as power-series in ${\displaystyle \xi ,\eta }$ in such a way that the following equations are true:

{\displaystyle {\begin{aligned}0&=\phi '(t_{0})\tau '+\phi _{1}(t_{0})\alpha '+\phi _{2}(t_{0})\beta '+(\tau ',\alpha ',\beta ')_{2}\\0&=\psi '(t_{0})\tau '+\psi _{1}(t_{0})\alpha '+\psi _{2}(t_{0})\beta '+(\tau ',\alpha ',\beta ')_{2}\\\xi &=\phi '(t_{1})\tau '+\phi _{1}(t_{1})\alpha '+\phi _{2}(t_{1})\beta '+(\tau ,\alpha ',\beta ')_{2}\\\eta &=\psi '(t_{1})\tau '+\psi _{1}(t_{1})\alpha '+\psi _{2}(t_{1})\beta '+(\tau ,\alpha ',\beta ')_{2}\end{aligned}}}

Since the determinant of these equations (Art. 151) is ${\displaystyle -\Theta (t_{1},t_{0})}$ and is different from zero, the point ${\displaystyle t_{1}}$ not being conjugate to ${\displaystyle t_{0}}$, it follows that the quantities ${\displaystyle \tau ,\tau ',\alpha ',\beta '}$ may be developed in powerseries in ${\displaystyle \xi ,\eta }$ which are convergent for small values of these quantities.

Consequently, a curve may be drawn through ${\displaystyle A}$ and ${\displaystyle P_{2}}$ which satisfies the differential equation ${\displaystyle G=0}$, and this curve will be neighboring the first curve and will deviate as little as we wish in direction from its initial direction, if ${\displaystyle \xi ,\eta }$, and consequently also ${\displaystyle \tau ,\tau ',\alpha ',\beta '}$ are sufficiently small.

If we form the determinant for the curve ${\displaystyle AP_{2}}$, which, when put equal to zero, is the equation for the determination of the point conjugate to ${\displaystyle A}$, it is seen that this determinant also may be developed as a power-series in ${\displaystyle \xi ,\eta }$, which becomes ${\displaystyle -\Theta (t_{1},t_{0})}$ when ${\displaystyle \xi =\eta =0}$. The function ${\displaystyle \Theta (t_{1},t_{0})}$ is different from zero when sufficiently small values are ascribed to ${\displaystyle \xi ,\eta }$. Consequently, within the interval ${\displaystyle AP_{2}}$, there is present no point which is conjugate to ${\displaystyle A}$.

We may therefore envelop the interval situated between two conjugate points of the original curve by a narrow surface area, which is of such a nattire that a curve, and only one, may be drawn from the point ${\displaystyle A}$ to any point within it, which satisfies the differential equation ${\displaystyle G=0}$, is neighboring the first curve and deviates in its initial direction only a little from it.

Article 154.
Let a portion of curve ${\displaystyle P_{0}P_{1}}$. satisfying the differential equation ${\displaystyle G=0}$, be given, which is of such a nature that for no point on it ${\displaystyle F_{1}=0}$ or ${\displaystyle \infty }$, and suppose that the point conjugate to ${\displaystyle P_{0}}$ does not lie before ${\displaystyle P_{1}}$. Between ${\displaystyle P_{0}}$ and ${\displaystyle P_{1}}$ take an arbitrary point ${\displaystyle P_{2}}$ and draw through ${\displaystyle P_{2}}$ a regular curve.[1] On this curve we choose a point ${\displaystyle P_{3}}$ so close to ${\displaystyle P_{2}}$ that a curve may be drawn through ${\displaystyle P_{0}}$ and ${\displaystyle P_{3}}$ which satisfies the differential equation ${\displaystyle G=0}$, and which lies entirely within the strip of surface defined above. Let us consider the change in the integral when we take it over ${\displaystyle P_{0}P_{3}+P_{3}P_{2}}$ instead of over ${\displaystyle P_{0}P_{2}}$. We may denote an integral taken over a curve that satisfies the differential equation by ${\displaystyle I}$, and one over an arbitrary curve by ${\displaystyle {\bar {I}}}$, and we may denote the direction of integration by added indices. We have therefore to compute the expression

${\displaystyle \Delta I=I_{03}+{\bar {I}}_{32}-I_{02}}$

or

${\displaystyle \Delta I=\left(\int _{\overline {P_{0}P_{3}}}F{\text{d}}t-\int _{\overline {P_{0}P_{2}}}F{\text{d}}t\right)+\int _{\overline {P_{0}P_{2}}}F{\text{d}}t}$

an expression which (Art. 79)

${\displaystyle =\epsilon \left(\int _{\overline {P_{0}P_{2}}}Gw{\text{d}}s+\left[\xi {\frac {\partial F}{\partial x'}}+\eta {\frac {\partial F}{\partial y'}}\right]^{t_{2}}\right)+(\epsilon ^{2})+\int _{\overline {P_{0}P_{2}}}F{\text{d}}t}$

where ${\displaystyle \xi ,\eta }$ are measured in the direction from ${\displaystyle P_{2}}$ to ${\displaystyle P_{3}}$.

At the point ${\displaystyle P_{2}}$ and along the curve ${\displaystyle P_{3}P_{2}}$ in the direction ${\displaystyle P_{3}P_{2}}$ we have

${\displaystyle \epsilon \xi =-{\bar {x}}_{2}'{\text{d}}{\bar {t}}+({\text{d}}{\bar {t}})^{2}\qquad \epsilon \eta =-{\bar {y}}_{2}'{\text{d}}{\bar {t}}+({\text{d}}{\bar {t}})^{2}}$

${\displaystyle {\text{d}}{\bar {t}}}$ denoting that this differential is taken with respect to the curve ${\displaystyle P_{3}P_{2}}$.

If we consider the arguments in ${\displaystyle F}$ expressed as functions of ${\displaystyle {\bar {t}}}$ along the curve ${\displaystyle P_{3}P_{2}}$, it follows that

${\displaystyle \int _{\overline {P_{0}P_{2}}}F{\text{d}}t=F(x_{2},y_{2},{\bar {x}}_{2}',{\bar {y}}_{2}'){\text{d}}{\bar {t}}+({\text{d}}{\bar {t}})^{2}}$

Hence at the point ${\displaystyle P_{2}}$, which is an arbitrary point of the curve ${\displaystyle P_{0}P_{1}}$, we have

${\displaystyle \Delta I=\left(F(x_{2},y_{2},{\bar {x}}_{2}',{\bar {y}}_{2}')-\left[{\bar {x}}_{2}'{\frac {\partial }{\partial x_{2}'}}F(x_{2},y_{2},x_{2}',y_{2}')+{\bar {y}}_{2}'{\frac {\partial }{\partial y_{2}'}}F(x_{2},y_{2},x_{2}',y_{2}')\right]\right){\text{d}}{\bar {t}}+({\text{d}}{\bar {t}})^{2}\qquad }$(a)

The function ${\displaystyle F}$ is homogeneous of the first order (Art. 68) with respect to its third and fourth arguments, so that (see Art. 72)

${\displaystyle F(x_{2},y_{2},{\bar {x}}_{2}',{\bar {y}}_{2}')={\bar {x}}_{2}'F^{(1)}(x_{2},y_{2},{\bar {x}}_{2}',{\bar {y}}_{2}')+{\bar {y}}_{2}'F^{(2)}(x_{2},y_{2},{\bar {x}}_{2}',{\bar {y}}_{2}')}$

We define by ${\displaystyle {\mathcal {E}}(x,y,x',y',{\bar {x}}',{\bar {y}}')}$ the expression

${\displaystyle 1)\qquad {\mathcal {E}}(x,y,x',y',{\bar {x}}',{\bar {y}}')={\bar {x}}'{\big (}F^{(1)}(x_{2},y_{2},{\bar {x}}_{2}',{\bar {y}}_{2}')-F^{(1)}(x_{2},y_{2},x_{2}',y_{2}'){\big )}+{\bar {y}}'{\big (}F^{(2)}(x_{2},y_{2},{\bar {x}}_{2}',{\bar {y}}_{2}')-F^{(2)}(x_{2},y_{2},x_{2}',y_{2}'){\big )}}$

Hence at the point ${\displaystyle P2}$ it follows that

${\displaystyle \Delta I={\mathcal {E}}(x,y,x',y',{\bar {x}}',{\bar {y}}'){\text{d}}{\bar {t}}+({\text{d}}{\bar {t}})^{2}}$

when in the function ${\displaystyle {\mathcal {E}}}$ we have substituted for the arguments those values that belong to the point ${\displaystyle P_{2}}$. The direction-cosines of the curve ${\displaystyle P_{0}P_{2}}$ at ${\displaystyle P}$ are denoted by

${\displaystyle p_{2}={\frac {x_{2}'}{\sqrt {x_{2}'^{2}+x_{2}'^{2}}}}\qquad {\text{and}}\qquad q_{2}={\frac {y_{2}'}{\sqrt {x_{2}'^{2}+x_{2}'^{2}}}}}$

and those of the curve ${\displaystyle P_{3}P_{2}}$ at ${\displaystyle P_{2}}$ by ${\displaystyle {\bar {p}}_{2}}$ and ${\displaystyle {\bar {q}}_{2}}$. It is evident from a consideration of the right-hand side of the formula defining ${\displaystyle {\mathcal {E}}}$ above (and cf. Art. 68) that

${\displaystyle {\frac {{\mathcal {E}}(x,y,x',y',{\bar {x}}',{\bar {y}}')}{\sqrt {{\bar {x}}_{2}'^{2}+{\bar {y}}_{2}'^{2}}}}={\mathcal {E}}(x,y,p,q,{\bar {q}},{\bar {q}})}$

Article 155.
If further we denote by ${\displaystyle \sigma }$ the differential of arc ${\displaystyle P_{3}P_{2}}$, we have finally

${\displaystyle 2)\qquad \Delta I={\mathcal {E}}(x,y,p,q,{\bar {p}}',{\bar {q}})\sigma +(\sigma )^{2}}$

Accordingly, if we take ${\displaystyle \sigma }$ sufficiently small; that is, if we choose the point ${\displaystyle P_{3}}$ very close to ${\displaystyle P_{2}}$, then we may always bring it about that the change in the integral has the same sign as that of the function ${\displaystyle {\mathcal {E}}}$.

The point ${\displaystyle P_{2}}$ was an arbitrary point on the curve ${\displaystyle P_{0}P_{1}}$, and ${\displaystyle P_{2}P_{3}}$ also represented an arbitrary direction.

It follows that if for any point ${\displaystyle P_{2}}$ and for any direction at ${\displaystyle P_{2}}$ the function ${\displaystyle {\mathcal {E}}}$ were negative, and for any other point and direction positive, then the given curve could vary in such a manner that the change in the integral is at one time positive and at another time negative. We have, therefore, the following theorem :

If the integral taken over the curve ${\displaystyle P_{0}P_{1}}$ which satisfies the differential equation ${\displaystyle G=0}$ is to be a maximum or a minimum, then the function ${\displaystyle {\mathcal {E}}}$ must have the same sign for every point of the curve, and at every point of the curve for any direction, and this sign m.ust be negative for a maximum and positive for a minimum,.

Article 156.
That the above condition is sufficient to assure the existence of a maximum or a minimum may be shown as follows : Let ${\displaystyle P_{0}(I)P_{1}}$ be a curve which satisfies the four conditions already established (and recapitulated in Art. 174), and let ${\displaystyle P_{0}(II)P_{1}}$ be any arbitrary curve that lies in the field about the curve ${\displaystyle P_{0}(I)P_{1}}$. It is subject only to the condition that it must be a regular curve and lie wholly in the given field.

By varying the parameters ${\displaystyle \alpha }$ and ${\displaystyle \beta }$ we can construct a system of curves as near as we like to one another, all satisfying the differential equation ${\displaystyle G=0}$. These curves cut the curve ${\displaystyle P_{0}(II)P_{1}}$ in two (or perhaps more) points. They do not cut the curve ${\displaystyle P_{0}(I)P_{1}}$ or intersect among themselves within the field in question. The function ${\displaystyle {\mathcal {E}}}$ must have the same sign along each of these curves as it has along the curve ${\displaystyle P_{0}(I)P_{1}}$. For, take an arbitrary point ${\displaystyle P}$ on any of these curves. Then on the curve ${\displaystyle P_{0}(I)P_{1}}$ there is a point for which the quantities ${\displaystyle x,y,p,q}$ differ only a little from the quantities that belong to the point ${\displaystyle P}$, and consequently ${\displaystyle {\mathcal {E}}}$ has the same sign for both points.

Consider now the variation in our integrals as we pass from ${\displaystyle P_{0}(I)P_{1}}$ to ${\displaystyle P_{0}P_{2}P_{3}P_{1}}$ and from ${\displaystyle P_{0}P_{2}P_{3}P_{1}}$ to ${\displaystyle P_{0}P_{4}P_{5}P_{1}}$, etc. As we saw in the preceding article, the variation caused by passing from ${\displaystyle P_{0}(I)P_{1}}$ to ${\displaystyle P_{0}P_{2}P_{3}P_{1}}$

${\displaystyle =\int _{t_{0}}^{t_{2}}F{\text{d}}t+\epsilon \left[{\frac {\partial F}{\partial x'}}\xi +{\frac {\partial F}{\partial y'}}\eta \right]_{t_{2}}^{t_{3}}+\int _{t_{3}}^{t_{1}}F{\text{d}}t}$
${\displaystyle =\int _{t_{0}}^{t_{2}}F{\text{d}}t-\left[{\bar {p}}{\frac {\partial F}{\partial x'}}+{\bar {q}}{\frac {\partial F}{\partial y'}}\right]^{t_{2}}\sigma +\int _{t_{3}}^{t_{1}}F{\text{d}}t-\left[{\bar {p}}{\frac {\partial F}{\partial x'}}+{\bar {q}}{\frac {\partial F}{\partial y'}}\right]^{t_{3}}\sigma }$

${\displaystyle {\bar {p}},{\bar {q}}}$ being the direction cosines of the tangent to the curve ${\displaystyle P_{0}(II)P_{1}}$ at the points ${\displaystyle P_{2}}$ and ${\displaystyle P_{3}}$, which, we notice, have opposite signs at these points.

If we denote the integration along the curves by the curves themselves, it is seen at once that the variation in these integrals may be expressed by

${\displaystyle P_{0}P_{2}P_{3}1P_{1}-P_{0}(I)P_{1}=[{\mathcal {E}}]^{t_{0}}\sigma +[{\mathcal {E}}]^{t_{1}}\sigma +(\sigma )^{2}}$

where the first is the length from ${\displaystyle P_{0}}$ to ${\displaystyle P_{2}}$ and the second from ${\displaystyle P_{3}}$ to ${\displaystyle P_{1}}$.

Similarly the differences in the integrals along

${\displaystyle P_{0}P_{4}P_{5}1P_{1}-P_{0}P_{2}P_{3}P_{1}=[{\mathcal {E}}]^{t_{2}}\sigma +[{\mathcal {E}}]^{t_{3}}\sigma +(\sigma )^{2}}$
${\displaystyle P_{0}P_{6}P_{7}1P_{1}-P_{0}P_{4}P_{5}P_{1}=[{\mathcal {E}}]^{t_{4}}\sigma +[{\mathcal {E}}]^{t_{5}}\sigma +(\sigma )^{2}}$
${\displaystyle ...................................}$
${\displaystyle P_{0}P_{2\nu }P_{2\nu +1}1P_{1}-P_{0}P_{2\nu -2}P_{2\nu -1}P_{1}=[{\mathcal {E}}]^{t_{2\nu -2}}\sigma +[{\mathcal {E}}]^{t_{2\nu -1}}\sigma +(\sigma )^{2}}$
${\displaystyle P_{0}(II)P_{1}-P_{0}P_{2\nu }P_{2\nu +1}P_{1}=[{\mathcal {E}}]^{t_{2\nu }}\sigma +[{\mathcal {E}}]^{t_{2\nu +1}}\sigma +(\sigma )^{2}}$

Adding these results together, we have the difference in the integrals along

${\displaystyle P_{0}(II)P_{1}-P_{0}(II)P_{1}=\int _{P_{0}(II)P_{1}}{\mathcal {E}}\sigma +(\sigma )^{2}}$

${\displaystyle \sigma }$ being a differential of arc along the curve ${\displaystyle P_{0}(I)P_{1}}$. This is a verification of the theorem stated at the end of the last Article.

We also see that, if we had not assured ourselves that none of the intermediary curves intersect, the signs of the ${\displaystyle \sigma }$'s would not all have been alike, and consequently the sum total of all these ${\displaystyle \sigma }$'s would not have constituted the curve ${\displaystyle P_{)}(II)P_{1}}$.

Article 157.
Another form of the function ${\displaystyle {\mathcal {E}}}$.

We have seen in the Integral Calculus that

${\displaystyle f(p_{1},q_{1})-f(p_{0},q_{0})=\int _{p_{0},q_{0}}^{p_{1},q_{1}}{\text{d}}f(p,q)}$
${\displaystyle \qquad =\int _{p_{0},q_{0}}^{p_{1},q_{1}}\left({\frac {\partial f(p,q)}{\partial p}}{\text{d}}p+{\frac {\partial f(p,q)}{\partial q}}{\text{d}}q\right)}$
${\displaystyle \qquad =\int _{k=0}^{k=1}{\big (}f^{(1)}[p_{0}+k(p_{1}-p_{0}),q_{0}+k(q_{1}-q_{0})](p_{1}-p_{0})+f^{(2)}[p_{0}+k(p_{1}-p_{0}),q_{0}+k(q_{1}-q_{0})](q_{1}-q_{0}){\big )}{\text{d}}k}$

Hence, if we write

${\displaystyle p_{k}=p+k({\bar {p}}-p)=(1-k)p+k{\bar {p}}}$
${\displaystyle q_{k}=q+k({\bar {q}}-q)=(1-k)q+k{\bar {q}}}$

it is seen that

${\displaystyle F^{(1)}(x,y,{\bar {p}},{\bar {q}})-F^{(1)}(x,y,p,q)=\int _{k=0}^{k=1}{\big (}F^{(11)}(x,y,p_{k},q_{k})({\bar {p}}-p)+F^{(12)}(x,y,p_{k},q_{k})({\bar {q}}-q){\big )}{\text{d}}k}$
${\displaystyle F^{(2)}(x,y,{\bar {p}},{\bar {q}})-F^{(2)}(x,y,p,q)=\int _{k=0}^{k=1}{\big (}F^{(21)}(x,y,p_{k},q_{k})({\bar {p}}-p)+F^{(22)}(x,y,p_{k},q_{k})({\bar {q}}-q){\big )}{\text{d}}k}$

Note that (see Art. 73)

${\displaystyle F^{(11)}=q_{k}^{2}F_{1}\qquad F^{(12)}=-p_{k}q_{k}F_{1}\qquad F^{(22)}=p_{k}^{2}F_{1}}$

and further that ${\displaystyle F^{(12)}=F^{(21)}}$.

By substituting these values in the above expressions, and in turn the resulting quantities in the expression for ${\displaystyle {\mathcal {E}}}$, we have

${\displaystyle {\mathcal {E}}(x,y,p,q,{\bar {p}}{\bar {q}})={\bar {p}}[F^{(1)}(x,y,{\bar {p}},{\bar {q}})-F^{(1)}(x,y,p,q)]+{\bar {q}}[F^{(2)}(x,y,{\bar {p}},{\bar {q}})-F^{(2)}(x,y,p,q)]}$
${\displaystyle \qquad =\int _{k=0}^{k=1}F_{1}(x,y,p_{k},q_{k}){\big (}[q_{k}({\bar {p}}-p)-p_{k}({\bar {q}}-q)]q_{k}{\bar {p}}+[-q_{k}({\bar {p}}-p)+p_{k}({\bar {q}}-q)]p_{k}{\bar {q}}{\big )}{\text{d}}k}$

The expression in the square brackets is

${\displaystyle (q_{k}{\bar {p}}-p_{k}{\bar {q}})[q_{k}({\bar {p}}-p)-p_{k}({\bar {q}}-q)]=(1-k)(q{\bar {p}}-p{\bar {q}})^{2}}$

and consequently

${\displaystyle 3)\qquad {\mathcal {E}}(x,y,p,q,{\bar {p}},{\bar {q}})=(q{\bar {p}}-p{\bar {q}})^{2}\int _{k=0}^{k=1}F_{1}(x,y,p_{k},q_{k}){\text{d}}k}$

This expression for ${\displaystyle {\mathcal {E}}}$ in the form of a definite integral is defective, in that it has a meaning only when ${\displaystyle F_{1}}$ remains finite for all values of ${\displaystyle p_{k}}$ and ${\displaystyle q_{k}}$, as ${\displaystyle k}$ varies between 0 and 1. For example, if ${\displaystyle k=1/2}$, then ${\displaystyle p_{1/2}=p+({\bar {p}}-p)/2=(p+{\bar {p}})/2}$, andi if ${\displaystyle {\bar {p}}=-p}$, then ${\displaystyle p_{1/2}=0}$; in the same way for ${\displaystyle k=1/2}$ and ${\displaystyle {\bar {q}}=-q}$, then also ${\displaystyle q_{1/2}=0}$. These two arguments being zero, ${\displaystyle F_{1}}$ becomes infinite (cf. Art. 73). Further, if the two directions ${\displaystyle p,q}$ and ${\displaystyle {\bar {p}},{\bar {q}}}$ coincide, then ${\displaystyle {\mathcal {E}}}$ becomes zero of the second order.

If ${\displaystyle OP}$ and ${\displaystyle O{\bar {P}}}$ are vectors of unit length with components ${\displaystyle p,q}$ and ${\displaystyle {\bar {p}},{\bar {q}}}$, then the components of ${\displaystyle OP'}$, when ${\displaystyle P'}$ travels along the line ${\displaystyle P{\bar {P}}}$, are ${\displaystyle p_{k},q_{k},k}$ varying between 0 and 1.

Article 158.
Another form was given by Weierstrass to the expression ${\displaystyle {\mathcal {E}}}$, in which he avoided the defect mentioned above, by integrating along the arc of a circle instead of along the straight line ${\displaystyle P{\bar {P}}}$. If we integrate along the arc of a circle of unit radius from the point ${\displaystyle P}$ to the point ${\displaystyle {\bar {P}}}$ we obtain an expression for ${\displaystyle {\mathcal {E}}}$ which is universally true.

We have as before, if ${\displaystyle POX=\tau }$, ${\displaystyle {\bar {P}}OX={\bar {\tau }}}$, ${\displaystyle \omega ={\bar {\tau }}-\tau ({\text{mod }}2\pi )}$, and ${\displaystyle -\pi \leq \omega \leq =\pi }$,

${\displaystyle {\mathcal {E}}(x,y,p,q,{\bar {p}},{\bar {q}})={\bar {p}}[F^{(1)}(x,y,{\bar {p}},{\bar {q}})-F^{(1)}(x,y,p,q)]+{\bar {q}}[F^{(2)}(x,y,{\bar {p}},{\bar {q}})-F^{(2)}(x,y,p,q)]}$
${\displaystyle =\cos {\bar {\tau }}[F^{(1)}(x,y,\cos {\bar {\tau }},\sin {\bar {\tau }})-F^{(1)}(x,y,\cos \tau ,\sin \tau )]+\sin {\bar {\tau }}[F^{(2)}(x,y,\cos {\bar {\tau }},\sin {\bar {\tau }})-F^{(2)}(x,y,\cos \tau ,\sin \tau )]}$
${\displaystyle =\cos {\bar {\tau }}\int _{\lambda =0}^{\lambda =\omega }{\text{d}}_{\lambda }F^{(1)}[x,y,\cos(\tau +\lambda ),\sin(\tau +\lambda )]+\sin {\bar {\tau }}\int _{\lambda =0}^{\lambda =\omega }{\text{d}}_{\lambda }F^{(2)}[x,y,\cos(\tau +\lambda ),\sin(\tau +\lambda )]}$

But, if ${\displaystyle F^{(1)}}$ denotes the derivative of ${\displaystyle F}$ with respect to its third argument, etc.,

${\displaystyle {\text{d}}_{\lambda }F^{(1)}[x,y,\cos(\tau +\lambda ),\sin(\tau +\lambda )]}$
${\displaystyle \qquad =[F_{\cos ,\cos }^{(11)}\sin(\tau +\lambda )+F_{\cos ,\sin }^{(12)}\cos(\tau +\lambda )]{\text{d}}\lambda }$
${\displaystyle \qquad =[-\sin ^{3}(\tau +\lambda )-\sin(\tau +\lambda )\cos ^{2}(\tau +\lambda )]F_{1}{\text{d}}\lambda }$
${\displaystyle \qquad =-\sin(\tau +\lambda )F_{1}[x,y,\cos(\tau +\lambda ),\sin(\tau +\lambda )]{\text{d}}\lambda }$

similarly,

${\displaystyle {\text{d}}_{\lambda }F^{(2)}[x,y,\cos(\tau +\lambda ),\sin(\tau +\lambda )]=\cos(\tau +\lambda )F_{1}[x,y,\cos(\tau +\lambda ),\sin(\tau +\lambda )]{\text{d}}\lambda }$

Hence, it follows that

${\displaystyle {\mathcal {E}}(x,y,p,q,{\bar {p}},{\bar {q}})=\int _{\lambda =0}^{\lambda =\omega }[-\cos {\bar {\tau }}\sin(\tau +\lambda )+\sin {\bar {\tau }}\cos(\tau +\lambda )]F_{1}{\text{d}}\lambda }$
${\displaystyle \qquad =\int _{\lambda =0}^{\lambda =\omega }\sin({\bar {\tau }}-\tau -\lambda )F_{1}{\text{d}}\lambda =\int _{\lambda =0}^{\lambda =\omega }\sin(\omega -\lambda )F_{1}[x,t,\cos(\tau +\lambda ),\sin(\tau +\lambda )]{\text{d}}\lambda }$

If we write

${\displaystyle \omega -\lambda =\lambda '}$

the integral just written is

${\displaystyle \int _{\lambda '=0}^{\lambda '=\omega }\sin \lambda 'F_{1}[x,y,\cos({\bar {\tau }}-\lambda '),\sin({\bar {\tau }}-\lambda ')]{\text{d}}\lambda '=F_{1}[x,y,\cos({\bar {\tau }}-\lambda _{2}'),\sin({\bar {\tau }}-\lambda _{2}')]\int _{\lambda '=0}^{\lambda '=\omega }{\text{d}}~\cos \lambda '}$

where ${\displaystyle \lambda _{2}'}$ is intermediary between 0 and ${\displaystyle \omega }$.

We therefore have finally

${\displaystyle 4)\qquad {\mathcal {E}}(x,y,p,q,{\bar {p}},{\bar {q}})=(1-\cos \omega )F_{1}[x,y,cos({\bar {\tau }}-\lambda _{2}'),\sin({\bar {\tau }}\lambda _{2}')]}$

If then ${\displaystyle F_{1}[x,y,cos({\bar {\tau }}-\lambda '),\sin({\bar {\tau }}\lambda _{'})]}$ has a constant sign between 0 and ${\displaystyle \omega }$ it follows also that ${\displaystyle {\mathcal {E}}(x,y,p,q,{\bar {p}},{\bar {q}})}$ has this sign, since ${\displaystyle \lambda _{2}'}$ is one of the values of ${\displaystyle \lambda '}$ within this interval.

The above formula is true for all values of ${\displaystyle \omega }$ situated between ${\displaystyle -\pi }$ and ${\displaystyle +\pi }$, and since ${\displaystyle \cos({\bar {\tau }}-\lambda _{2}')}$ and ${\displaystyle \sin({\bar {\tau }}-\lambda _{2}')}$ cannot both be zero at the same time, it is seen that

${\displaystyle F_{1}[x,y,\cos({\bar {\tau }}-\lambda _{2}'),\sin({\bar {\tau }}-\lambda _{2}')]\neq \infty }$

and consequently the expression 4 for ${\displaystyle {\mathcal {E}}}$ has not the same defect as the one given in the preceding article.

Article 159.
For any displacement of the curve ${\displaystyle \omega \neq 0}$, and consequently ${\displaystyle 1-\cos \omega }$ is a positive quantity. Hence ${\displaystyle {\mathcal {E}}}$ has the same sign as ${\displaystyle F_{1}}$. If ${\displaystyle F_{1}[x,y,\cos({\bar {\tau }}-\lambda ),\sin({\bar {\tau }}-\lambda )]}$ is found by examination to have always the same sign independently of ${\displaystyle \cos({\bar {\tau }}-\lambda )}$, ${\displaystyle \sin({\bar {\tau }}-\lambda )}$ for every point of the curve within the interval in question, then we may be convinced that there is a maxim.um, or a minimumof the integral without the derivation and examination of the function ${\displaystyle {\mathcal {E}}}$. By this process, however, we have shown without the second variation that the function ${\displaystyle F_{1}(x,y,p,q)}$ can change its sign for no point on the curve, and for no direction of the tangent to the curve at a point.

Article 160.
It is evident that if ${\displaystyle F_{1}}$, considered as a function of its third and fourth arguments, has a definite sign, then ${\displaystyle {\mathcal {E}}}$ has also the same sign; but if ${\displaystyle {\mathcal {E}}}$ retains a definite sign, ${\displaystyle p}$ and ${\displaystyle q}$ being fixed while ${\displaystyle {\bar {p}}}$ and ${\displaystyle {\bar {q}}}$ are varied, it does not then follow that ${\displaystyle F_{1}}$ always has a definite sign. This is illustrated in the following example, due to Schwarz :

Let

${\displaystyle F(x,y,x',y')=\alpha {\sqrt {x'^{2}+y'^{2}}}+beta{\frac {x'y'^{2}}{x'^{2}+y'^{2}}}=(\alpha +\beta \cos \tau \sin ^{2}\tau ){\sqrt {x'^{2}+y'^{2}}}}$

It follows that

${\displaystyle F_{1}(x,y,x',y')={\frac {\alpha }{(x'^{2}+y'^{2})^{3/2}}}+2\beta {\frac {x'(x'^{2}-3y'^{2})}{(x'^{2}+y'^{2})^{3}}}=(\alpha +2\beta \cos 3\tau ){\frac {1}{(x'^{2}+y'^{2})^{3/2}}}}$

and, since ${\displaystyle x'^{2}+y'^{2}=\cos ^{2}\lambda +\sin ^{2}\lambda =1}$,

${\displaystyle {\mathcal {E}}(x,y,p,q,{\bar {p}},{\bar {q}})=\int _{\lambda =0}^{\lambda =\omega }\sin(\omega -\lambda )F_{1}[x,y,\cos(\tau +\lambda ),\sin(\tau +\lambda )]{\text{d}}\lambda }$
${\displaystyle \qquad =\int _{\lambda =0}^{\lambda =\omega }\sin(\omega -\lambda )(\alpha +2\beta \cos 3\lambda ){\text{d}}\lambda }$

where we have written ${\displaystyle \tau +\lambda =\lambda }$ or ${\displaystyle \tau =0}$; i.e., we have taken the ${\displaystyle X}$-axis as the initial direction, from which ${\displaystyle \omega }$ is measured.

Noting that

${\displaystyle \sin(\omega +2\lambda )+\sin(\omega -4\lambda )=2\sin(\omega -\lambda )\cos 3\lambda }$

it is seen that

${\displaystyle {\mathcal {E}}(x,y,p,q,{\bar {p}},{\bar {q}})=(1-\cos \omega )[\alpha +\beta (\cos \omega +\cos ^{2}\omega )]}$

The greatest and least values that ${\displaystyle \cos \omega +\cos ^{2}\omega }$ can have are 2 and ${\displaystyle -1/4}$, the corresponding values of ${\displaystyle \omega }$ being 0 and ${\displaystyle 2\pi /3}$. Hence, if we we make ${\displaystyle \alpha =l}$ and ${\displaystyle /beta=1}$, the function ${\displaystyle {\mathcal {E}}}$ is situated between the values

${\displaystyle {\frac {3}{4}}(1-\cos \omega )~{\text{and}}~3(1-\cos \omega )}$

and can consequently vanish only for ${\displaystyle \omega =0}$, and is never negativeOn the other hand, ${\displaystyle 1+2\cos 3\tau }$ changes sign repeatedly, for example, when ${\displaystyle \tau =40^{\circ }}$.

Article 161.
The proof stated at the end of Art. 155 is of paramount importance in the determination whether there exists a true maximum or minimum. The proof of the sufficiency of this theorem, as illustrated in Art. 156, was given in a somewhat different form by Prof. Schwarz. Owing to its importance we add another proof, taken from the lectures of Weierstrass.

Let ${\displaystyle OO_{1}1}$ be the curve which satisfies the differential equation ${\displaystyle G=0}$, and let ${\displaystyle 0_{1}31}$ be the arbitrary curve in the field, as defined in Art. 156. Let 3 be any point on the arbitrary curve, whose coordinates we consider as functions of length of arc ${\displaystyle s}$ (instead of ${\displaystyle t}$, as before). The point ${\displaystyle O_{1}}$ is taken between 0 and 1 so that the curve ${\displaystyle 0_{1}31}$ may lie wholly within the field, since the field might terminate in a point at 0. From the point 0 we draw a curve to 3 which satisfies the differential equation ${\displaystyle G=0}$. We consider the sum of integrals ${\displaystyle I_{03}+{\bar {I}}_{31}}$ as a function of ${\displaystyle s}$. This function we denote by ${\displaystyle \int (s)}$. Further, take on the arbitrary curve a point 2 in the neighborhood of the point 3 and before it. Join the points 0 and 2 by a curve which satisfies the differential equation ${\displaystyle G=0}$. Then, if we denote the increment of ${\displaystyle s}$ by ${\displaystyle \sigma }$, it is seen that

${\displaystyle 5)\qquad \int (s-\sigma )-\int (s)=I_{02}+{\bar {I}}_{21}-I_{03}-{\bar {I}}_{31}=I_{02}-I_{03}+{\bar {I}}_{23}={\mathcal {E}}(x_{3},y_{3},p_{3},q_{3},{\bar {p}}_{3},{\bar {q}}_{3})\sigma +(\sigma )^{2}}$

In the same manner take a point 4 immediately after the point 3 on the arbitrary curve and join this point with the point 0 by a curve which satisfies the differential equation ${\displaystyle G=0}$. Then we have

${\displaystyle 6)\qquad \int (s-\sigma )-\int (s)=I_{04}-I_{03}-{\bar {I}}_{34}=-{\mathcal {E}}(x_{3},y_{3},p_{3},q_{3},{\bar {p}}_{3},{\bar {q}}_{3})\sigma +(\sigma )^{2}}$

It therefore follows that

${\displaystyle 7)\qquad \lim _{\sigma =0}{\frac {\int (s-\sigma )-\int (s)}{-\sigma }}=\lim _{\sigma =0}{\frac {\int (s+\sigma )-\int (s)}{\sigma }}=-{\mathcal {E}}(x_{3},y_{3},p_{3},q_{3},{\bar {p}}_{3},{\bar {q}}_{3})}$

that is, the quantity ${\displaystyle -{\mathcal {E}}(x_{3},y_{3},p_{3},q_{3},{\bar {p}}_{3},{\bar {q}}_{3})}$ is the differential quotient of the function ${\displaystyle \int (s)}$ at the point 3.

If, then, along the curve ${\displaystyle O_{1}31}$ the function ${\displaystyle {\mathcal {E}}}$ is nowhere positive, the function ${\displaystyle \int (s)}$ continuously diminishes when the point 3 slides from ${\displaystyle O_{1}}$ toward the point 1.

Let the point [/itex]O_{1}[/itex], which was taken very near the point 0, coincide with this point; then we can say :

If the function ${\displaystyle {\mathcal {E}}}$ is nowhere positive and is not zero at every point of the arbitrary curve 031 the integral taken over the original curve is always greater than the integral extended over the curve 031 ; and if the function ${\displaystyle {\mathcal {E}}}$ is not negative and not zero at evevy point of the curve 031 then the integral taken over the original curve 01 is continuously less than the integral extended over the arbitrary curve 031.

Article 162.
It remains yet to see if it is possible for the function ${\displaystyle {\mathcal {E}}}$ to vanish along the whole curve 031. It appears from the formula 3) that this is possible only when along the whole curve we have

${\displaystyle (p{\bar {q}}-q{\bar {p}})^{2}=0{\text{ or }}p{\bar {q}}-q{\bar {p}}=0}$

In this case every curve 03 which satisfies the differential equation ${\displaystyle G=0}$ has a common tangent at the point 3 with the curve 031.

We shall show that the curve ${\displaystyle MN}$ which is formed of the points conjugate to the point 0 has this property, and that no curve having this property can be drawn from 0 within the region that is bounded by ${\displaystyle MN}$. In other words, ${\displaystyle {\mathcal {E}}}$ is equal to zero along the curve ${\displaystyle MN}$, but is not equal to zero for all the points of any other curve that can be drawn within the region that is enveloped by ${\displaystyle MN}$.

All the curves that satisfy the differential equation ${\displaystyle G=0}$, which pass through one point, and whose initial directions differ from one another by very small quantities, may be represented (Art. 148) in the form

${\displaystyle x=\phi (t,k)\qquad y=\psi (t,k)}$

where the values of ${\displaystyle k}$ are within certain limits.

To each curve corresponds a different value of ${\displaystyle k}$. If, therefore, we fix a value of ${\displaystyle k}$ and take a second value ${\displaystyle k+k'}$ the curve which corresponds to this value may be expressed by the equations

${\displaystyle x+\xi =\phi (t+\tau ',k+k')\qquad y+\eta =\psi (t+\tau ',k+k')}$

where the same value of ${\displaystyle t}$ corresponds to the initial directions of both curves.

If the latter curve is cut by the former we must have

${\displaystyle 0=\phi '(t)\tau '+{\frac {\partial \phi }{\partial k}}k'+(\tau ',k')_{2}\qquad 0=\psi '(t)\tau '+{\frac {\partial \psi }{\partial k}}k'+(\tau ',k')_{2}}$

The determinant of the linear terms of the equations just written gives, when put equal to zero, the equation for the determination of the point conjugate to the initial point, i. e.,

${\displaystyle \phi '(t){\frac {\partial \psi }{\partial k}}-\psi '(t){\frac {\partial \phi }{\partial k}}=0}$ \qquad (A)

The smallest root of this equation, which is greater than the value ${\displaystyle t_{0}}$ of ${\displaystyle t}$, gives the value of ${\displaystyle t}$, which belongs to the conjugate point. If this value is ${\displaystyle t_{1}}$, then the coordinates of the point are

${\displaystyle {\bar {x}}=\phi (t_{1},k)\qquad {\bar {y}}=\psi (t_{1},k)}$

If we consider ${\displaystyle t_{1}}$ as a function of ${\displaystyle k}$, defined through the equation (A), and if we give to ${\displaystyle k}$ a series of values, the two equations just written represent the curve that is constituted of the points conjugate to 0.

The direction-cosines of the tangent to this curve are proportional to the quantities ${\displaystyle {\frac {\partial {\bar {x}}}{\partial k}},{\frac {\partial {\bar {y}}}{\partial k}}}$. But we also have

${\displaystyle {\frac {\partial {\bar {x}}}{\partial k}}={\frac {\partial \phi (t_{1},k)}{\partial t_{1}}}{\frac {{\text{d}}t_{1}}{{\text{d}}k}}+{\frac {\partial \phi (t_{1},k)}{\partial k}}\qquad {\frac {\partial {\bar {y}}}{\partial k}}={\frac {\partial \psi (t_{1},k)}{\partial t_{1}}}{\frac {{\text{d}}t_{1}}{{\text{d}}k}}+{\frac {\partial \psi (t_{1},k)}{\partial k}}}$

Multiply the first of these equations by ${\displaystyle {\frac {\partial \psi (t_{1},k)}{\partial t_{1}}}=\psi '(t_{1})}$, and subtract from it the second after it has been multiplied by ${\displaystyle {\frac {\partial \phi (t_{1},k)}{\partial t_{1}}}=\phi '(t_{1})}$. We have then, with the aid of (A),

${\displaystyle \phi '(t_{1}){\frac {{\text{d}}{\bar {y}}}{{\text{d}}k}}-\psi '(t_{1}){\frac {{\text{d}}{\bar {x}}}{{\text{d}}k}}=0}$

Since ${\displaystyle \phi '(t_{1}),\psi '(t_{1})}$ are proportional to the direction-cosines of the tangent at a point ${\displaystyle t_{1}}$ of the curve through ${\displaystyle t_{0}}$ and ${\displaystyle t_{1}}$, which satisfies the differential equation ${\displaystyle G=0}$, it follows from the above equation that the tangents to both curves at the point ${\displaystyle t_{1}}$ coincide. Hence, the locus of the conjugate points to is the envelope of the curves through 0, which satisfy the differential equation ${\displaystyle G=0}$.

Article 163.
Let ${\displaystyle {\bar {x}}=f(u)}$ and ${\displaystyle {\bar {y}}=g(u)}$ be an arbitrary curve 031, which passes through the point 0, and is situated entirely within the region bounded by the envelope. Further, suppose that 031 does not coincide throughout its whole extent with any of the curves passing through 0, which satisfy the differential equation ${\displaystyle G=0}$. Suppose, however, that 031 is touched by the curves that pass through and satisfy the differential equation ${\displaystyle G=0}$. At the point of contact we must have

${\displaystyle \phi (t,k)=f(u)\qquad \psi (t,k)=g(u)}$

and

${\displaystyle {\frac {\partial \phi }{\partial t}}{\frac {{\text{d}}g}{{\text{d}}u}}-{\frac {\partial \psi }{\partial t}}{\frac {{\text{d}}f}{{\text{d}}u}}=0}$ \qquad (B)

The values of ${\displaystyle t}$ and ${\displaystyle u}$, which belong to the point of contact, are determined as functions of ${\displaystyle k}$ through the first two equations.

These equations, being true for sufficiently small values of ${\displaystyle k}$, may be differentiated with respect to ${\displaystyle k}$, and we thus have:

${\displaystyle {\frac {\partial \phi }{\partial t}}{\frac {{\text{d}}t}{{\text{d}}k}}+{\frac {\partial \phi }{\partial k}}={\frac {{\text{d}}f}{{\text{d}}u}}{\frac {{\text{d}}u}{{\text{d}}k}}\qquad {\frac {\partial \psi }{\partial t}}{\frac {{\text{d}}t}{{\text{d}}k}}+{\frac {\partial \psi }{\partial k}}={\frac {{\text{d}}g}{{\text{d}}u}}{\frac {{\text{d}}u}{{\text{d}}k}}}$

If we multiply the first of these equations by ${\displaystyle {\frac {{\text{d}}g}{{\text{d}}u}}}$ and the second by ${\displaystyle -{\frac {{\text{d}}f}{{\text{d}}u}}}$ and add we have with the aid of (B)

${\displaystyle {\frac {\partial \phi }{\partial k}}{\frac {{\text{d}}g}{{\text{d}}u}}-{\frac {\partial \psi }{\partial k}}{\frac {{\text{d}}f}{{\text{d}}u}}=0}$

If between this equation and the equation (B) we eliminate the quantities ${\displaystyle {\frac {{\text{d}}g}{{\text{d}}u}}}$ and ${\displaystyle {\frac {{\text{d}}f}{{\text{d}}u}}}$, we have

${\displaystyle {\frac {\partial \phi }{\partial t}}{\frac {\partial \psi }{\partial k}}-{\frac {\partial \psi }{\partial t}}{\frac {\partial \phi }{\partial k}}=0}$

an equation, which served for the determination of the point conjugate to the initial point. Consequently the point of contact of the curve, that passes through 0 and satisfies the differential equation ${\displaystyle G=0}$, with the arbitrary curve must be the point conjugate to 0.

But this is possible only if the curve ${\displaystyle {\bar {x}}=f(u),{\bar {y}}=g(u)}$ coincides with the envelope ; while according to our supposition the curve 031 is to lie entirely within the region that is bounded by the envelope. It follows that there can be within the region no curve 031 such that each of the curves which satisfies the differential equation ${\displaystyle G=0}$, and which joins the point 0 with a point of 031, touches 031 at the same time.

Hence, the quantity ${\displaystyle q{\bar {p}}-p{\bar {q}}}$ can be everywhere zero only when the arbitrary curve between 0 and 1 coincides throughout its whole extent with one of the curves that passes through 0 and satisfies the differential equation ${\displaystyle G=0}$. But since, within the strip of surface inclosing the field as we have defined it, there can be only one curve drawn through 0 and 1 which satisfies the differential equation ${\displaystyle G=0}$, it follows that the arbitrary curve 031 can coincide only with the original curve 01, and then it is not a variation of that curve. It therefore follows that the function ${\displaystyle {\mathcal {E}}}$ cannot vanish for all the points of the curve that has been subjected to variation.

Article 164.
It is not necessary that the curve 031 be a single trace of a regular curve in its whole extent. If we assume that 031 is composed of an arbitrary number of regular portions of curve, the integral may be regarded as the sum of the integrals over the single portions, and the conclusions made above are also applicable.

It may happen that one of the portions of curve coincides throughout its whole extent with a portion of one of the curves that goes through 0 and satisfies the differential equation ${\displaystyle G=O}$. If this is the case for 23, for example, so that ${\displaystyle {\mathcal {E}}}$ is equal to zero along 23, then we may replace this portion of curve by an arbitrary portion of curve ${\displaystyle 2'3}$, which lies very near 23. Then the theorem proved above is true for the curve ${\displaystyle 02'31}$, viz., that

${\displaystyle I_{03}{\overset {>}{<}}I_{02'}+{\bar {I}}_{2'3}}$

according as the function ${\displaystyle {\mathcal {E}}}$ is nowhere positive or nowhere negative along the curve ${\displaystyle 02'31}$. Now, if we bring the curve ${\displaystyle 2'3}$ as near to the curve 23 as we wish, the absolute value of the difference ${\displaystyle I_{03}-I_{02'}-{\bar {I}}_{2'3}}$ can be made smaller than any arbitrarily small quantity ${\displaystyle \delta }$; and, in accordance with what was proved above, in the first case the difference ${\displaystyle I_{03}-I_{02'}-{\bar {I}}_{2'3}}$ is certainly not negative, and in the second case it is not positive.

If we shove the point 3 further along the arbitrary curve toward 1, then, when 3 takes a position in the neighborhood of 4, it follows again that ${\displaystyle I_{04}-I_{03}-{\bar {I}}_{34}}$ is greater or less than zero, and, as above, we see that the integral ${\displaystyle I_{01}}$, extended over the curve that satisfies the differential equation ${\displaystyle G=0}$, is greater or less than the integral taken over the arbitrary curve 0231, according as the function ${\displaystyle {\mathcal {E}}}$ is nowhere negative or nowhere positive.

Article 165.
Further, it is not necessary that the single portions of the curve which has been subjected to variation be regular in order that our conclusions be correctly drawn, if only the coordinates can be expressed as functions of some quantity, and if these functions have derivatives. Finally, if we consider the variation made quite arbitrary, so that only the positions of the points are given, while it is not known whether their coordinates have derivatives, then indeed the integral taken over this curve has no longer any meaning. But the meaning of the integral may be extended so that it has a signification even in this case. For if at first we assume that the coordinates of the curve, which has been subjected to variation, are expressible through functions that have derivatives, then the integral taken over the curve is

${\displaystyle \int _{t_{0}}^{t_{1}}F[f(t),g(t),f'(t),g'(t)]{\text{d}}t}$

This integral distributed into a sum of integrals (corresponding to the intervals ${\displaystyle t_{0}\ldots \tau _{1},\tau _{1}\ldots \tau _{2},\ldots ,\tau _{n}\ldots t_{1}}$ is equal to

${\displaystyle \int _{t_{0}}^{\tau _{1}}F{\text{d}}t+\int _{\tau _{1}}^{\tau _{2}}F{\text{d}}t+\cdots +\int _{\tau _{n}}^{t_{1}}F{\text{d}}t}$ \qquad (C)

We assume that the points ${\displaystyle x_{0},y_{0};x_{1},y_{1};\ldots x_{n},y_{n};x_{n+1},y_{n+1}}$ correspond to the values ${\displaystyle t_{0}m\tau _{1},\ldots \tau _{n},t_{1}}$.

We then have:

${\displaystyle x_{1}-x_{0}=f'(t_{0})(\tau _{1}-t_{0})+(\tau _{1}-t_{0})[\tau _{1}-t_{0}]}$
${\displaystyle ..........................................}$
${\displaystyle x_{n+1}-x_{n}=f'(\tau _{n})(t_{1}-\tau _{n})+(t_{1}-\tau _{n})[t_{1}-\tau _{n}]}$
${\displaystyle y_{1}-y_{0}=g'(t_{0})(\tau _{1}-t_{0})+(\tau _{1}-t_{0})[\tau _{1}-t_{0}]}$
${\displaystyle ..........................................}$
${\displaystyle y_{n+1}-y_{n}=g'(\tau _{n})(t_{1}-\tau _{n})+(t_{1}-\tau _{n})[t_{1}-\tau _{n}]}$

where ${\displaystyle [\tau _{v}-\tau _{v-1}]}$ denotes a quantity which becomes indefinitely small at the same time with ${\displaystyle \tau _{v}-\tau _{v-1}}$.

For the first of the integrals in the expression (C) we write:

${\displaystyle x=x_{0}+x_{0}'(t-t_{0})+(t-t_{0})[t-t_{0}]}$
${\displaystyle y=y_{0}+y_{0}'(t-t_{0})+(t-t_{0})[t-t_{0}]}$
${\displaystyle x'=x_{0}'+x_{0}''(t-t_{0})+(t-t_{0})[t-t_{0}]}$
${\displaystyle y'=y_{0}'+y_{0}''(t-t_{0})+(t-t_{0})[t-t_{0}]}$

for the second integral we write

${\displaystyle x=x_{1}+x_{1}'(t-\tau _{1})+(t-\tau _{1})[t-\tau _{1}]}$
${\displaystyle y=y_{1}+y_{1}'(t-\tau _{1})+(t-\tau _{1})[t-\tau _{1}]}$

and similarly for the other integrals.

These expressions we write in the sum of integrals (C), and, developing them in power-series, we have through integration

${\displaystyle (\tau _{1}-t_{0})F(x_{0},y_{0},x_{0}',y_{0}')+(\tau _{2}-\tau _{1})F(x_{1},y_{1},x_{1}',y_{1}')+\cdots +(t_{1}-\tau _{n})F(x_{n},y_{n},x_{n}',y_{n}')\ldots }$

plus a similar number of terms, which become indefinitely small of the second order with respect to the quantities ${\displaystyle \tau _{v}-\tau _{v-1}}$.

We may therefore write the integral in the form

${\displaystyle \lim _{n\rightarrow \infty }\left(\sum _{v=1}^{n+1}F(x_{v-1},y_{v-1},x_{v-1}',y_{v-1}')\right)}$

where we must understand by ${\displaystyle \tau _{0}}$ the value ${\displaystyle t_{0}}$, and by ${\displaystyle t_{n+1}}$ the value ${\displaystyle t_{1}}$.

Since ${\displaystyle \tau _{v}-\tau _{v-1}}$ are positive quantities, and the functions ${\displaystyle F}$ in regard to ${\displaystyle x_{1}',y_{1}'\ldots }$ are homogeneous of the first degree, we may write the above limit in the form

${\displaystyle \lim _{n\rightarrow \infty }\left(\sum _{v=1}^{n+1}F(x_{v-1},y_{v-1},(\tau _{v}-\tau _{v-1})x_{v-1}',(\tau _{v}-\tau _{v-1})y_{v-1}')\right)}$

or, since

${\displaystyle x_{v}-x_{v-1}=(\tau _{v}-\tau _{v-1})x_{v-1}'+(\tau _{v}-\tau _{v-1})[\tau _{v}-\tau _{v-1}]}$

the above expression is

${\displaystyle \lim _{n\rightarrow \infty }{\big (}F(x_{0},y_{0},x_{1}-x_{0},y_{1}-y_{0})+\cdots +F(x_{n},y_{n},x_{n+1}-x_{n},y_{n+1}-y_{n}){\big )}}$

Article 166.
The integral in the above form has a more general meaning than the one hitherto employed, with which, however, it coincides in every particular where that one has a meaning. We may assume, with respect to any arbitrary variation, a series of points ${\displaystyle x_{0},y_{0};x_{1},y_{1};\ldots x_{n},y_{n};x_{n+1},y_{n+1}}$ of such a nature that the distance between, say, two successive points does not exceed a certain quantity ${\displaystyle \delta }$.

We then form the sum

${\displaystyle F(x_{0},y_{0},x_{1}-x_{0},y_{1}-y_{0})+\cdots +F(x_{n},y_{n},x_{n+1}-x_{n},y_{n+1}-y_{n})}$

If we make ${\displaystyle \delta }$ smaller and smaller by increasing the number of points, it may happen that this sum approaches a definite limit. We call this limit the value of the integral taken over the curve. It may also happen that the limit does not approach a definite value; for example, it may vacillate between two values. We then say the integral taken over this curve has no meaning.

If we think of the series of points that are taken upon the curve, joined together successively by a broken line, the integral taken over this broken line will approach the same limit as will the integral taken over the curve, if the integral has a meaning.

If, therefore, a curve 01 is given, which satisfies all the conditions that have hitherto been made for a maximum or a minimum, and if this curve varies in an arbitrary manner, then if the integral taken over the curve, which has been subjected to variation, has a meaning as defined above, we ma)' draw a broken line, the integral over which deviates as little as we wish from the integral taken over the curve that has been caused to vary and to which the theorem of Art. 161 is applicable. Consequently, we may say, in the case of a maximum, the integral taken over the curve subjected to variation cannot be greater than the integral taken over the original curve, and in the case of a m,inimum,, it cannot be less than the integral taken over the original curve.

Since we may make the region as narrow as we wish within which all the variations are to lie, we ma)' assume that upon the curve which has been varied a point 3 lies so near to 01 (but not upon it) that two curves 03, 31 can be drawn between the points and 3 and between 3 and 1, which also satisfy all the conditions of the problem.

For the sake of brevity, let us assume that we have to do with a maximum. Then, as we have just seen, the integrals over 03 and 31 cannot at all events be smaller than the integrals over the corresponding parts of the curve which has been varied ; but, after the preceding theorems, the integral taken over 01 is greater than the sum of the integrals taken over 03 and 31, and consequently also greater than the integral over the curve that has been varied. A maximum is therefore in reality present.

Article 167.
We may now investigate the behavior of the function ${\displaystyle {\mathcal {E}}}$ in the case of the four problems which we last considered in Arts. 140–144.

The problem of the surface of rotation of minimum area.

We saw that the catenary between limits, within which were situated no pair of conjugate points, was the curve that described a surface of minimum area when rotated around the axis of the half-plane. From the point ${\displaystyle P_{0}}$ we may draw in any direction a curve which satisfies the differential equation ${\displaystyle G=0}$ (a catenary); the function ${\displaystyle F_{1}}$ is positive for each of these curves as soon as we limit ourselves to the half-plane in which ${\displaystyle y}$ is positive. A true minimum will therefore in reality enter. For if ${\displaystyle p,q}$ are the direction-cosines of the tangent to the catenary at any point, ${\displaystyle {\bar {p}},{\bar {q}}}$ those of the tangent to any arbitrary curve through the same point, then, owing to the relations

${\displaystyle F^{(1)}(x,y,x',y')={\frac {yx'}{\sqrt {x'^{2}+y'^{2}}}}\qquad F^{(2)}(x,y,x',y')={\frac {yy'}{\sqrt {x'^{2}+y'^{2}}}}}$

it follows that

${\displaystyle F^{(1)}(x,y,x',y')=yp\qquad F^{(2)}(x,y,x',y')=yq}$

since

${\displaystyle p^{2}+q^{2}=1}$

and consequently

${\displaystyle {\mathcal {E}}(x,y,p,q,{\bar {p}},{\bar {q}})=y{\big (}({\bar {p}}-p){\bar {p}}+{\bar {q}}-q){\bar {q}}{\big )}=y{\big (}1-(p{\bar {p}}+q{\bar {q}}){\big )}}$

The expression ${\displaystyle p{\bar {p}}+q{\bar {q}}}$ is the cosine of the angle between the two tangents. Hence we see that the function ${\displaystyle {\mathcal {E}}}$ is negative for no point which comes under consideration, and for no two directions ${\displaystyle p,q}$ and ${\displaystyle {\bar {p}},{\bar {q}}}$.

If, therefore, ${\displaystyle y=0}$ for no point of the curve, our former conclusions are applicable, and a true minimum of the integral has, in reality, been found.

Article 168.
The Brachistochrone. We saw that this curve is the cycloid

${\displaystyle x=g+r(1-\sin t)\qquad y+a=r(1-\cos t)}$

We assume that the point ${\displaystyle A}$, from which the moving point starts, having an initial velocity proportional to the quantity ${\displaystyle {\sqrt {a}}}$, is the origin of coordinates, and that the ${\displaystyle Y}$-axis is the direction of gravity. We saw that the cycloid could then be generated by a point described by a circle which rolls upon the straight line ${\displaystyle y=-a}$. If ${\displaystyle a}$ is different from zero, an arc of a cycloid may be constructed through ${\displaystyle A}$ in any direction. If the curve passes through a singular point it does not minimize the integral, as was shown in Art. 104. If ${\displaystyle A}$ and ${\displaystyle B}$ are not singular points, the function ${\displaystyle F_{1}}$ has a positive value different from zero everywhere along this curve and in the neighborhood of it in every direction.

Between two arbitrary points (see Art. 105), when the quantity ${\displaystyle a}$ is given, there can alwaj'-s be drawn one, and only one, arc of a cycloid which has no singular points between these two points. If, therefore, ${\displaystyle a}$ is different from zero, and consequently ${\displaystyle A}$ and ${\displaystyle B}$ are not singular points, then (see Art. 159) it follows that the curve, in reality, causes the integral to have a minimum value. Suppose that ${\displaystyle A}$ or ${\displaystyle B}$ is a singular point; then at this point ${\displaystyle F_{1}}$ becomes infinite, a case which we consider in the next Article.

Article 169.
Suppose ${\displaystyle A}$ is a singular point and ${\displaystyle a=0}$. Draw an arbitrary curve between ${\displaystyle A}$ and ${\displaystyle B}$. Take upon this curve in the neighborhood of ${\displaystyle A}$ a point ${\displaystyle A_{1}}$, and through ${\displaystyle A_{1}}$ and ${\displaystyle B}$ draw a cycloid which cuts the ${\displaystyle X}$-axis at ${\displaystyle A_{1}'}$. The material point under the action of gravity passes through ${\displaystyle A_{1}}$ with the same velocity which it would have at an equal distance below the ${\displaystyle X}$-axis if it traversed the cycloid drawn through ${\displaystyle A}$ and ${\displaystyle B}$.

The following notation may be introduced :

${\displaystyle I_{01}}$ to denote the time of falling between ${\displaystyle A_{1}'}$ and ${\displaystyle B}$ upon the cycloid ${\displaystyle A_{1}'B}$,

${\displaystyle I_{01}'}$ to denote the time of falling between ${\displaystyle A}$ and ${\displaystyle A_{1}}$ upon the arbitrary curve ${\displaystyle AB}$,

${\displaystyle I}$ to denote the time of falling between ${\displaystyle A_{1}}$ and ${\displaystyle B}$ upon the cycloid ${\displaystyle A_{1}B}$,

${\displaystyle I'}$ to denote the time of falling between ${\displaystyle A_{1}}$ and ${\displaystyle B}$ upon the arbitrary curve ${\displaystyle A_{1}B}$.

We proved that

${\displaystyle I'>I}$

and therefore, if we write

${\displaystyle {\bar {I}}=I_{01}'+I'}$

it follows that

${\displaystyle {\bar {I}}>I+I_{01}'}$

Now, let the point ${\displaystyle A_{1}}$ approach nearer and nearer the point ${\displaystyle A}$, so that the integral ${\displaystyle I}$ approaches the limit ${\displaystyle I_{01}}$, while ${\displaystyle I_{01}'}$ becomes indefinitely small. We must then have

${\displaystyle {\bar {I}}\geq I_{01}}$

That ${\displaystyle {\bar {I}}}$ is greater than ${\displaystyle I_{01}}$ may be seen as follows: As soon as ${\displaystyle G\neq 0}$ along a portion of curve, we may always vary it in such a way that the increment in the corresponding integral may have any sign. If, then, ${\displaystyle G\neq 0}$ along the whole curve ${\displaystyle AA_{1}B}$, we may substitute another curve, for which, if ${\displaystyle I''}$ is the value of the integral which belongs to it,

${\displaystyle I''<{\bar {I}}}$

But since we also have

${\displaystyle I''\geq I_{01}}$

it follows that

${\displaystyle {\bar {I}}>I_{01}}$

If, on the other hand, ${\displaystyle G=0}$ along the whole curve ${\displaystyle AA_{1}B}$, then this curve must consist of several cycloidal arcs ; since, if it were only one, the curves ${\displaystyle AA_{1}B}$ and ${\displaystyle AB}$ would be identical. These arcs must have different tangents at the point where they come together ; for, since this point cannot lie on the ${\displaystyle X}$-axis, a consecutive point having the same direction must lie on the same cycloidal arc. If corners were present, however, they could be so rounded off that there would be a shorter path between the two points, and consequently, the velocity being the same, the time of falling would be shorter.

Hence the arc of a cycloid also minimizes the time of falling between ${\displaystyle A}$ and ${\displaystyle B}$ in the case where ${\displaystyle A}$ is a singular point ; that is, when the material point starts from ${\displaystyle A}$ with an initial velocity that is zero.

The conclusions just made are also applicable, if ${\displaystyle B}$ is a singular point ; for it makes no difference whether the material point ascends from ${\displaystyle B}$ to ${\displaystyle A}$ or falls from ${\displaystyle A}$ to ${\displaystyle B}$, if we allow the material point to go back with the same initial velocity with which it arrived at ${\displaystyle B}$. On the way back it will reach ${\displaystyle A}$ with its original velocity. Its velocity will be the same in both cases at all points of the curve, but directed toward opposite directions. The integral taken over the curve has the same value in both cases ; and consequently the curve which caused the integral to have a minimum value will also, in the second case, minimize the integral.

Article 170.
The problem of the geodesic line on a sphere offers here nothing of special interest. It is found that the function ${\displaystyle {\mathcal {E}}}$ retains a positive sign along the arc of a great circle situated between two poles.

Article 171.
Problem of the surface of revolution which offers the least resistance.

In this problem

${\displaystyle F()x,y,x',y')={\frac {xx'^{3}}{x'^{2}+y'^{2}}}}$

and since

${\displaystyle p^{2}+q^{2}=1}$

it follows that

${\displaystyle F(x,y,p,q)={\frac {xp^{3}}{p^{2}+q^{2}}}=xp^{3}}$
${\displaystyle {\frac {\partial F}{\partial p}}=x(p^{4}+3p^{2}q^{2})\qquad {\frac {\partial F}{\partial q}}=-2xp^{3}q}$

Substituting these values in

${\displaystyle {\mathcal {E}}(x,y,p,q,{\bar {p}},{\bar {q}})=F(x,y,{\bar {p}},{\bar {q}})-{\bar {p}}{\frac {\partial F}{\partial p}}-{\bar {q}}{\frac {\partial F}{\partial q}}}$

we have

${\displaystyle {\mathcal {E}}(x,y,p,q,{\bar {p}},{\bar {q}})=x[{\bar {p}}^{3}-{\bar {p}}p^{2}-2{\bar {p}}p^{2}q^{2}+2{\bar {q}}p^{3}q]}$
${\displaystyle \qquad =x[{\bar {p}}({\bar {p}}^{2}-p^{2})+2p^{2}q(p{\bar {q}}-{\bar {p}}q)]}$
${\displaystyle \qquad =x[{\bar {p}}({\bar {p}}^{2}(p^{2}+q^{2})+2p^{2}q(p{\bar {q}}-{\bar {p}}q))+2p^{2}q(p{\bar {q}}-{\bar {p}}q)]}$
${\displaystyle \qquad =x({\bar {p}}q-p{\bar {q}})[{\bar {p}}({\bar {p}}q+p{\bar {q}})-2p^{2}q]}$
${\displaystyle \qquad =x({\bar {p}}q-p{\bar {q}})[{\bar {q}}({\bar {q}}q-p{\bar {q}})(p^{2}+q^{2})+2p{\bar {p}}{\bar {q}}(p^{2}+q^{2})-2p^{2}q({\bar {p}}^{2}+{\bar {q}}^{2})]}$
${\displaystyle \qquad =x({\bar {p}}q-p{\bar {q}})[{\bar {p}}({\bar {p}}q-p{\bar {q}})(p^{2}+q^{2})-2p^{2}{\bar {p}}({\bar {p}}q-p{\bar {q}})+2pq{\bar {q}}({\bar {p}}q-p{\bar {q}})]}$
${\displaystyle \qquad =x({\bar {p}}q-p{\bar {q}})[{\bar {p}}(p^{2}+q^{2})-2p^{2}{\bar {p}}+2pq{\bar {q}}]}$
${\displaystyle \qquad =x({\bar {p}}q-p{\bar {q}})[{\bar {p}}(q^{2}-p^{2})+2pq{\bar {q}}]}$

Writing

${\displaystyle \cos {\bar {\tau }}={\bar {p}}\qquad \sin {\bar {\tau }}={\bar {q}}\qquad \cos \tau =p\qquad \sin \tau =q}$

we have

${\displaystyle {\mathcal {E}}(x,y,p,q,{\bar {p}},{\bar {q}})=-x\sin({\bar {\tau }}-\tau )^{2}\cos({\bar {\tau }}+2\tau )}$

Therefore, the sign of ${\displaystyle {\mathcal {E}}}$ is the same as that of ${\displaystyle -\cos({\bar {\tau }}+2\tau )}$, and may be either positive or negative by properly choosing ${\displaystyle {\bar {\tau }}}$, an angle which depends upon ${\displaystyle {\bar {p}},{\bar {q}}}$.

At every point of the curve for which ${\displaystyle x\neq 0}$ the function ${\displaystyle {\mathcal {E}}}$ can have different signs, and consequently a maximum or a minimum value of the integral does not exist. We saw in Art. 109 that ${\displaystyle x}$ must be different from zero for all points of the arc.

Article 172.
Legendre (Mimoire sur la manihre de distinguer les maxima des minima dans le Calcul des Variations) showed that by taking a zigzag line for the generating curve, the resistance could be made as small as we wish.

Suppose that the arc ${\displaystyle P_{1}P_{2}}$, had the desired property of generating a surface of least resistance, and suppose that the tangent to this curve is nowhere parallel to the ${\displaystyle X}$-axis. Writing ${\displaystyle p={\frac {{\text{d}}x}{{\text{d}}y}}}$, it follows that ${\displaystyle p\neq 0}$ along the arc ${\displaystyle P_{1}P_{2}}$.

We have then (Art. 108)

${\displaystyle I_{1,2}=\int _{t_{1}}^{t_{2}}{\frac {xx'^{3}}{x'^{2}+y'^{2}}}{\text{d}}t=\int _{x_{1}}^{x_{2}}{\frac {p^{2}x}{1+p^{2}}}{\text{d}}x}$

Since ${\displaystyle p}$ is finite and continuous along the arc in question, it follows that ${\displaystyle {\frac {p^{2}}{1+p^{2}}}}$ has the same properties along the arc, and therefore

${\displaystyle I_{1,2}={\frac {p_{0}^{2}x}{1+p_{0}^{2}}}\int _{x_{1}}^{x_{2}}x~{\text{d}}x={\frac {p_{0}^{2}x}{1+p_{0}^{2}}}{\frac {x_{2}^{2}-x_{1}^{2}}{2}}}$

where ${\displaystyle p_{0}}$ is a mean value of ${\displaystyle p}$, lying between the points ${\displaystyle P_{1}}$ and ${\displaystyle P_{2}}$ of the curve.

Between the ordinates at ${\displaystyle P_{1}}$ and ${\displaystyle P_{2}}$ draw a line parallel to the ${\displaystyle Y}$-axis, and on this line take a point ${\displaystyle P_{3}}$ whose ordinate is longer than those of the points ${\displaystyle P_{1}}$ and ${\displaystyle P_{2}}$. Draw the straight lines ${\displaystyle P_{1}P_{3}}$ and ${\displaystyle P_{2}P_{3}}$, and let ${\displaystyle p_{1}}$ and ${\displaystyle p_{2}}$ be the values of ${\displaystyle {\frac {{\text{d}}x}{{\text{d}}y}}}$ for these lines. The integral ${\displaystyle \int F{\text{d}}t}$ taken over the broken line ${\displaystyle P_{1}P_{3}P_{2}}$ may be denoted by ${\displaystyle I_{13}+I_{32}}$, where

${\displaystyle I_{13}=\int _{x_{1}}^{x_{3}}{\frac {p_{1}^{2}x}{1+p_{1}^{2}}}{\text{d}}x={\frac {p_{1}^{2}}{1+p_{1}^{2}}}{\frac {x_{3}^{2}-x_{1}^{2}}{2}}}$

and

${\displaystyle I_{32}=\int _{x_{3}}^{x_{2}}{\frac {p_{2}^{2}x}{1+p_{2}^{2}}}{\text{d}}x={\frac {p_{2}^{2}}{1+p_{2}^{2}}}{\frac {x_{2}^{2}-x_{3}^{2}}{2}}}$

We have then

${\displaystyle I_{132}-I_{12}=I_{13}+I_{32}-I_{12}={\frac {p_{1}^{2}(x_{3}^{2}-x_{1}^{2})}{2(1+p_{1}^{2})}}+{\frac {p_{2}^{2}(x_{2}^{2}-x_{3}^{2})}{2(1+p_{2}^{2})}}-{\frac {p_{0}^{2}(x_{2}^{2}-x_{1}^{2})}{2(1+p_{0}^{2})}}}$

The first two terms of this expression may be made as small as we choose by sufficiently diminishing the quantities ${\displaystyle p_{1}}$ and ${\displaystyle p_{2}}$, which is done by removing indefinitely the point ${\displaystyle P_{3}}$ along its ordinate. Hence, their sum is less than the third term, so that, consequently,

${\displaystyle I_{132}

This result may also be derived as follows :

${\displaystyle {\frac {I_{13}+I_{32}}{I_{12}}}={\frac {p_{1}^{2}(1+p_{0}^{2})}{p_{0}^{2}(1+p_{1}^{2})}}{\frac {x_{3}^{2}-x_{1}^{2}}{x_{2}^{2}-x_{1}^{2}}}+{\frac {p_{2}^{2}(1+p_{0}^{2})}{p_{0}^{2}(1+p_{2}^{2})}}{\frac {x_{2}^{2}-x_{3}^{2}}{x_{2}^{2}-x_{1}^{2}}}<{\frac {p_{1}^{2}(1+p_{0}^{2})}{p_{0}^{2}(1+p_{1}^{2})}}+{\frac {p_{2}^{2}(1+p_{0}^{2})}{p_{0}^{2}(1+p_{2}^{2})}}}$

since ${\displaystyle x_{1}.

Hence also for a greater reason

${\displaystyle {\frac {I_{13}+I_{32}}{I_{12}}}=<(p_{1}^{2}+p_{2}^{2}){\frac {1+p_{0}^{2}}{p_{0}^{2}}}}$

From this it is seen that the ratio ${\displaystyle {\frac {I_{13}+I_{32}}{I_{12}}}}$ may be indefinitely dimmished by properly choosing ${\displaystyle p_{1}}$ and ${\displaystyle p_{2}}$. There is then no limit to the least possible resistance.

The method just given does not replace the ${\displaystyle {\mathcal {E}}}$-criterion which shows that no surface of minimal resistance exists. It shows simply that no rotational surface exists, which gives an absolute minimum of resistance—a resistance less than any other neighboring surface. The ${\displaystyle {\mathcal {E}}}$-criterion shows that no minimum exists in the sense of giving a resistance less than that given any neighboring curve within a limited neighborhood.

Article 173.
In the general case, when ${\displaystyle F(x,y,x',y')}$ is a rational function of ${\displaystyle x'}$ and ${\displaystyle y'}$, neither a maximum nor a minimum can exist. For in this case

${\displaystyle {\mathcal {E}}=F(x,y,{\bar {p}},{\bar {q}})-{\bar {p}}{\frac {\partial F}{\partial p}}-{\bar {q}}{\frac {\partial F}{\partial q}}}$

is also a rational function of ${\displaystyle {\bar {p}}}$ and ${\displaystyle {\bar {q}}}$ and homogeneous in these quantities of the first degree. Consequently

${\displaystyle {\mathcal {E}}(x,y,k{\bar {p}},k{\bar {q}})=k{\mathcal {E}}(x,y,{\bar {x}},{\bar {y}})}$

and therefore

${\displaystyle {\mathcal {E}}(x,y,-{\bar {p}},-{\bar {q}})=-{\mathcal {E}}(x,y,{\bar {p}},{\bar {q}})}$

It is thus seen that we have only to reverse the direction of the displacement to effect a change of sign in the function ${\displaystyle {\mathcal {E}}}$.

Article 174.
We have now completely solved the four problems that were proposed in Chapter I, and at the same time one of the principal parts of the Calculus of Variations has been finished. After stating succinctly the four criteria that have been established, we shall take up the second part, which has as its object the theoretical and practical solution of problems, a general type of which were the Problems V and VI of Chapter I.

These criteria may be summarized as follows (cf. Art. 125): There exists a minimum or a maximum value of the integral

${\displaystyle I=\int _{t_{0}}^{t_{1}}F(x,y,x',y'){\text{d}}t}$

where ${\displaystyle F}$ is a one-valued, regular function of its four arguments and homogeneous of the first degree in ${\displaystyle x'}$ and ${\displaystyle y'}$, if

1) the differential equation ${\displaystyle G=0}$ is satisfied for every point of the curve;

2) ${\displaystyle F_{1}}$ is positive or negative throughout the whole interval ${\displaystyle t_{0}\ldots t_{1}}$;

3) there are no conjugate points of the curve within the interval ${\displaystyle t_{0}\ldots t_{1}}$ (limits included);

4) the function ${\displaystyle {\mathcal {E}}}$ is positive or negative throughout the whole interval ${\displaystyle t_{0}\ldots t_{1}}$

In this discussion we have excluded the cases where

1) the extremities of the curve are conjugate points;

2) ${\displaystyle F_{1}=0}$ for some point of the curve;

3) ${\displaystyle F_{1}=0}$ for some stretch of the curve;

4) ${\displaystyle {\mathcal {E}}=0}$ for some point or stretch of the curve.

A general treatment of the first three cases would require the extension of the theory to variations of a higher order. Otherwise particular devices must be employed in every example in which one of the above exceptional cases is found.

Article 175.
Before we begin the consideration of Relative Maxima and Minima, we may, at least, indicate the natural extensions and generalizations of the theory which has already been presented : Instead of the determination of a structure of the first kind[2] in the domain of two quantities, it may be required to determine a structure of the first hind in the domain of ${\displaystyle n}$ quantities.

If a structure of the first kind is determined in the domain of the ${\displaystyle n}$ quantities ${\displaystyle x_{1},x_{2},\ldots ,x_{2}}$, then ${\displaystyle n-1}$ of these quantities may be expressed as functions of the remaining one, say, ${\displaystyle x_{1}}$.

Writing

${\displaystyle u=\int F\left(x_{1},x_{2},\ldots ,x_{n},{\frac {{\text{d}}x_{2}}{{\text{d}}x_{1}}},{\frac {{\text{d}}x_{3}}{{\text{d}}x_{1}}},\ldots ,{\frac {{\text{d}}x_{n}}{{\text{d}}x_{1}}}\right){\text{d}}x_{1}}$

it is seen that ${\displaystyle u}$ is so connected with the ${\displaystyle n-1}$ functions that

${\displaystyle {\frac {{\text{d}}u}{{\text{d}}x_{1}}}=F\left(x_{1},x_{2},\ldots ,x_{n},{\frac {{\text{d}}x_{2}}{{\text{d}}x_{1}}},{\frac {{\text{d}}x_{3}}{{\text{d}}x_{1}}},\ldots ,{\frac {{\text{d}}x_{n}}{{\text{d}}x_{1}}}\right)}$

The difference of the values of ${\displaystyle u}$ at the initial-point and at the end-point of the structure is expressed by a definite integral.

This integral takes the form, when we consider the ${\displaystyle x}$'s expressed as functions of ${\displaystyle t}$, say, ${\displaystyle x_{1}=x_{1}(t),x_{2}=x_{2}(t),\ldots ,x_{n}=x_{n}(t)}$,

${\displaystyle I=\int _{t_{0}}^{t_{1}}F(x_{1},x_{2},\ldots ,x_{n},x_{1}',x_{2}',\ldots ,x_{n}'){\text{d}}t}$

The function ${\displaystyle F}$ must be a one-valued, regular function of its arguments in the whole or a limited portion of the fixed domain.

The value of the integral ${\displaystyle I}$ is independent of the manner in which the variables ${\displaystyle x_{1},x_{2},\ldots ,x_{n}}$ have been expressed as functions of ${\displaystyle t}$. It therefore follows after the analogon of Art. 68 that the function ${\displaystyle F}$ is subjected to the further restriction :

${\displaystyle kF(x_{1},x_{2},\ldots ,x_{n},x_{1}',x_{2}',\ldots ,x_{n}')=F(x_{1},x_{2},\ldots ,x_{n},kx_{1}',kx_{2}',\ldots ,kx_{n}')}$

where ${\displaystyle k}$ is a positive constant.

The indicated generalization of the problem given in Art. 13 may accordingly be expressed as follows :

The ${\displaystyle n}$ quantities ${\displaystyle x_{1},x_{2},\ldots ,x_{n}}$ are to be determined as functions of a quantity ${\displaystyle t}$ in such a manner that for the analytical structure that is defined through the equations

${\displaystyle x_{1}-x_{1}(t),x_{2}=x_{2}(t),\ldots ,x_{n}=x_{n}(t)}$

the value of the integral

${\displaystyle I=\int _{t_{0}}^{t_{1}}F(x_{1},x_{2},\ldots ,x_{n},x_{1}',x_{2}',\ldots ,x_{n}'){\text{d}}t}$

is a maximum or a m.inim,um,; in other words, if one causes the above analytical structure to vary indefinitely little, the change in the integral thereby produced must in the case of a maximum be constantly negative, and in the case of a minimum it must be constantly positive. Further, the function ${\displaystyle F}$ is to be considered a one-valued, regular function of its arguments, and indeed, with respect to ${\displaystyle x_{1}',x_{2}',\ldots ,x_{n}'}$, a homogeneous function of the first degree.

Article 176.
The treatment of the above problem is found to be the complete analogon of the problem given in Art. 13. A greater complication arises when there are present equations of condition among the variables ${\displaystyle x_{1},x_{2},\ldots ,x_{n}}$. An example of this kind we had in Problem III of Chapter I.

This problem may be expressed thus : Among all the curves in space which belong to the surface

${\displaystyle f(x,y,z)=0}$

determine that one for which the integral

${\displaystyle \int _{t_{0}}^{t_{1}}{\sqrt {x'^{2}+y'^{2}+z'^{2}}}{\text{d}}t}$

is a minimum.

The general problem may be formulated as follows : Among the structures of the first kind in the domain of the quantities ${\displaystyle x_{1},x_{2},\ldots ,x_{n}}$, for which the ${\displaystyle m}$ equations

${\displaystyle f_{\mu }(x_{1},x_{2},\ldots ,x_{n})=0\qquad \mu =1,2,\ldots ,m;~~m

exist, that one is to be determined for which the integral

${\displaystyle I=\int _{t_{0}}^{t_{1}}F(x_{1},x_{2},\ldots ,x_{n},x_{1}',x_{2}',\ldots ,x_{n}'){\text{d}}t}$

is a maximum or a minimum.

This problem may be reduced to the one of the preceding Article in an analogous manner as is done in Art. 10 for the case of the shortest line upon a surface. The ${\displaystyle m}$ equations of condition may be satisfied by introducing for the variables ${\displaystyle x_{1},x_{2},\ldots ,x_{n}}$ functions of ${\displaystyle n-m}$ new variables after the method given in the Lectures on the Theory of Maxima and Minima, etc., Chapter I, Art. 15. The new variables are independent of one another, so that the above integral may be replaced by one in which the variables are free from extraneous conditions ; or we may proceed as was done in the Theory of Maxima and Minima where the variables are subject to subsidiary conditions (loc. cit., p. 54).

Article 177.
The more general problem of the Calculus of Variations, in so far as it has to do with the structures of the first kind, may be stated as follows:

Among the structures of the first kind in the domain of the ${\displaystyle n}$ quantities ${\displaystyle x_{1},x_{2},\ldots ,x_{n}}$, for which definite equations of condition exist, not only among the ${\displaystyle n}$ quantities themselves, but also among their first derivatives, that structure is to be determined for which the integral

${\displaystyle I=\int _{t_{0}}^{t_{1}}F(x_{1},x_{2},\ldots ,x_{n},x_{1}',x_{2}',\ldots ,x_{n}'){\text{d}}t}$

becomes a maximum or a minimum

It may be easily shown that the apparently more general case in which ${\displaystyle F}$ is a function of ${\displaystyle x_{1},x_{2},\ldots ,x_{n}}$ and of the first and higher derivatives of these quantities, is contained in the problem just stated. For the sake of simplicity, take the case where only two variables are involved and write

${\displaystyle u=\int _{x_{1}}^{x_{2}}F\left(x,y,{\frac {{\text{d}}y}{{\text{d}}x}},{\frac {{\text{d}}^{2}y}{{\text{d}}x^{2}}},\ldots \right){\text{d}}x}$

If in this integral we express ${\displaystyle x}$ and ${\displaystyle y}$ as functions of ${\displaystyle t}$, we have

${\displaystyle {\frac {{\text{d}}y}{{\text{d}}t}}={\frac {{\text{d}}y}{{\text{d}}x}}x'\qquad {\frac {{\text{d}}^{2}y}{{\text{d}}t^{2}}}={\frac {{\text{d}}^{2}y}{{\text{d}}x^{2}}}x'^{2}+{\frac {{\text{d}}y}{{\text{d}}x}}x''}$

We may consequently change the integral ${\displaystyle u}$ into

${\displaystyle u=\int _{t_{0}}^{t_{1}}F(x,y,x',y',x'',y'',\ldots ){\text{d}}t}$

We further have

${\displaystyle {\frac {{\text{d}}x'}{{\text{d}}t}}-x''=0\qquad {\frac {{\text{d}}y'}{{\text{d}}t}}-y''=0}$

We may therefore write

${\displaystyle u=\int _{t_{0}}^{t_{1}}F\left(x,y,x_{1},y_{1},{\frac {{\text{d}}x_{1}}{{\text{d}}t}},{\frac {{\text{d}}y_{1}}{{\text{d}}t}}\right){\text{d}}t}$

with the equations of condition :

${\displaystyle {\frac {{\text{d}}x}{{\text{d}}t}}=x_{1}\qquad {\frac {{\text{d}}y}{{\text{d}}t}}=y_{1}}$

If, then, there appear in ${\displaystyle F}$ only the first and second derivatives, it is seen that ${\displaystyle F}$ depends upon the four functions ${\displaystyle x,y,x_{1},y_{1}}$ which are to be determined, while at the same time the two equations of condition just written must be satisfied. One of the classes of problems belonging to the general problem just stated is the one which was formulated in Art. 17 and which is treated in the following Chapters.

Article 178.
It may be mentioned finally that the problem of the Calculus of Variations may be further generalized, if we require the determination of structures of a higher kind. For example, in the simplest case the three quantities ${\displaystyle x,y,z}$ may be determined as functions of two independent variables ${\displaystyle u}$ and ${\displaystyle v}$. We have then instead of the single integral the double integral

${\displaystyle \int \int F\left(x,y,z,{\frac {\partial x}{\partial u}},{\frac {\partial y}{\partial u}},{\frac {\partial z}{\partial u}},{\frac {\partial x}{\partial v}},{\frac {\partial y}{\partial v}},{\frac {\partial z}{\partial v}}\right){\text{d}}u{\text{d}}v}$

which must be a maximum or a minimum.

The treatment of this problem would give a theory of Minimal Surfaces.

## References

1. The shaded curves do not satisfy the differential equation ${\displaystyle G=0}$.
2. See my Lectures on the Theory of Maxima and Minima of Functions of Several Variables, pp. IS and 86.