Ordinary Differential Equations/Exact 1

From Wikibooks, open books for an open world
< Ordinary Differential Equations
Jump to: navigation, search

First Order Differential Equations

This page details a method for trying to find solutions to equations of the form

This is often written as a differential form

.

Subsequently, we will refer to this expression as ODE. Differential forms frequently come up in multivariable calculus while studying line integrals.

Exact Differential Equations[edit]

Before we begin identifying and solving exact differential equation it helps to make a few observations. We will begin reminding ourselves of the chain rule from multivariable calculus. Which states how to compute the derivative of a composition of two or more functions. Suppose that is a function of two real variables, and we are given functions and which are functions of a single real variable. Then the function is simply a function of , with and being plugged into as and . The chain rule from multivariable calculus tells us how to calculate the derivative of . It states that:

If we slightly abuse the notation and call the two functions and (instead of and ) then we can write the chain rule as

As an example we could let and we could let and . Then according to the chain rule

Of course this would have been seen more directly by substituting for and to discover that , but this simply gives us a verification that we took the derivative correctly.

We will use this theory to evaluate:

If we examine this expression carefully it looks equal to the left hand side of our ODE above. Specifically if and then our ODE is:

This type of equation is especially easy to solve. The only functions whose derivatives are 0 are constant functions or simple constants. Thus the solution to our ODE, i.e. integrating it, will be given by

Now consider the following example, applying what we have just figured out.

In this example , , and . Notice that if then and . By the way, if you ever wanted to check for yourself what we are doing here and your calculus is a bit rusty, get maxima (http://maxima.sourceforge.net or prepackaged for your favorite Linux distribution or for Android). The derivation of we have just made can easily be replayed in maxima, like so,

(%i1) psi:x*y-(x^3/3);

and

(%i2) diff(psi,x);

yielding

(%o1) y - x^2.

Or going from to ,

(%i1) psi_prime:y-x^2;

and

(%i2) integrate(psi_prime,x);

yielding

(%o1) x*y-(x^3/3)

Turning back to our problem, by our observations above the solution of this equation should be given by , or in other words:

This particular equation is linear so we may easily verify that the solution obtained in this way is correct. When there is a function so that and then the equation is called exact. Unfortunately not every differential equation of the form is exact. In order for this to be an effective method for solving differential equation we need a way to distinguish if a differential equation is exact, and what the function is if the function is exact.

In order to see that and could not be arbitrary, remember form multivariable calculus that (read: the order of the partial derivatives of is exchangeable) whenever the derivatives exist and are continuous. Since , then , similarly . Hence, if the equation is exact we would definitely need

or the same stated differently

This can be put into a theorem.

Theorem

Suppose that and have continuous partial derivatives and they satisfy the relationship Then there is a function so that and

Proof. We prove this by giving an explicit construction of . First notice that if exists then by integrating the expression in we get:

Here just taking an anti-derivative with respect to x, treating y as a constant. It is necessary to add a function , because for any function . So that with the above definition .

Now we need to determine . To do this we use that .

Note that when this is the case, Pdx+Qdy and must be the same, meaning that and . This implies that . We will now prove that this is also a sufficient condition when the mixed derivative is continuous.


Proof:

First, take the integral

This obviously satisfies the condition that P=.

In order for it to satisfy the other condition, Q(x,y)= meaning that



Canceling Q(x,y) from both sides, we get .

This proves that the equation is exact and that

is an integral of the differential equation.

Note that only C is the arbitrary constant. Changing only changes the integral by a constant value, which is absorbed by the C. Changing will also only change it by a constant because of the fact that .

Example[edit]

Consider the following DE:

Note that:

and 1 is a continuous function so this equation is exact by what has been proven above.

Therefore, the integral is

Take and

Which is

Integrating Factors for an Ordinary Linear Differential Equation of the First Order[edit]

Consider an equation of the form

where P(x), Q(x) and y are all functions of x. This is a first-order linear differential equation as discussed previously. For this to work this form must be closely adhered to - the derivative must be by itself.

In general these equations are not exact. They can, however, be made exact by multiplying through by an integrating factor, I(x), another function of x as we have done previously.

Multiply our original equation by I(x):

(1):

This will be our new, solvable, DE. Now consider the derivative of the product below:

(2):

Now, if we make the RHS of (1) equal to the LHS of (2), then

(3):

Which, by the other halves of the equations, makes:

(4):

Which simplifies to:

(5):

By equating the equations to get (3) forces multiplying by I(x) to produce a derivative of a product on the RHS of (1), i.e.

(6):

The new DE is therefore exact, and can be solved more easily. We now find the function I(x) from (5). We will change notation slightly here.

We take this to be our integrating factor. We can ignore the negative factor, because when both sides of the DE are multiplied by it, they will cancel. So, our integrating factor is:

To solve the DE, we then multiply by this factor, and solve the equation, given that one side will be able to be turned into a derivative of a product.

General Integrating Factors[edit]

Now we generalize first-order linear differential equations to functions of the sort Pdx+Qdy=0 which is our ODE from the beginning this section. They are sometimes not exact, too. However, when multiplied by a function h(x,y), the product hPdx+hQdy=0 may be exact.

Theorem: An equation of the form Pdx+Qdy=0 which has exactly one integral solution with one arbitrary constant C has infinitely many integrating factors.

Proof: Suppose that the solution is f(x,y)=C. The differential is

Since f(x,y)=c is a solution of Pdx+Qdy=0, it must hold true that

Which means that a function h exists such that

and

.

Obviously, this is an integrating factor. Furthermore, let S(f) be any function of f.

Then would equal so hS(f) is also an integrating factor.