# Ordinary Differential Equations/Non Homogenous 1

Higher Order Differential Equations

## Contents

## Non-Homogeneous Equations[edit]

A **non-homogeneous** equation of constant coefficients is an equation of the form

where *c*_{i} are all constants and *f*(*x*) is not 0.

### Complementary Function[edit]

Every non-homogeneous equation has a **complementary function** (CF), which can be found by replacing the *f*(*x*) with 0, and solving for the homogeneous solution. For example, the CF of

is the solution to the differential equation

### Superposition Principle[edit]

The superposition principle makes solving a non-homogeneous equation fairly simple. The final solution is the sum of the solutions to the complementary function, and the solution due to *f*(*x*), called the **particular integral** (PI). In other words,

- General Solution = CF + PI

## Method of Undetermined Coefficients[edit]

The **method of undetermined coefficients** is an easy shortcut to find the particular integral for some *f*(*x*). The method works only if a finite number of derivatives of *f*(*x*) eventually reduces to 0, or if the derivatives eventually fall into a pattern in a finite number of derivatives. If this is true, we then know part of the PI - the sum of all derivatives before we hit 0 (or all the derivatives in the pattern) multiplied by arbitrary constants. This is the **trial** PI. We can then plug our trial PI into the original equation to solve it fully.

As we will see, we may need to alter this trial PI depending on the CF. If the trial PI contains a term that is also present in the CF, then the PI will be absorbed by the arbitrary constant in the CF, and therefore we will not have a full solution to the problem.

*f*(*x*) = Constant[edit]

The simplest case is when *f*(*x*) is constant, for example

- .

The solution to the CF is

We now need to find a trial PI. When we differentiate *y=3*, we get zero. Therefore, our trial PI is the sum of a functions of *y* before this, that is, 3 multiplied by an arbitrary constant, which gives another arbitrary constant, K.

We now set *y* equal to the PI and find the derivatives up to the order of the DE (here, the second).

We can now substitute these into the original DE:

By summing the CF and the PI, we can get the general solution to the DE:

*f*(*x*) = Polynomial[edit]

This is the general method which includes the above example. A polynomial of order *n* reduces to 0 in exactly *n*+1 derivatives (so 1 for a constant as above, three for a quadratic, and so on). So we know that our PI is

For an example, lets take

First off, we know that our PI is

In order to plug in, we need to calculate the first two derivatives of this:

Plugging in we get:

Solving gives us

So, our PI is

However, we need to get the complementary function as well. To get that, set *f*(*x*) to 0 and solve just like we did in the last section. For this equation, the roots are -3 and -2. So that makes our CF

Summing gives us our general solution:

*f*(*x*) = Power of *e*[edit]

Powers of *e* don't ever reduce to 0, but they do become a pattern. In fact it does so in only 1 differentiation, since it's its own derivative. So we know

where K is our constant and *p* is the power of *e* givin in the original DE.

For example, consider

- .

We make our trial PI

- .

Which then gives

Plugging in, we get

That's the particular integral. We found the CF earlier. So the general solution is

*f*(*x*) = Polynomial × Power of *e*[edit]

Polynomials multiplied by powers of *e* also form a loop, in *n* derivatives (where *n* is the highest power of *x* in the polynomial). So we know that our trial PI is

where C is a constant and *p* is the power of *e* in the equation.

For example, lets try

We can now set our PI to

- .

Plugging in, we get

Thats the particular solution. We found the homogeneous solution earlier. So the total solution is

*f*(*x*) = Trigonometric Functions[edit]

Trig functions don't reduce to 0 either. But they do have a loop of 2 derivatives - the derivative of sin *x* is cos *x*, and the derivative of cos *x* is -sin *x*. So we put our PI as

where C is a constant and *p* is the term inside the trig. function in the original DE.

For example, lets try

We set our PI to

Plugging in, we get

Thats the particular solution. We found the homogeneous solution earlier. So the total solution is

*f*(*x*) = A Mixture[edit]

Not only are any of the above solvable by the method of undetermined coefficients, so is the sum of one or more of the above. This is because the sum of two things whose derivatives either go to 0 or loop must also have a derivative that goes to 0 or loops. The would be the sum of the individual functions.

### [edit]

When dealing with , or sometimes with polynomials (if the homogeneous equation has roots of 0) as *f*(*x*), you may get the same term in both the trial PI and the CF. If this happens, the PI will be absorbed into the arbitrary constants of the CF, which will not result in a full solution. To overcome this, multiply the affected terms by *x* as many times as needed until it no longer appears in the CF.

As an example, let's take

First, solve the homogeneous equation to get the CF.

The auxiliary polynomial is

Find the roots of the auxiliary polynomial. In this case, they are

The CF is

Now for the particular integral. Since *f*(*x*) is a polynomial of degree 1, we would normally use A*x*+B. However, since both a term in *x* and a constant appear in the CF, we need to multiply by *x*² and use

- .

We solve this as we normally do for A and B.

So the total solution is

## Variation of Parameters[edit]

Variation of parameters is a method for finding a particular solution to the equation if the general solution for the corresponding homogeneous equation is known. We will now derive this general method.

We already know the general solution of the homogenous equation: it is of the form . We will look for a particular solution of the non-homogenous equation of the form , with u and v functions of the independent variable x. Differentiating this we get

Now notice that there is currently only one condition on , namely that . We now impose another condition, that

This means that will have no second derivatives of and . Thus, these new parameters (hence the name "variation of parameters") will be the solutions to some first order differential equation, which can be solved. Let us finish the problem:

and

,

where the last step follows from the fact that and are solutions of the homogeneous equation.

Therefore, we have and . Multiplying the first equation by and the second by and adding gives

A similar procedure gives

.

Now it is only necessary to evaluate these expressions and integrate them with respect to to get the functions and , and then we have our particular solution . The general solution to the differential equation is therefore .

Note that the main difficulty with this method is that the integrals involved are often extremely complicated. If the integral does not work out well, it is best to use the method of undetermined coefficients instead.

The quantity that appears in the denominator of the expressions for and is called the Wronskian of and .

## Laplace Transforms[edit]

The Laplace transform is a very useful tool for solving nonhomogenous initial-value problems. It allows us to reduce the problem of solving the differential equation to that of solving an algebraic equation. We begin with some setup.

### Definition of the Laplace Transform[edit]

The Laplace transform of is defined as

.

This can also be written as . When writing this on paper, you may write a cursive capital "L" and it will be generally understood. There is also an inverse Laplace transform , but calculating it requires an integration with respect to a complex variable. Luckily, it is frequently possible to find without resorting to this integration, using a variety of tricks which will be described later. However, it is first necessary to prove some facts about the Laplace transform.

### Two Properties of the Laplace Transform[edit]

*Property 1.* The Laplace transform is a linear operator; that is, .

- The proof of this property follows immediately from the definition of the Laplace transform and is left to the reader.

*Property 2.* If , then

- Proof:
- (integrating by parts)

It is property 2 that makes the Laplace transform a useful tool for solving differential equations. As a corollary of property 2, note that .

### Laplace Transforms of Simple Functions[edit]

The last two can be easily calculated using Euler's formula .

In order to find more Laplace transforms, in particular the transform of , we will derive two more properties of the transform.

### Two More Properties of the Laplace Transform[edit]

*Property 3.* If , then .

*Property 4.* If , then .

- The proofs are straightforward and are left to the reader.

Now we can easily see that . Applying Property 3 multiple times, we can find that . At last we are ready to solve a differential equation using Laplace transforms.

### Using Laplace Transforms to Solve Non-Homogeneous Initial-Value Problems[edit]

In general, we solve a second-order linear non-homogeneous initial-value problem as follows: First, we take the Laplace transform of both sides. This immediately reduces the differential equation to an algebraic one. We then solve for . Finally, we take the inverse transform of both sides to find .

Let's begin by using this technique to solve the problem

- .

We begin by taking the Laplace transform of both sides and using property 1 (linearity):

- (using the initial conditions)

Now we isolate :

Here we have factored in preparation for the next step. We now attempt to take the inverse transform of both sides; in order to do this, we will have to break down the right hand side into partial fractions.

The first two fractions imply that . Setting gives , while setting gives . The other three fractions similarly give and . Therefore:

And finally we can take the inverse transform (by inspection, of course) to get

.

### The Convolution[edit]

The convolution is a method of combining two functions to yield a third function. The convolution has applications in probability, statistics, and many other fields because it represents the "overlap" between the functions. We are not concerned with this property here; for us the convolution is useful as a quick method for calculating inverse Laplace transforms.

*Definition.* The convolution is defined as .

The convolution has several useful properties, which are stated below:

*Property 1.* (Associativity)

*Property 2.* (Commutativity)

*Property 3.* (Distribution over addition)

We now prove the result that makes the convolution useful for calculating inverse Laplace transforms.

*Theorem.*

- Proof:
- (changing order of integration)

- Now let .

Let's solve another differential equation:

Taking Laplace transforms of both sides gives

We now have to find . To do this, we notice that , so by the Theorem above. Thus, the solution to our differential equation is the convolution of sine with itself. We proceed to calculate this:

Therefore, the solution to the original equation is

.