17.1 Second-Order Differential Equations
The easiest second-order differential equations to solve are those which we can integrate directly, for example
When we integrate this once, we get that
and then integrating again, we get that
Which is the general solution to this particular second-order differential equation. With first-order differential equations we only have one constant, and we can determine the value of the constant given a single point on the curve. 11 1 In more formal notation, for the differential equation with solution we can determine the value of given a value of , and the corresponding value of , at that point.
For a second-order differential equation, however, we have two constants, so we definitely can’t solve the equation given only point. We need either two points on the curve, or one point on the original curve and one point on its first derivative.
17.1.2 Homogenous second-order ODEs
These are a bit icky, because they’re not solvable in general. Fortunately a lot of them are solvable.
Homogenous (i.e. everything is a function of only one variable) second-order, linear differential equations can be solved without too much difficulty. We can reduce an equation in the form
to a quadratic by setting .
Here there are a number of possibilities for the value of the discriminant 22 2 If you’ve no idea what this is, review the earlier section on quadratics..
, which is the "straightforward" case
, in which case there’s only one value of
, in which case we can use complex numbers and trigonometry.
In the case where 33 3 note that means ”the disciminant” we have two possible solutions to the differential equation,
In the case where we have only
And in the case where we have the same case as in Equation 17.7, except that there’s a complex part to the root44 4 And, as Euler might point out,
17.1.3 Why two linearly independent solutions?55 5 If two vectors (which we can call and ) are linearly independent, then the only values of and which solve the equation are . If there were a different combination, then we could write that And thus they wouldn’t be linearly independent, as one of the vectors is a multiple of the other. This can be generalised to vectors.
Let’s suppose we have a differential equation involving and its first and second derivative. We may want to solve this subject to the initial conditions
In order to be able to solve for every possible initial condition, we need a linear combination of two "linearly independent" solutions, using which we can satisfy any possible initial condition. This means that if we have two solutions and and neither can be written as a multiple of the other, then the "general solution" (i.e. the one with the constants in it, like for first-order differential equations when we integrate them) is of the form
The good news is that for homogenous second-order differential equations we do have two linearly independent solutions!66 6 Except in the case where , but there’s a way to get around that (more on that later). For example if our roots of the auxiliary are , then we would have that the two solutions are
Proof that we need two linearly independent solutions, and that two linearly independent solutions are sufficient to solve any initial condition (note: no A Level exam board examines this). The reason we need two linearly independent solutions is that if we have two sets of linearly independent initial conditions we certainly can’t write both of them the as linear combinations of a single solution. 88 8 This is a lot like how we can write any vector in a 2D space in terms of two vectors, so long as those vectors are not parallel (same for 3D space, except with three vectors). For example, suppose we have two possible initial conditions, one where
and another where
These are linearly independent solutions, and we can’t write them both as a linear combination of a single solution. Thus, there is at least one case where we need at least two linearly independent solutions.
The next thing to prove is that if we have two solutions ( and ) which are linearly independent, then for suitable values of and we can satisfy any initial condition using a solution of the form
Does this even solve the differential equation though? Yes! We can prove this with a bunch of algebra (there’s an easier way to do this by introducing some new notation, but that’s for later99 9 Note: I have yet to write about this new notation)
and thus that
And thus is a solution to Equation 17.6
If that was messy, it gets worse. 1111 11 I’ve had nightmares about drowning in a differential equation algebra soup. Literal soup made of algebra - it was a very strange dream.
From Equation 17.13 we know that the derivative of our solution will be
as we want to show that this can satisfy any set of initial conditions, where and , we can start by writing a set of simultaneous equations
we can write these in matrix form and obtain that
These equations can be solved whenever the determinant of the 2x2 matrix above is not equal to zero, i.e. whenever
Then the two equations have a unique solution. The easiest way to go from here is to prove this by contradiction (as we have lots of techniques for dealing with equalities () and not very many for dealing with inequalities involving ). We can proceed by assuming that two linearly independent solutions are not sufficient to determine the general solution of any second-order differential equation, and write that
We can now manipulate this a little
Dividing through leads to the formula
Here, though it looks like we have a contraction. Why? Let’s set
Then we can write that
But we specified earlier that and are linearly independent! And now we’ve found that in Equation 17.29 that they’re not linearly independent, and thus we’ve found that assuming that two linearly independent solutions doesn’t solve any set of initial conditions for a second-order differential equation leads to a contradiction. Hence, two linearly independent solutions are sufficient.