For example, these boundary problems came up in the study of the heat equation \(u_t =
k u_{xx}\) when we were trying to solve the equation by the method of separation of variables in Section 5.6. Dirichlet conditions correspond to applying a zero temperature at the ends, Neumann means insulating the ends, etc. Other types of endpoint conditions also arise naturally, such as the Robin boundary conditions
for some constant \(h\text{.}\) These conditions come up when the ends are immersed in some medium.
In the separation of variables computation we encountered an eigenvalue problem and found the eigenfunctions \(X_n(x)\text{.}\) We then found the eigenfunction decomposition of the initial temperature \(f(x) = u(x,0)\text{,}\)
Once we had this decomposition and found suitable \(T_n(t)\) such that \(T_n(0) = 1\) and such that \(T_n(t)X_n(x)\) were solutions to the heat equation, we wrote the solution to the original problem, including the initial condition, as
To study more general problems with this method, we must study more general eigenvalue problems. First, we study second order linear equations of the form
\begin{equation}
\frac{d}{dx} \left( p(x) \frac{dy}{dx} \right)
- q(x) y + \lambda r(x) y = 0 .\tag{6.1}
\end{equation}
Essentially any second order linear equation of the form \(a(x) y'' + b(x) y' + c(x) y + \lambda d(x) y = 0\) can be written as (6.1) after multiplying by a proper factor.
Example6.1.1.
[Bessel] Put the following equation into the form (6.1):
\begin{equation*}
\begin{split}
\frac{1}{x} \left( x^2 y'' + xy' + \left(\lambda x^2 - n^2\right)y \right)
& =
x y'' + y' + \left(\lambda x - \frac{n^2}{x}\right)y
\\
& =
\frac{d}{dx} \left( x \frac{dy}{dx} \right)
- \frac{n^2}{x} y + \lambda x y = 0.
\end{split}
\end{equation*}
The Bessel equation turns up for example in the solution of the two-dimensional wave equation. If you want to see how one solves the equation, you can look at Subsection 4.3.3.
The so-called Sturm–Liouville problem 1 is to seek nontrivial solutions to
In particular, we seek \(\lambda\)s that allow for nontrivial solutions. The \(\lambda\)s that admit nontrivial solutions are called the eigenvalues and the corresponding nontrivial solutions are called eigenfunctions. The constants \(\alpha_1\) and \(\alpha_2\) should not be both zero, same for \(\beta_1\) and \(\beta_2\text{.}\)
Theorem6.1.1.
Suppose \(p(x)\text{,}\)\(p'(x)\text{,}\)\(q(x)\) and \(r(x)\) are continuous on \([a,b]\) and suppose \(p(x) > 0\) and \(r(x) > 0\) for all \(x\) in \([a,b]\text{.}\) Then the Sturm–Liouville problem (6.2) has an increasing sequence of eigenvalues
and such that to each \(\lambda_n\) there is (up to a constant multiple) a single eigenfunction \(y_n(x)\text{.}\)
Moreover, if \(q(x) \geq 0\) and \(\alpha_1, \alpha_2, \beta_1, \beta_2 \geq 0\text{,}\) then \(\lambda_n \geq 0\) for all \(n\text{.}\)
Problems satisfying the hypothesis of the theorem (including the “Moreover”) are called regular Sturm–Liouville problems, and we will only consider such problems here. That is, a regular problem is one where \(p(x)\text{,}\)\(p'(x)\text{,}\)\(q(x)\) and \(r(x)\) are continuous, \(p(x) > 0\text{,}\)\(r(x) > 0\text{,}\)\(q(x) \geq 0\text{,}\) and \(\alpha_1, \alpha_2, \beta_1, \beta_2 \geq 0\text{,}\) where neither \(\alpha_1\) and \(\alpha_2\) are both zero, nor \(\beta_1\) and \(\beta_2\) are both zero. Note: Be careful about the signs. Also be careful about the inequalities for \(r\) and \(p\text{,}\) they must be strict for all \(x\) in the interval \([a,b]\text{,}\) including the endpoints!
When zero is an eigenvalue, we usually start labeling the eigenvalues at 0 rather than at 1 for convenience. That is we label the eigenvalues \(\lambda_0 < \lambda_1 < \lambda_2 < \cdots\text{.}\)
Example6.1.2.
The problem \(y''+\lambda y\text{,}\)\(0 < x < L\text{,}\)\(y(0) = 0\text{,}\) and \(y(L) = 0\) is a regular Sturm–Liouville problem: \(p(x) = 1\text{,}\)\(q(x) = 0\text{,}\)\(r(x) = 1\text{,}\) and we have \(p(x) = 1 > 0\) and \(r(x) = 1 > 0\text{.}\) We also have \(a=0\text{,}\)\(b=L\text{,}\)\(\alpha_1 = \beta_1 = 1\text{,}\)\(\alpha_2 = \beta_2 = 0\text{.}\) The eigenvalues are \(\lambda_n = \frac{n^2 \pi^2}{L^2}\) and eigenfunctions are \(y_n(x) = \sin\bigl(\frac{n\pi}{L} x\bigr)\text{.}\) All eigenvalues are nonnegative as predicted by the theorem.
Identify the \(p, q, r, \alpha_j, \beta_j\text{.}\) Can you use the theorem above to make the search for eigenvalues easier? Hint: Consider the condition \(-y'(0)=0\text{.}\)
Example6.1.3.
Find eigenvalues and eigenfunctions of the problem
\begin{equation*}
\begin{aligned}
& y''+\lambda y = 0, \quad 0 < x < 1 , \\
& hy(0)- y'(0) = 0, \quad y'(1) = 0, \quad h > 0.
\end{aligned}
\end{equation*}
These equations give a regular Sturm–Liouville problem.
By Theorem 6.1.1, \(\lambda \geq 0\text{.}\) So the general solution (without boundary conditions) is
\begin{equation*}
\begin{aligned}
& y(x) = A \cos ( \sqrt{\lambda}\, x) + B \sin (
\sqrt{\lambda}\, x) & & \qquad \text{if } \; \lambda > 0 , \\
& y(x) = A x + B & & \qquad \text{if } \; \lambda = 0 .
\end{aligned}
\end{equation*}
Let us see if \(\lambda = 0\) is an eigenvalue: We must satisfy \(0 = hB - A\) and \(A = 0\text{,}\) hence \(B=0\) (as \(h > 0\)). Therefore, 0 is not an eigenvalue (no nonzero solution, so no eigenfunction).
Now let us try \(\lambda > 0\text{.}\) We plug in the boundary conditions:
\begin{equation*}
\begin{aligned}
& 0 = h A - \sqrt{\lambda}\, B , \\
& 0 = -A \sqrt{\lambda}\, \sin (\sqrt{\lambda}) +B \sqrt{\lambda}\,
\cos (\sqrt{\lambda}) .
\end{aligned}
\end{equation*}
If \(A=0\text{,}\) then \(B=0\) and vice versa, hence both are nonzero. So \(B = \frac{hA}{\sqrt{\lambda}}\text{,}\) and \(0 = -A \sqrt{\lambda}\, \sin ( \sqrt{\lambda}) + \frac{hA}{\sqrt{\lambda}}
\sqrt{\lambda}\, \cos ( \sqrt{\lambda})\text{.}\) As \(A \not= 0\) we get
We use a computer to find \(\lambda_n\text{.}\) There are tables available, though using a computer or a graphing calculator is far more convenient nowadays. Easiest method is to plot the functions \(\nicefrac{h}{x}\) and \(\tan x\) and see for which \(x\) they intersect. There is an infinite number of intersections. Denote the first intersection by \(\sqrt{\lambda_1}\text{,}\) the second intersection by \(\sqrt{\lambda_2}\text{,}\) etc. For example, when \(h=1\text{,}\) we get \(\sqrt{\lambda_1} \approx 0.86\text{,}\)\(\sqrt{\lambda_2} \approx 3.43\text{,}\) .... That is \(\lambda_1 \approx 0.74\text{,}\)\(\lambda_2 \approx 11.73\text{,}\) .... A plot for \(h=1\) is given in Figure 6.1. The appropriate eigenfunction (let \(A = 1\) for convenience, then \(B=\nicefrac{h}{\sqrt{\lambda}}\)) is
\begin{equation*}
y_n(x) = \cos ( \sqrt{\lambda_n}\, x ) + \frac{h}{\sqrt{\lambda_n}}
\sin (\sqrt{\lambda_n} \, x ) .
\end{equation*}
When \(h=1\) we get (approximately)
\begin{equation*}
y_1(x) \approx \cos (0.86\, x ) + \frac{1}{0.86}
\sin (0.86 \, x ) , \qquad
y_2(x) \approx \cos (3.43\, x ) + \frac{1}{3.43}
\sin (3.43 \, x ) , \qquad \ldots .
\end{equation*}
Subsection6.1.2Orthogonality
We have seen the notion of orthogonality before. For example, we have shown that \(\sin (nx)\) are orthogonal for distinct \(n\) on \([0,\pi]\text{.}\) For general Sturm–Liouville problems we need a more general setup. Let \(r(x)\) be a weight function (any function, though generally we assume it is positive) on \([a,b]\text{.}\) Two functions \(f(x)\text{,}\)\(g(x)\) are said to be orthogonal with respect to the weight function \(r(x)\) when
\begin{equation*}
\langle f , g \rangle \overset{\text{def}}{=} \int_a^b f(x) \, g(x) \, r(x)
\,dx ,
\end{equation*}
and then say \(f\) and \(g\) are orthogonal whenever \(\langle f , g \rangle = 0\text{.}\) The results and concepts are again analogous to finite-dimensional linear algebra.
The idea of the given inner product is that those \(x\) where \(r(x)\) is greater have more weight. Nontrivial (nonconstant) \(r(x)\) arise naturally, for example from a change of variables. Hence, you could think of a change of variables such that \(d\xi = r(x)\, dx\text{.}\)
Eigenfunctions of a regular Sturm–Liouville problem satisfy an orthogonality property, just like the eigenfunctions in Section 5.1. Its proof is very similar to the analogous Theorem 5.1.1.
that is, \(y_j\) and \(y_k\) are orthogonal with respect to the weight function \(r\text{.}\)
Subsection6.1.3Fredholm alternative
The Fredholm alternative theorem we talked about before (Theorem 5.1.2) holds for all regular Sturm–Liouville problems. We state it here for completeness.
Theorem6.1.3.Fredholm alternative.
Suppose that we have a regular Sturm–Liouville problem. Then either
has a unique solution for any \(f(x)\) continuous on \([a,b]\text{.}\)
This theorem is used in much the same way as we did before in Section 5.4. It is used when solving more general nonhomogeneous boundary value problems. The theorem does not help us solve the problem, but it tells us when a unique solution exists, so that we know when to spend time looking for it. To solve the problem we decompose \(f(x)\) and \(y(x)\) in terms of eigenfunctions of the homogeneous problem, and then solve for the coefficients of the series for \(y(x)\text{.}\)
Subsection6.1.4Eigenfunction series
What we want to do with the eigenfunctions once we have them is to compute the eigenfunction decomposition of an arbitrary function \(f(x)\text{.}\) That is, we wish to write
where \(y_n(x)\) are eigenfunctions. We wish to find out if we can represent any function \(f(x)\) in this way, and if so, we wish to calculate \(c_n\) (and of course we would want to know if the sum converges). OK, so imagine we could write \(f(x)\) as (6.3). We will assume convergence and the ability to integrate the series term by term. Because of orthogonality we have
Note that \(y_m\) are known up to a constant multiple, so we could have picked a scalar multiple of an eigenfunction such that \(\langle y_m , y_m \rangle = 1\) (if we had an arbitrary eigenfunction \(\tilde{y}_m\text{,}\) divide it by \(\sqrt{\langle \tilde{y}_m , \tilde{y}_m \rangle}\)). When \(\langle y_m , y_m \rangle = 1\) we have the simpler form \(c_m = \langle f, y_m \rangle\text{.}\) The following theorem holds more generally, but the statement given is enough for our purposes.
Theorem6.1.4.
Suppose \(f\) is a piecewise smooth continuous function on \([a,b]\text{.}\) If \(y_1,
y_2, \ldots\) are eigenfunctions of a regular Sturm–Liouville problem, one for each eigenvalue, then there exist real constants \(c_1, c_2, \ldots\) given by (6.4) such that (6.3) converges and holds for \(a < x < b\text{.}\)
The above is a regular Sturm–Liouville problem, and Theorem 6.1.1 says that if \(\lambda\) is an eigenvalue then \(\lambda \geq 0\text{.}\)
Suppose \(\lambda = 0\text{.}\) The general solution is \(y(x) = Ax + B\text{.}\) We plug in the initial conditions to get \(0=y(0) = B\text{,}\) and \(0 = y'(\nicefrac{\pi}{2}) = A\text{.}\) Hence \(\lambda = 0\) is not an eigenvalue.
So let us consider \(\lambda > 0\text{,}\) where the general solution is
\begin{equation*}
y(x) = A \cos ( \sqrt{\lambda} \, x ) + B \sin ( \sqrt{\lambda} \, x) .
\end{equation*}
Plugging in the boundary conditions we get \(0 = y(0) = A\) and \(0 = y'(\nicefrac{\pi}{2})
= \sqrt{\lambda} \, B \cos \bigl(\sqrt{\lambda} \, \frac{\pi}{2}\bigr)\text{.}\) Since \(A\) is zero, then \(B\) cannot be zero. Hence \(\cos \bigl( \sqrt{\lambda} \,
\frac{\pi}{2}\bigr) = 0\text{.}\) This means that \(\sqrt{\lambda} \,\frac{\pi}{2}\) is an odd integral multiple of \(\nicefrac{\pi}{2}\text{,}\) i.e. \((2n-1)\frac{\pi}{2} = \sqrt{\lambda_n} \,\frac{\pi}{2}\text{.}\) Solving for \(\lambda_n\) we get
Note that the series converges to an odd \(2\pi\)-periodic extension of \(f(x)\text{.}\) With the regular sine series we would expect a function with period \(2 \, \frac{\pi}{2} = \pi\text{.}\)
Let us compute an example. Consider \(f(x) = x\) for \(0 < x < \nicefrac{\pi}{2}\text{.}\) Some calculus later we find
Both sums converge are equal to \(f(x)\) for \(0 < x < \nicefrac{\pi}{2}\text{,}\) but the eigenfunctions involved come from different eigenvalue problems.
Suppose that you had a Sturm–Liouville problem on the interval \([0,1]\) and came up with \(y_n(x) = \sin (\gamma n x)\text{,}\) where \(\gamma > 0\) is some constant. Decompose \(f(x) = x\text{,}\)\(0 < x < 1\) in terms of these eigenfunctions.
Hint: First write the system as a constant coefficient system to find general solutions. Do note that Theorem 6.1.1 guarantees \(\lambda \geq 0\text{.}\)
Put the following problems into the standard form for Sturm–Liouville problems, that is, find \(p(x)\text{,}\)\(q(x)\text{,}\)\(r(x)\text{,}\)\(\alpha_1\text{,}\)\(\alpha_2\text{,}\)\(\beta_1\text{,}\) and \(\beta_2\text{,}\) and decide if the problems are regular or not.
\(x y'' + \lambda y = 0\) for \(0 < x < 1\text{,}\)\(y(0) = 0\text{,}\)\(y(1) = 0\text{.}\)
\((1+x^2) y'' + 2xy' + (\lambda-x^2) y = 0\) for \(-1 < x < 1\text{,}\)\(y(-1) = 0\text{,}\)\(y(1)+y'(1) = 0\text{.}\) 4
a) \(p(x) = 1\text{,}\)\(q(x) = 0\text{,}\)\(r(x) = \frac{1}{x}\text{,}\)\(\alpha_1 = 1\text{,}\)\(\alpha_2 =
0\text{,}\)\(\beta_1 = 1\text{,}\)\(\beta_2 = 0\text{.}\) The problem is not regular. b) \(p(x) = 1+x^2\text{,}\)\(q(x) = x^2\text{,}\)\(r(x) = 1\text{,}\)\(\alpha_1 = 1\text{,}\)\(\alpha_2 =
0\text{,}\)\(\beta_1 = 1\text{,}\)\(\beta_2 = 1\text{.}\) The problem is regular.
In an earlier version of this book, a typo rendered the equation as \((1+x^2) y'' - 2xy' + (\lambda-x^2) y = 0\) ending up with something harder than intended. Try this equation for a further challenge.
For a higher quality printout use the PDF version: https://www.jirka.org/diffyqs/diffyqs.pdf