Introductory university-level calculus, linear algebra, abstract algebra, probability, statistics, and stochastic processes.
Special solution of inhomogeneous linear differential equations
Get link
Facebook
X
Pinterest
Email
Other Apps
-
The general solution of an inhomogeneous linear differential equation can be obtained as the sum of a special solution and the general solution of the corresponding homogeneous differential equation. We study the method of variation of parameters in particular.
Consider the linear differential equation of the form
\[F(D)y = q(x)\]
where \(F(t)\) is a polynomial and \(q(x)\) is a function. When \(q(x) = 0\), this is a homogeneous linear differential equation. Here, we assume \(q(x) \neq 0\).
Suppose we can factorize \(F(t)\) as \(F(t) = G(t)H(t)\). If \(y = y(x)\) is a solution of \(F(D)y = q\), then \(z = H(D)y\) is a solution of \(G(D)z = q\). Thus, the given differential equation \(F(D)y = q\) is decomposed into two parts:
\(G(D)z = q\) (a linear differential equation of \(z\), given \(q\)),
\(H(D)y = z\) (a linear differential equation of \(y\), given \(z\)),
and we can process one after the other. Thus, by factorizing the polynomial \(F(D)\), we only need to consider the case where \(F(t) = t - \alpha\) (\(\alpha \in \mathbb{C}\)). That is, \(y' - \alpha y = q.\) This ODE can be readily solved by using the method of variation of parameters. The general solution of this ODE is obtained as
Note that \(F(t) = t^2 + t - 6 = (t - 2)(t + 3)\) so the given ODE is \((D - 2)(D+3)y = \cos x\). Solving the homogeneous ODE \((D - 2)(D + 3)y = 0\), we have
\(y = Ae^{2x} + Be^{-3x}\) where \(A, B\) are constants. Let \(z = (D+3)y\), then we have \((D-2)z = \cos x\) and its special solution is given by
\[z = e^{2x}\int e^{-2x}\cos x dx = -\frac{2}{5}\cos x + \frac{1}{5}\sin x.\]
Similarly, solving \((D+3)y = -\frac{2}{5}\cos x + \frac{1}{5}\sin x\), we have a special solution
\[y = e^{-3x}\int e^{3x}\left(-\frac{2}{5}\cos x + \frac{1}{5}\sin x\right)\,dx = -\frac{7}{50}\cos x + \frac{1}{50}\sin x.\]
Thus, the general solution is
\[y = -\frac{7}{50}\cos x + \frac{1}{50}\sin x + Ae^{2x} + Be^{-3x}\]
where \(A, B\) are constants. □
Let's consider the more general case where \(F(t) = t^2 + at + b = (t - \alpha)(t - \beta)\) with \(\alpha \neq \beta\). In particular, when \(a^2 - 4b < 0\), we have \(\beta = \bar{\alpha}\). We want to solve the following second-order ODE:
\[(D - \alpha)(D - \beta)y = q.\]
Let \(z = (D-\beta)y\), and we have
\[(D - \alpha)z = q.\tag{eq:odea}\]
We first solve the homogeneous case \((D - \alpha)z = 0\) which yields
\[z = Ae^{\alpha x}.\]
Now, regarding \(A\) as a function of \(x\), substitute it into (eq:odea), we have
Note the symmetry between the two terms on the right-hand side. This means that swapping \(\alpha\) and \(\beta\) does not change the result. This is expected as the differential operators are commutative: \((D-\alpha)(D - \beta) = (D -\beta)(D - \alpha)\).
Example. Let us solve
\[y'' + y = \frac{1}{\cos x}.\]
The corresponding homogeneous equation is \(y'' + y = 0\). Noting \(D^2 + 1 = (D -i)(D+i)\), its solution is
\[y = Ae^{ix} + Be^{-ix} \tag{eg:hom}\]
where \(A, B \in \mathbb{C}\) are constants. Now, we use the method of variation of parameters by regarding \(A\) and \(B\) as functions of \(x\). According to the above result, we have
Using Euler's formula (\(e^{\pm i x} = \cos x \pm i \sin x\)), this can be further "simplified" (exercise!) as
\[y = (\log|\cos x| + C)\cos x + (x + D)\sin x\]
where \(C\) and \(D\) are constants. Note that this is a real-valued function if \(C\) and \(D\) are real. □
Next, let's consider \((D-\alpha)^2y = q.\) Of course, \((t - \alpha)^2 = (t-\alpha)(t-\alpha)\) so that we can apply the technique above. Let \(y_1 = (D-\alpha)y\), and we solve \((D-\alpha)y_1 = q\). As we have seen above,
\[y_1 = e^{\alpha x}\int e^{-\alpha x}q(x)\,dx.\]
Next, we solve \((D-\alpha)y_2 = y_1\). This gives
We can continue the same process to solve \((D-\alpha)^my = q\) for any \(m = 3, 4, \cdots\). Let the solution of \((D-\alpha)^my = q\) be \(y_m\) for \(m = 0, 1, 2, \cdots\). We can see that
Defining the birth process Consider a colony of bacteria that never dies. We study the following process known as the birth process , also known as the Yule process . The colony starts with \(n_0\) cells at time \(t = 0\). Assume that the probability that any individual cell divides in the time interval \((t, t + \delta t)\) is proportional to \(\delta t\) for small \(\delta t\). Further assume that each cell division is independent of others. Let \(\lambda\) be the birth rate. The probability of a cell division for a population of \(n\) cells during \(\delta t\) is \(\lambda n \delta t\). We assume that the probability that two or more births take place in the time interval \(\delta t\) is \(o(\delta t)\). That is, it can be ignored. Consequently, the probability that no cell divides during \(\delta t\) is \(1 - \lambda n \delta t - o(\delta t)\). Note that this process is an example of the Markov chain with states \({n_0}, {n_0 + 1}, {n_0 + 2}...
In mathematics, we must prove (almost) everything and the proofs must be done logically and rigorously. Therefore, we need some understanding of basic logic. Here, I will informally explain some rudimentary formal logic. Definitions (Proposition): A proposition is a statement that is either true or false. "True" and "false" are called the truth values, and are often denoted \(\top\) and \(\bot\). Here is an example. "Dr. Akira teaches at UBD." is a statement that is either true or false (we understand the existence of Dr. Akira and UBD), hence a proposition. The following statement is also a proposition, although we don't know if it's true or false (yet): Any even number greater than or equal to 4 is equal to a sum of two primes. See also: Goldbach's conjecture Next, we define several operations on propositions. Note that propositions combined with these operations are again propositions. (Conjunction, logical "and"): Let \(P\)...
Taylor's theorem As we have seen in a previous post, the essence of differentiation is a linear approximation. See also: Partial and total differentiation of multivariate functions Given a totally differentiable function \(z = f(x,y)\), we have in the neighbor of \(P = (a,b)\), \[f(x,y) = f(a,b) + f_x(a,b)(x - a) + f_y(a,b)(y - b) + o(\|X - P\|)\] where \(X = (x,y).\) By discarding the higher order terms \(o(\|X - P\|)\), we have the equation of the tangent plane at \(P\): \[z = f(a,b) + f_x(a,b)(x - a) + f_y(a,b)(y - b)\] that gives the linear approximation of \(z = f(x,y)\) around the point \(P\). But we may obtain better approximations by looking into the details of \(o(\|X - P\|)\), that is, by incorporating higher-order derivatives. Just as in the univariate case, we have a multivariate version of Taylor's theorem, which can be stated concisely and proved easily by using differential operators. See also: Differential operators Theorem (Taylor's theorem) Let \(f(x,y)...
Comments
Post a Comment