Introductory university-level calculus, linear algebra, abstract algebra, probability, statistics, and stochastic processes.
Convergence and divergence of sequences
Get link
Facebook
X
Pinterest
Email
Other Apps
-
Consider the sequence \(a_1, a_2, a_3, \cdots\) where each \(a_i\) is a real number, and there is a real number \(a_n\) for each natural number \(n\). We usually denote such a sequence by \(\{a_n\}\) or \(\{a_n\}_{n=1}^{\infty}\). Note that a sequence is not just a set, its order matters. \(a_n\) may ``converge'' to some real number as \(n\) becomes larger and larger. But what does that mean?
Definition (Limits)
Let \(\{a_n\}\) be a sequence. If \(a_n\) approaches arbitrarily close to a constant value \(\alpha\) as \(n\) becomes arbitrarily large, we call this \(\alpha\) the limit of the sequence \(\{a_n\}\) and write
\[\lim_{n\to\infty}a_n = \alpha\]
or
\[a_n \to \alpha \text{ as } n \to \infty,\]
and we say the sequence \(\{a_n\}\) converges to \(\alpha\). In this case, we also say ``the limit of \(\{a_n\}\) is \(\alpha\).''
If the sequence \(\{a_n\}\) does not converge, we say that the sequence diverges.
Remark. When we say ``\(a_n\) approaches to \(\alpha\) as \(n\to \infty\)'', it means \(|a_n - \alpha|\) becomes smaller as \(n\) becomes larger. But we will make this notion more precise below. □
Example. Let \(\{a_n\}\) be a sequence with each term defined by \(a_n = \frac{(-1)^n}{n}\) for each \(n\in\mathbb{N}\). Then the sequence is
Example. Let \(\{a_n\}\) be a sequence defined by \(a_n = 3n - 1\). We have
\[2, 5, 8, 11, 14, \cdots.\]
The numbers become arbitrarily large and positive numbers. Thus, this sequence diverges. □
In cases such as this last example, we say the sequence \(\{a_n\}\) diverges to the positive infinity and write
\[\lim_{n\to\infty}a_n = \infty\]
or
\[a_n \to \infty ~ (n \to \infty).\]
Example. Let \(\{a_n\}\) be defined by \(a_n = -n^2\) so we have
\[-1, -4, -9, -16, -25, \cdots.\]
The numbers become arbitrarily large negative numbers. So this sequence diverges. □
As in this example, when a sequence diverges to arbitrarily large negative values, we say the sequence \(\{a_n\}\) diverges to the negative infinity, and write
\[\lim_{n\to\infty}a_n = -\infty\]
or
\[a_n \to -\infty ~ (n \to \infty).\]
Example. Let \(\{a_n\}\) be defined by \(a_n = (-1)^n\). We have
\[-1, 1, -1, 1, \cdots\]
and this sequence diverges, but neither to the positive nor negative infinities. □
What do we exactly mean by convergence? Consider the sequence \(a_n = \frac{(-1)^n}{n}\).
\(\lim_{n\to\infty}a_n = 0\) means that \(|a_n| = |a_n - 0|\) becomes arbitrarily small as \(n\) becomes larger. For example,
For \(n > 100\), \(|a_n - 0| < 0.01\).
For \(n > 1,000\), \(|a_n - 0| < 0.001\).
Generalizing these observations, for any arbitrarily small positive number, say \(\varepsilon = 0.0001\), we can always find some natural number, say \(N = 100,000\), such that if \(n > N\), then \(|a_n - 0| < \varepsilon\). Hence the following definition.
Definition (Convergence of a sequence)
The sequence \(\{a_n\}\) is said to converge to a real number \(\alpha\) if and only if the following condition holds:
For any \(\varepsilon > 0\), there exists \(N\in\mathbb{N}\) such that for any \(n\in\mathbb{N}\), if \(n\geq N\), then \(|a_n - \alpha| < \varepsilon\).
(Here, we implicitly assume \(\varepsilon \in\mathbb{R}\).)
Remark. We call this type of argument using \(\varepsilon\) and \(N\) the ``\(\varepsilon-N\) argument.'' □
Remark. In a logical form, the above condition for convergence can be expressed as
Example. If \(a_n = \frac{n}{n+1}\), then \(\{a_n\}\) converges to \(\alpha = 1\). For example, let \(\varepsilon = 0.01\). If \(|a_n - \alpha| < \varepsilon\), then
\[1 - \frac{n}{n+1} < 0.01.\]
By solving this, we have
\[n > 99.\]
Therefore, if we set \(N = 100\), then for any \(n \geq N\), we have \(|a_n - \alpha| < \varepsilon\).
The same procedure can be applied for any values of \(\varepsilon > 0\). □
Example. Let us prove that
\[\lim_{n\to\infty}\frac{1}{n} = 0.\]
By Archimedes' principle, for any given \(\varepsilon > 0\), we can find a natural number \(N\) such that \(\varepsilon N > 1\), and hence \(\frac{1}{N} < \varepsilon\). For \(n\in\mathbb{N}\), if \(n\geq N\), then \(\frac{1}{n} \leq \frac{1}{N}\), so that
Remark. Note that the choice of ``\(N\)'' depends on the value of \(\varepsilon\). For each \(\varepsilon >0\), we choose an appropriate \(N\in\mathbb{N}\). We cannot choose one \(N\in\mathbb{N}\) for all possible values of \(\varepsilon\). If that's the case, then we should have \(\frac{1}{N} < \varepsilon\) for any \(\varepsilon > 0\), which implies \(\frac{1}{N} = 0\), which is nonsense.
Example. Let us prove that \(a_n = (-1)^n\) does not converge by using the \(\varepsilon-N\) argument.
We prove it by contradiction.
Suppose that \(\{a_n\}\) converges to some real number \(\alpha\). Let's pick \(\varepsilon = 1\). There should exist some \(N\in\mathbb{N}\) such that, if \(n\geq N\), then
\[|a_n -\alpha| < 1.\]
When \(n\) is even, \(a_n = 1\) so \(|1 - \alpha| < 1\), in particular,
\[1 - \alpha < 1.\]
This implies that \(\alpha > 0\).
When \(n\) is odd, \(a_n = -1\) so \(|-1 -\alpha| < 1\), in particular,
\[1 + \alpha < 1.\]
This implies \(\alpha < 0\).
Therefore \(\alpha > 0\) and \(\alpha < 0\), which is a contradiction. Hence \(\{a_n\}\) does not converge. □
Definition (Divergence to \(\pm \infty\))
The sequence \(\{a_n\}\) is said to diverge to the positive infinity, denoted \(\lim_{n\to\infty}a_n = +\infty\), if the following condition is satisfied.
For any \(M\in\mathbb{R}\), there exists \(N\in\mathbb{N}\) such that for any \(n\in\mathbb{N}\), if \(n\geq N\), then \(a_n > M\).
Or, in a logical form,
\[\forall M\in\mathbb{R}, \exists N\in\mathbb{N}, \forall n\in\mathbb{N} ~(n \geq N \implies a_n > M).\]
The sequence \(\{a_n\}\) is said to diverge to the negative infinity, denoted \(\lim_{n\to\infty}a_n = -\infty\), if the following condition is satisfied.
For any \(M\in\mathbb{R}\), there exists \(N\in\mathbb{N}\) such that for any \(n\in\mathbb{N}\), if \(n\geq N\), then \(a_n < M\).
Or, in a logical form,
\[\forall M\in\mathbb{R}, \exists N\in\mathbb{N}, \forall n\in\mathbb{N} ~(n \geq N \implies a_n < M).\]
Example. Consider the sequence \(\{3n^2\}\). For a given arbitrary positive real number \(M>0\), let \(N = \left[\sqrt{\frac{M}{3}}\right]+1\) where \(\left[x\right]\) indicates the integer part of \(x\). Note, in particular, that \(N > \sqrt{\frac{M}{3}}\). Now, suppose \(n \geq N\). We have \(3n^2 \geq 3N^2 > M\). Therefore, the sequence \(\{3n^2\}\) diverges to \(+\infty\). □
Defining the birth process Consider a colony of bacteria that never dies. We study the following process known as the birth process , also known as the Yule process . The colony starts with \(n_0\) cells at time \(t = 0\). Assume that the probability that any individual cell divides in the time interval \((t, t + \delta t)\) is proportional to \(\delta t\) for small \(\delta t\). Further assume that each cell division is independent of others. Let \(\lambda\) be the birth rate. The probability of a cell division for a population of \(n\) cells during \(\delta t\) is \(\lambda n \delta t\). We assume that the probability that two or more births take place in the time interval \(\delta t\) is \(o(\delta t)\). That is, it can be ignored. Consequently, the probability that no cell divides during \(\delta t\) is \(1 - \lambda n \delta t - o(\delta t)\). Note that this process is an example of the Markov chain with states \({n_0}, {n_0 + 1}, {n_0 + 2}...
Generational growth Consider the following scenario (see the figure below): A single individual (cell, organism, etc.) produces \(j (= 0, 1, 2, \cdots)\) descendants with probability \(p_j\), independently of other individuals. The probability of this reproduction, \(\{p_j\}\), is known. That individual produces no further descendants after the first (if any) reproduction. These descendants each produce further descendants at the next subsequent time with the same probabilities. This process carries on, creating successive generations. Figure 1. An example of the branching process. Let \(X_n\) be the random variable representing the population size (number of individuals) of generation \(n\). In the above figure, we have \(X_0 = 1\), \(X_1=4\), \(X_2 = 7\), \(X_3=12\), \(X_4 = 9.\) We shall assume \(X_0 = 1\) as the initial condition. Ideally, our goal would be to find how the population size grows through generations, that is, to find the probability \(\Pr(X_n = k)\) for e...
The birth-death process Combining birth and death processes with birth and death rates \(\lambda\) and \(\mu\), respectively, we expect to have the following differential-difference equations for the birth-death process : \[\begin{eqnarray}\frac{{d}p_0(t)}{{d}t} &=& \mu p_1(t),\\\frac{{d}p_n(t)}{{d}t} &=& \lambda(n-1)p_{n-1}(t) - (\lambda + \mu)np_n(t) + \mu(n+1)p_{n+1}(t),~~(n \geq 1).\end{eqnarray}\] You should derive the above equations based on the following assumptions: Given a population with \(n\) individuals, the probability that an individual is born in the population during a short period \(\delta t\) is \(\lambda n \delta t + o(\delta t)\). Given a population with \(n\) individuals, the probability that an individual dies in the population is \(\mu n \delta t + o(\delta t)\). The probability that multiple individuals are born or die during \(\delta t\) is negligible. (The probability of one birth and one death during \(\delta t\) is also negligible.) Consequ...
Comments
Post a Comment