Introductory university-level calculus, linear algebra, abstract algebra, probability, statistics, and stochastic processes.
Convergence and divergence of sequences
Get link
Facebook
X
Pinterest
Email
Other Apps
-
Consider the sequence \(a_1, a_2, a_3, \cdots\) where each \(a_i\) is a real number, and there is a real number \(a_n\) for each natural number \(n\). We usually denote such a sequence by \(\{a_n\}\) or \(\{a_n\}_{n=1}^{\infty}\). Note that a sequence is not just a set, its order matters. \(a_n\) may ``converge'' to some real number as \(n\) becomes larger and larger. But what does that mean?
Definition (Limits)
Let \(\{a_n\}\) be a sequence. If \(a_n\) approaches arbitrarily close to a constant value \(\alpha\) as \(n\) becomes arbitrarily large, we call this \(\alpha\) the limit of the sequence \(\{a_n\}\) and write
\[\lim_{n\to\infty}a_n = \alpha\]
or
\[a_n \to \alpha \text{ as } n \to \infty,\]
and we say the sequence \(\{a_n\}\) converges to \(\alpha\). In this case, we also say ``the limit of \(\{a_n\}\) is \(\alpha\).''
If the sequence \(\{a_n\}\) does not converge, we say that the sequence diverges.
Remark. When we say ``\(a_n\) approaches to \(\alpha\) as \(n\to \infty\)'', it means \(|a_n - \alpha|\) becomes smaller as \(n\) becomes larger. But we will make this notion more precise below. □
Example. Let \(\{a_n\}\) be a sequence with each term defined by \(a_n = \frac{(-1)^n}{n}\) for each \(n\in\mathbb{N}\). Then the sequence is
Example. Let \(\{a_n\}\) be a sequence defined by \(a_n = 3n - 1\). We have
\[2, 5, 8, 11, 14, \cdots.\]
The numbers become arbitrarily large and positive numbers. Thus, this sequence diverges. □
In cases such as this last example, we say the sequence \(\{a_n\}\) diverges to the positive infinity and write
\[\lim_{n\to\infty}a_n = \infty\]
or
\[a_n \to \infty ~ (n \to \infty).\]
Example. Let \(\{a_n\}\) be defined by \(a_n = -n^2\) so we have
\[-1, -4, -9, -16, -25, \cdots.\]
The numbers become arbitrarily large negative numbers. So this sequence diverges. □
As in this example, when a sequence diverges to arbitrarily large negative values, we say the sequence \(\{a_n\}\) diverges to the negative infinity, and write
\[\lim_{n\to\infty}a_n = -\infty\]
or
\[a_n \to -\infty ~ (n \to \infty).\]
Example. Let \(\{a_n\}\) be defined by \(a_n = (-1)^n\). We have
\[-1, 1, -1, 1, \cdots\]
and this sequence diverges, but neither to the positive nor negative infinities. □
What do we exactly mean by convergence? Consider the sequence \(a_n = \frac{(-1)^n}{n}\).
\(\lim_{n\to\infty}a_n = 0\) means that \(|a_n| = |a_n - 0|\) becomes arbitrarily small as \(n\) becomes larger. For example,
For \(n > 100\), \(|a_n - 0| < 0.01\).
For \(n > 1,000\), \(|a_n - 0| < 0.001\).
Generalizing these observations, for any arbitrarily small positive number, say \(\varepsilon = 0.0001\), we can always find some natural number, say \(N = 100,000\), such that if \(n > N\), then \(|a_n - 0| < \varepsilon\). Hence the following definition.
Definition (Convergence of a sequence)
The sequence \(\{a_n\}\) is said to converge to a real number \(\alpha\) if and only if the following condition holds:
For any \(\varepsilon > 0\), there exists \(N\in\mathbb{N}\) such that for any \(n\in\mathbb{N}\), if \(n\geq N\), then \(|a_n - \alpha| < \varepsilon\).
(Here, we implicitly assume \(\varepsilon \in\mathbb{R}\).)
Remark. We call this type of argument using \(\varepsilon\) and \(N\) the ``\(\varepsilon-N\) argument.'' □
Remark. In a logical form, the above condition for convergence can be expressed as
Example. If \(a_n = \frac{n}{n+1}\), then \(\{a_n\}\) converges to \(\alpha = 1\). For example, let \(\varepsilon = 0.01\). If \(|a_n - \alpha| < \varepsilon\), then
\[1 - \frac{n}{n+1} < 0.01.\]
By solving this, we have
\[n > 99.\]
Therefore, if we set \(N = 100\), then for any \(n \geq N\), we have \(|a_n - \alpha| < \varepsilon\).
The same procedure can be applied for any values of \(\varepsilon > 0\). □
Example. Let us prove that
\[\lim_{n\to\infty}\frac{1}{n} = 0.\]
By Archimedes' principle, for any given \(\varepsilon > 0\), we can find a natural number \(N\) such that \(\varepsilon N > 1\), and hence \(\frac{1}{N} < \varepsilon\). For \(n\in\mathbb{N}\), if \(n\geq N\), then \(\frac{1}{n} \leq \frac{1}{N}\), so that
Remark. Note that the choice of ``\(N\)'' depends on the value of \(\varepsilon\). For each \(\varepsilon >0\), we choose an appropriate \(N\in\mathbb{N}\). We cannot choose one \(N\in\mathbb{N}\) for all possible values of \(\varepsilon\). If that's the case, then we should have \(\frac{1}{N} < \varepsilon\) for any \(\varepsilon > 0\), which implies \(\frac{1}{N} = 0\), which is nonsense.
Example. Let us prove that \(a_n = (-1)^n\) does not converge by using the \(\varepsilon-N\) argument.
We prove it by contradiction.
Suppose that \(\{a_n\}\) converges to some real number \(\alpha\). Let's pick \(\varepsilon = 1\). There should exist some \(N\in\mathbb{N}\) such that, if \(n\geq N\), then
\[|a_n -\alpha| < 1.\]
When \(n\) is even, \(a_n = 1\) so \(|1 - \alpha| < 1\), in particular,
\[1 - \alpha < 1.\]
This implies that \(\alpha > 0\).
When \(n\) is odd, \(a_n = -1\) so \(|-1 -\alpha| < 1\), in particular,
\[1 + \alpha < 1.\]
This implies \(\alpha < 0\).
Therefore \(\alpha > 0\) and \(\alpha < 0\), which is a contradiction. Hence \(\{a_n\}\) does not converge. □
Definition (Divergence to \(\pm \infty\))
The sequence \(\{a_n\}\) is said to diverge to the positive infinity, denoted \(\lim_{n\to\infty}a_n = +\infty\), if the following condition is satisfied.
For any \(M\in\mathbb{R}\), there exists \(N\in\mathbb{N}\) such that for any \(n\in\mathbb{N}\), if \(n\geq N\), then \(a_n > M\).
Or, in a logical form,
\[\forall M\in\mathbb{R}, \exists N\in\mathbb{N}, \forall n\in\mathbb{N} ~(n \geq N \implies a_n > M).\]
The sequence \(\{a_n\}\) is said to diverge to the negative infinity, denoted \(\lim_{n\to\infty}a_n = -\infty\), if the following condition is satisfied.
For any \(M\in\mathbb{R}\), there exists \(N\in\mathbb{N}\) such that for any \(n\in\mathbb{N}\), if \(n\geq N\), then \(a_n < M\).
Or, in a logical form,
\[\forall M\in\mathbb{R}, \exists N\in\mathbb{N}, \forall n\in\mathbb{N} ~(n \geq N \implies a_n < M).\]
Example. Consider the sequence \(\{3n^2\}\). For a given arbitrary positive real number \(M>0\), let \(N = \left[\sqrt{\frac{M}{3}}\right]+1\) where \(\left[x\right]\) indicates the integer part of \(x\). Note, in particular, that \(N > \sqrt{\frac{M}{3}}\). Now, suppose \(n \geq N\). We have \(3n^2 \geq 3N^2 > M\). Therefore, the sequence \(\{3n^2\}\) diverges to \(+\infty\). □
Defining the birth process Consider a colony of bacteria that never dies. We study the following process known as the birth process , also known as the Yule process . The colony starts with \(n_0\) cells at time \(t = 0\). Assume that the probability that any individual cell divides in the time interval \((t, t + \delta t)\) is proportional to \(\delta t\) for small \(\delta t\). Further assume that each cell division is independent of others. Let \(\lambda\) be the birth rate. The probability of a cell division for a population of \(n\) cells during \(\delta t\) is \(\lambda n \delta t\). We assume that the probability that two or more births take place in the time interval \(\delta t\) is \(o(\delta t)\). That is, it can be ignored. Consequently, the probability that no cell divides during \(\delta t\) is \(1 - \lambda n \delta t - o(\delta t)\). Note that this process is an example of the Markov chain with states \({n_0}, {n_0 + 1}, {n_0 + 2}...
In mathematics, we must prove (almost) everything and the proofs must be done logically and rigorously. Therefore, we need some understanding of basic logic. Here, I will informally explain some rudimentary formal logic. Definitions (Proposition): A proposition is a statement that is either true or false. "True" and "false" are called the truth values, and are often denoted \(\top\) and \(\bot\). Here is an example. "Dr. Akira teaches at UBD." is a statement that is either true or false (we understand the existence of Dr. Akira and UBD), hence a proposition. The following statement is also a proposition, although we don't know if it's true or false (yet): Any even number greater than or equal to 4 is equal to a sum of two primes. See also: Goldbach's conjecture Next, we define several operations on propositions. Note that propositions combined with these operations are again propositions. (Conjunction, logical "and"): Let \(P\)...
Taylor's theorem As we have seen in a previous post, the essence of differentiation is a linear approximation. See also: Partial and total differentiation of multivariate functions Given a totally differentiable function \(z = f(x,y)\), we have in the neighbor of \(P = (a,b)\), \[f(x,y) = f(a,b) + f_x(a,b)(x - a) + f_y(a,b)(y - b) + o(\|X - P\|)\] where \(X = (x,y).\) By discarding the higher order terms \(o(\|X - P\|)\), we have the equation of the tangent plane at \(P\): \[z = f(a,b) + f_x(a,b)(x - a) + f_y(a,b)(y - b)\] that gives the linear approximation of \(z = f(x,y)\) around the point \(P\). But we may obtain better approximations by looking into the details of \(o(\|X - P\|)\), that is, by incorporating higher-order derivatives. Just as in the univariate case, we have a multivariate version of Taylor's theorem, which can be stated concisely and proved easily by using differential operators. See also: Differential operators Theorem (Taylor's theorem) Let \(f(x,y)...
Comments
Post a Comment