Introductory university-level calculus, linear algebra, abstract algebra, probability, statistics, and stochastic processes.
More on the limit of functions
Get link
Facebook
X
Pinterest
Email
Other Apps
-
There are a few variations of limits.
Definition (Divergence of a function)
If the function \(f(x)\) does not converge to any value as \(x \to a\), we say that \(f(x)\) diverges as \(x \to a\). In particular, if the value of \(f(x)\) increases arbitrarily as \(x \to a\), we say that \(f(x)\) diverges to positive infinity and write
\[\lim_{x\to a}f(x) = \infty.\]
If the value of \(f(x)\) is negative and its absolute value increases arbitrarily, we say that \(f(x)\) diverges to negative infinity and write
\[\lim_{x\to a}f(x) = -\infty.\]
Example. Consider the function
\[f(x) = \frac{1}{(x - 1)^2}\]
defined on \(\mathbb{R}\setminus\{1\}\). We have
\[\lim_{x \to 1}f(x) = \infty.\]
You should verify this by drawing a graph. □
There are a few variants of the notion of limit.
Definition (Left and right limits)
We define the left limit of the function \(f(x)\) at \(x=a\) \[\lim_{x \to a -0}f(x) = \alpha\] if the following is satisfied
For any \(\varepsilon > 0\), there exists \(\delta > 0\) such that for any \(x\in\text{dom}(f)\), if \(0 < a - x < \delta\) then \(|f(x) - \alpha| < \varepsilon\)
This mean \(x\) approaches \(a\) from the left (so ``\(x < a\)'' is kept).
We define the right limit of the function \(f(x)\) at \(x=a\) \[\lim_{x \to a +0}f(x) = \alpha\] if the following is satisfied
For any \(\varepsilon > 0\), there exists \(\delta > 0\) such that for any \(x\in\text{dom}(f)\), if \(0 < x - a < \delta\) then \(|f(x) - \alpha| < \varepsilon\).
This mean \(x\) approaches \(a\) from the right (so ``\(a < x\)'' is kept).
Remark. We write \(x \to +0\) instead of \(x \to 0+0\). Similarly \(x\to -0\) rather than \(x \to 0-0\). □
Example. Consider the function
\[f(x) = \frac{x(x + 1)}{|x|}\]
defined on \(\mathbb{R}\setminus\{0\}\).
If \(x > 0\), we have
\[f(x) = \frac{x(x + 1)}{x} = x + 1\]
so that
\[\lim_{x \to +0}f(x) = 1.\]
If \(x < 0\), we have
\[f(x) = \frac{x(x + 1)}{-x} = -x - 1\]
so that
\[\lim_{x \to -0}f(x) = -1.\]
□
Example. For the function \(f(x) = \frac{1}{x}\) defined on \(\mathbb{R}\setminus\{0\}\), we have
We define the limit at \(x \to \infty\) \[\lim_{x \to \infty}f(x) = \alpha\] if the following is satisfied
For any \(\varepsilon > 0\), there exists \(\delta > 0\) such that for any \(x\in\text{dom}(f)\), if \(x > \delta\) then \(|f(x) - \alpha| < \varepsilon\).
We define the limit at \(x \to -\infty\) \[\lim_{x \to -\infty}f(x) = \alpha\] if the following is satisfied
For any \(\varepsilon > 0\), there exists \(\delta > 0\) such that for any \(x\in\text{dom}(f)\), if \(x < -\delta\) then \(|f(x) - \alpha| < \varepsilon\).
Theorem
Let \(f(x)\) be a function defined on the open interval \(I = (a, b)\) \((a < b)\) such that \(\lim_{x\to a+0}f(x) = \alpha\) \((\alpha \in\mathbb{R})\).
For a given \(c\in\mathbb{R}\), if \(f(x) \geq c\) for all \(x\in I\), then \(\alpha \geq c\).
For a given \(d\in\mathbb{R}\), if \(f(x) \leq d\) for all \(x\in I\), then \(\alpha \leq d\).
Proof. We prove only part 1 and prove it by contradiction. Part 2 is similar.
Suppose \(\alpha < c\) and let \(\varepsilon = c - \alpha > 0\). Since \(\lim_{x\to a+0}f(x) = \alpha\), we can find \(\delta > 0\) such that \(0 < x - a < \delta\) implies \(|f(x) - \alpha| < \varepsilon\). But then \(f(x) < \alpha + \varepsilon = c\), which is a contradiction. ■
Theorem
Let \(f(x)\) be a function defined on an interval that includes \(x= a\).
If \(\lim_{x\to a}f(x) = \alpha\) exists, then the corresponding left and right limits also exist and \(\lim_{x\to a+0}f(x) = \lim_{x\to a-0}f(x) = \alpha\).
If both \(\lim_{x\to a+0}f(x)\) and \(\lim_{x\to a-0}f(x)\) exist and \(\lim_{x\to a+0}f(x) = \lim_{x\to a-0}f(x) = \alpha\), then \(\lim_{x\to a}f(x) = \alpha\).
In order to show \(e = \lim_{x\to 0}(1 + x)^{\frac{1}{x}}\), we need to show \(e = \lim_{x\to +0}(1 + x)^{\frac{1}{x}}\) and \(e = \lim_{x\to -0}(1 + x)^{\frac{1}{x}}\). By changing variables \(x = \frac{1}{y}\), \(\lim_{x\to +0}(1 + x)^{\frac{1}{x}} = \lim_{y\to\infty}\left(1 + \frac{1}{y}\right)^y\) and \(\lim_{x\to -0}(1 + x)^{\frac{1}{x}} = \lim_{y\to -\infty}\left(1 + \frac{1}{y}\right)^y\). Furthermore, by changing variables \(y = -t\), assuming \(t > 1\), we have
Since \(\lim_{t\to\infty}\left(1 + \frac{1}{t-1}\right) = 1\), we can see that we only need to show \(e = \lim_{x\to\infty}\left(1 + \frac{1}{x}\right)^{x}\).
For any \(x \geq 1\), we can find \(n\in\mathbb{N}\) such that
A similar result hold for the convergence of functions.
Theorem (Cauchy criterion for the right limit of a function)
Let \(f(x)\) be a function on \(I = (a,b]\). The right limit \(\lim_{x\to a+0}f(x)\) exists if and only if the following condition is satisfied:
For any \(\varepsilon > 0\), there exists a \(\delta > 0\) such that for all \(x, y\in I\), if \(a < x < a + \delta\) and \(a < y < a + \delta\) then \(|f(x) - f(y)| < \varepsilon\). (\(\star\))
Remark. Similarly, the left limit \(\lim_{x\to b-0}f(x)\) exists if and only if
For any \(\varepsilon > 0\), there exists a \(\delta > 0\) such that for all \(x, y\in I\), if \(b - \delta < x < b\) and \(b -\delta < y < b\) then \(|f(x) - f(y)| < \varepsilon\).
The limit \(\lim_{x\to\infty}f(x)\) exists if and only if
For any \(\varepsilon > 0\), there exists a \(\delta > 0\) such that for all \(x, y\in I\), if \(x > \delta\) and \(y > \delta\) then \(|f(x) - f(y)| < \varepsilon\).
The limit \(\lim_{x\to-\infty}f(x)\) exists if and only if
For any \(\varepsilon > 0\), there exists a \(\delta > 0\) such that for all \(x, y\in I\), if \(x < -\delta\) and \(y < -\delta\) then \(|f(x) - f(y)| < \varepsilon\).
□
Proof. (\(\Rightarrow\)) Suppose \(\lim_{x\to a+0}f(x)\) exists. Then there exists some real number \(\alpha\) such that \(\lim_{x\to a+0}f(x) = \alpha\). For any \(\varepsilon > 0\), there exists \(\delta > 0\) such that, for all \(x\in I\), if \(a < x < a + \delta\) then \(|f(x) - \alpha| < \frac{\varepsilon}{2}\). Let \(x\) and \(y\) be arbitrary real numbers such that \(a < x < a + \delta\) and \(a < y < a + \delta\). Then
(\(\Leftarrow\)) Suppose (\(\star\)) is satisfied. Let us define a sequence \(\{a_n\}\) by \(a_n = a + \frac{b-a}{n}\), \(n\in\mathbb{N}\). For any \(\varepsilon > 0\), choose a \(\delta > 0\) satisfying (\(\star\)). By Archimedes' principle, we can find an \(N\in \mathbb{N}\) such that \(N\delta > b - a\). Then for any \(n \geq N\), \(a_n - a = \frac{b-a}{n} < \frac{b-a}{N} < \delta\) so that \(a < a_n < a + \delta\). Therefore by (\(\star\)), for all \(n, m > N\), \(|f(a_n) - f(a_m)| < \varepsilon\) so that \(\{f(a_n)\}\) is a Cauchy sequence, and hence this sequence converges: \(\lim_{n\to\infty}f(a_n) = \alpha\) for some \(\alpha \in \mathbb{R}\).
Now, let us show that \(\lim_{x\to a+0}f(x) = \alpha\). By (\(\star\)), for any \(\varepsilon > 0\), we can find \(\delta > 0\) such that for all \(x,y \in (a, a+\delta)\), \(|f(x) - f(y)|< \frac{\varepsilon}{2}\). Also, since \(\lim_{n\to\infty}f(a_n) = \alpha\), we can find an \(N\in\mathbb{N}\) such that for all \(n \geq N\), \(|f(a_n) - \alpha| < \frac{\varepsilon}{2}\). If \(a < x < a + \delta\), choose \(n \geq N\) such that \(a < a_n < a+\delta\). Then
Defining the birth process Consider a colony of bacteria that never dies. We study the following process known as the birth process , also known as the Yule process . The colony starts with \(n_0\) cells at time \(t = 0\). Assume that the probability that any individual cell divides in the time interval \((t, t + \delta t)\) is proportional to \(\delta t\) for small \(\delta t\). Further assume that each cell division is independent of others. Let \(\lambda\) be the birth rate. The probability of a cell division for a population of \(n\) cells during \(\delta t\) is \(\lambda n \delta t\). We assume that the probability that two or more births take place in the time interval \(\delta t\) is \(o(\delta t)\). That is, it can be ignored. Consequently, the probability that no cell divides during \(\delta t\) is \(1 - \lambda n \delta t - o(\delta t)\). Note that this process is an example of the Markov chain with states \({n_0}, {n_0 + 1}, {n_0 + 2}...
Generational growth Consider the following scenario (see the figure below): A single individual (cell, organism, etc.) produces \(j (= 0, 1, 2, \cdots)\) descendants with probability \(p_j\), independently of other individuals. The probability of this reproduction, \(\{p_j\}\), is known. That individual produces no further descendants after the first (if any) reproduction. These descendants each produce further descendants at the next subsequent time with the same probabilities. This process carries on, creating successive generations. Figure 1. An example of the branching process. Let \(X_n\) be the random variable representing the population size (number of individuals) of generation \(n\). In the above figure, we have \(X_0 = 1\), \(X_1=4\), \(X_2 = 7\), \(X_3=12\), \(X_4 = 9.\) We shall assume \(X_0 = 1\) as the initial condition. Ideally, our goal would be to find how the population size grows through generations, that is, to find the probability \(\Pr(X_n = k)\) for e...
The birth-death process Combining birth and death processes with birth and death rates \(\lambda\) and \(\mu\), respectively, we expect to have the following differential-difference equations for the birth-death process : \[\begin{eqnarray}\frac{{d}p_0(t)}{{d}t} &=& \mu p_1(t),\\\frac{{d}p_n(t)}{{d}t} &=& \lambda(n-1)p_{n-1}(t) - (\lambda + \mu)np_n(t) + \mu(n+1)p_{n+1}(t),~~(n \geq 1).\end{eqnarray}\] You should derive the above equations based on the following assumptions: Given a population with \(n\) individuals, the probability that an individual is born in the population during a short period \(\delta t\) is \(\lambda n \delta t + o(\delta t)\). Given a population with \(n\) individuals, the probability that an individual dies in the population is \(\mu n \delta t + o(\delta t)\). The probability that multiple individuals are born or die during \(\delta t\) is negligible. (The probability of one birth and one death during \(\delta t\) is also negligible.) Consequ...
Comments
Post a Comment