Branching processes: Mean and variance
Generational growth
Consider the following scenario (see the figure below):
- A single individual (cell, organism, etc.) produces \(j (= 0, 1, 2, \cdots)\) descendants with probability \(p_j\), independently of other individuals.
- The probability of this reproduction, \(\{p_j\}\), is known.
- That individual produces no further descendants after the first (if any) reproduction.
- These descendants each produce further descendants at the next subsequent time with the same probabilities.
- This process carries on, creating successive generations.
Let's consider the probability generating function for \(p_j\):
\[G(s) = \sum_{j=0}^{\infty}p_js^j.\]
Since we are assuming \(X_0 = 1\), \(G(s)\) is the generating function for the random variable \(X_1\) (this means \(\Pr(X_1 = j) = p_j\)).
Let \(G_2(s)\) be the generating function of the random variable \(X_2\) (the population size of the second generation). Since there are \(X_1\) individuals, assuming each produce \(Y_k (k=1, 2, \cdots, X_1)\) offsprings, we have
\[X_2 = Y_1 + Y_2 + \cdots + Y_{X_1}.\]
In the case of the above Figure 1, we have a realized value of \(X_2\) as
\[x_2 = y_1 + y_2 + y_3 + y_4 = 4 + 1 + 0 + 2 = 7.\]
In general,
\[\Pr(Y_k = j) = p_j,~~ \Pr(X_1 = m) = \Pr(X_1 = m|X_0 = 1) = p_m.\]
(Recall that \(X_0 = 1\) is the assumed initial condition.)
Let \[h_n = \Pr(X_2 = n).\]
Using the partition theorem (or the law of total probability), we have
\[\begin{eqnarray} h_n &=& \Pr(X_2 = n) = \sum_{r=0}^{\infty}\Pr(X_2=n|X_1 = r)\Pr(X_1 = r)\\ &=&\sum_{r=0}^{\infty}p_r\Pr(X_2=n|X_1 = r). \end{eqnarray}\]
Multiply both sides by \(s^n\) and sum over \(n\),
\[G_2(s) = \sum_{n=0}^{\infty}h_ns^n = \sum_{r=0}^{\infty}p_r\sum_{n=0}^{\infty}P(X_2=n|X_1=r)s^n.\]
The last factor (the second sum) can be calculated as follows.
\[\begin{eqnarray} \sum_{n=0}^{\infty}P(X_2=n|X_1=r)s^n &=& \mathbb{E}(s^{X_2}|X_1 = r) ~~~ \text{(conditional expectation)}\\ &=& \mathbb{E}(s^{Y_1 + Y_2 + \cdots + Y_r})\nonumber\\ &=& \mathbb{E}(s^{Y_1})\mathbb{E}(s^{Y_2})\cdots \mathbb{E}(s^{Y_r})~~~\text{(by independence)}\nonumber\\ &=& G(s)G(s)\cdots G(s)~~~\text{(identically distributed)}\nonumber\\ &=& [G(s)]^r \end{eqnarray}\]
since \(\mathbb{E}(s^{Y_1}) = \mathbb{E}(s^{Y_2}) = \cdots = G(s)\). Each \(Y_k\) is (the random variable representing) the number of offsprings produced by an individual, and they are i.i.d. Therefore,
\[G_2(s) = \sum_{r=0}^{\infty}p_r[G(s)]^{r} = G(G(s)).\]
Since \(G(1) = 1\), we have \(G(G(1)) = G(1) = 1\).
The same result holds for any successive generations. Let \(G_m(s)\) be the generating function of \(X_m\) (the population size of generation \(m\)), then
\[G_m(s) = G_{m-1}(G(s)) = G(G(\cdots(G(s))\cdots))\]
where \(G(s)\) is nested \(m\) times. Note that \(G_m(1) = 1\) follows from $G(1) = 1$.
Mean and variance
Let \(\mu_n\) and \(\sigma_n^2\) be the mean and variance of the population size of the \(n\)-th generation. Using the pgf \(G_n(s)\), we have
\[\begin{eqnarray} \mu_n &=& \mathbb{E}(X_n) = G_n'(1),\\ \sigma_n^2 &=& \mathbb{V}(X_n) = G_n''(1) + \mu_n - \mu_n^2. \end{eqnarray}\]
It is understood that \(G_1(s) = G(s)\). For \(n = 2\), we have
\[\begin{eqnarray} \mu_2 &=& \frac{{d}}{{d}s}G_2(s)|_{s=1} = \frac{{d}}{{d}s}[G(G(s))]_{s=1} \nonumber\\ &=& G'(1)G'(1) = \mu_1^2. \end{eqnarray}\]
In general,
\[\begin{eqnarray} \mu_n &=& \left.\frac{\mathrm{d}G_n(s)}{\mathrm{d}s}\right|_{s=1}\nonumber\\ &=& \left.\frac{\mathrm{d}G(G_{n-1}(s))}{\mathrm{d}s}\right|_{s=1}\nonumber\\ &=& G'(1)G_{n-1}'(1) = \mu_1\mu_{n-1} = \mu_1^2\mu_{n-2} = \cdots = \mu_1^{n}. \end{eqnarray}\]
Exercise. Prove this by mathematical induction. □
To obtain the variance \(\sigma_n^2\), we need second derivatives.
\[\begin{eqnarray} G_n'(s) &=& G'(G_{n-1}(s))G_{n-1}'(s),\\ G_n''(s) &=& G''(G_{n-1}(s))[G_{n-1}'(s)]^2 + G'(G_{n-1}(s))G_{n-1}''(s). \end{eqnarray}\]
With \(s=1\) and recalling \(G_n(1) = 1\) for any \(n\), we have
\[\begin{eqnarray} G_n''(1) &=& G''(G_{n-1}(1))[G_{n-1}'(1)]^2 + G'(G_{n-1}(1))G_{n-1}''(1)\nonumber\\ &=& G''(1)[G_{n-1}'(1)]^2 + G'(1)G_{n-1}''(1)\nonumber\\ &=& (\sigma_1^2 - \mu_1 +\mu_1^2)\mu_{n-1}^2 + \mu_1G_{n-1}''(1). \end{eqnarray}\]
Let \(u_n = G_{n}''(1)\) and we have
\[u_n - \mu_1u_{n-1} = (\sigma_1^2 - \mu_1 +\mu_1^2)\mu_{1}^{2n-2}\]
since \(\mu_{n-1} = \mu_1^{n-1}\). This is a first-order linear inhomogeneous difference equation. We need to distinguish between two cases depending on \(\mu_1 = 1\) or not.
Case 1: \(\mu_1 \neq 1\)
The corresponding homogeneous difference equation is
\[u_n - \mu_1u_{n-1} = 0\]
which is just a geometric sequence. So the general solution is
\[u_n = B\mu_1^n\]
with \(B\) being a constant. For the particular solution, try \(u_n = C\mu_1^{2n}\) with some constant \(C\). This tentative solution satisfies the difference equation if
\[C = \frac{\sigma_1^2 - \mu_1 + \mu_1^2}{\mu_1(\mu_1 - 1)}.\]
Exercise. Verify this result. □
Combining the two solutions (the general and particular), we have
\[u_n = B\mu_1^n + \frac{\sigma_1^2 - \mu_1 + \mu_1^2}{\mu_1(\mu_1 - 1)}\mu_1^{2n}.\]
The constant \(B\) can be determined from the case with \(n=1\) where we have
\[u_1 = \sigma_1^2 -\mu_1 + \mu_1^2\]
so that
\[B = -\frac{\sigma_1^2 - \mu_1 + \mu_1^2}{\mu_1(\mu_1 - 1)}.\]
After all, we have
\[u_n = \frac{\sigma_1^2 - \mu_1 + \mu_1^2}{\mu_1(\mu_1 - 1)}\mu_1^{n}(\mu_1^n - 1).\]
I hope you still remember that \(u_n = G_n''(1)\). Therefore,
\[\sigma_n^2 = u_n + \mu_1^n - \mu_1^{2n} = \frac{\sigma_1^2\mu_1^{n-1}(\mu_1^n - 1)}{\mu_1 - 1}.\]
Case 2: \(\mu_1 = 1\)
In this case, the original difference equation becomes
\[u_n - u_{n-1} = \sigma_1^2\]
which is just an arithmetic sequence, so the general solution is
\[u_n = n\sigma_1^2 + A,\],
where \(A\) is a constant. From \(u_1 = \sigma_1^2\), we have \(A=0\). Hence,
\[\sigma_n^2 = G_n''(1) = n\sigma_1^2.\]
Comments
Post a Comment