Introductory university-level calculus, linear algebra, abstract algebra, probability, statistics, and stochastic processes.
Fourier series: Introduction
Get link
Facebook
X
Pinterest
Email
Other Apps
-
The theory of the Fourier series is based on a wild assumption: any "well-behaved" periodic function can be represented as a linear combination of sine and cosine functions, and that expression is unique for the given function. This theory (eventually) provided much of the foundations of modern mathematics. But it is also of tremendous practical importance.
Trigonometric functions are periodic. But what is a periodic function in general?
Definition (Periodic function)
A function \(f(x)\) is said to be a periodic function if there exists a real number \(T > 0\) such that
\[f(x + T) = f(x).\]
In this case, \(T\) is called a period of the function. Note that if \(T\) is a period of \(f(x)\), then \(2T\), \(3T\), \(\cdots, nT\) (\(n \in \mathbb{N}\)) is are also periods of \(f(x)\). In fact,
The smallest (non-zero) period is called the fundamental period.
Remark. If we say simply a period, we usually mean the fundamental period. □
Example. For any \(n\in \mathbb{N}\), \(\cos(nx)\) and \(\sin(nx)\) have a period of \(2\pi\), their fundamental periods are \(\frac{2\pi}{n}\) for each \(n\). □
where \(a_n\) (\(n = 0, 1, \cdots\)) and \(b_n\) (\(n=1, 2,\cdots\)) are (usually) real constants.
The factor \(\frac{1}{2}\) in the constant term (\(\frac{1}{2}a_0\)) is by convention as well as for convenience. Each term of the series has a period of \(2\pi\), so the domain of the above function of \(x\) may be \(\mathbb{R}\) or \([-\pi, \pi]\) or \([0, 2\pi]\).
Definition (Fourier expansion of a function)
Let \(f(x)\) be a function on \(\mathbb{R}\) that has a period of \(2\pi\). If
holds for all except for finitely many \(x \in \mathbb{R}\), the right-hand side is said to be
a Fourier expansion or Fourier series expansion of the function \(f\).
If term-wise integration is allowed, the coefficients \(a_n\) and \(b_n\) are readily determined. Note the following formulae: For any \(m, n \in \mathbb{N}\),
where \(\delta_{m,n}\) is Kronecker's delta. That is, the functions \(\cos x, \cos(2x), \cdots, \sin x, \sin(2x), \cdots\) are orthogonal. By multiplying (eq:FF) by \(\cos(mx)\) or \(\sin(mx)\) and then integrating, we have
Since we are assuming the period of \(2\pi\) for \(f\), the range of integration can be \([-\pi, \pi]\) instead of \([0, 2\pi]\).
The sequences \(\{a_m\}\) and \(\{b_m\}\) defined by (eq:fab) are called the Fourier coefficients of \(f\). The Fourier coefficients of \(f\) can be determined if \(f\) is integrable on \([0,2\pi]\). Given the Fourier coefficients of \(f\), we can formally define the following series, which is called the Fourier series of \(f\), denoted by \(S[f]\):
Note that, in this case, \(S[f]\) may not be the same function as \(f\). In fact, whether and/or when \(f = S[f]\) (note it's "\(=\)", not "\(\sim\)") is a fundamental question in the theory of Fourier series.
The fundamental problems in the theory of the Fourier series are
Under what conditions on \(f\) does \(S[f]\) converge?
What does the sum \(S[f]\) represent?
The continuity of \(f\) alone is known to be insufficient. To state the sufficient condition, we need the language of Lebesgue integral, or the measure theory, which is far beyond the scope of this post. We give it below anyway without proof.
Theorem (Carlson (1966))
Let \(f\) be a function that is measurable on \([0, 2\pi]\) and \(L^2\)-integrable, i.e.,
\[\int_0^{2\pi}|f(x)|^2dx < +\infty,\]
then the Fourier series \(S[f]\) converges to \(f\) almost everywhere.
Remark. The technical terms such as measurable, \(L^2\)-integrable, and almost everywhere come from the theory of Lebesgue integral. The Lebesgue integral is a generalization of the Riemann integral. Much of modern mathematics depends on the Lebesgue integral. □
Defining the birth process Consider a colony of bacteria that never dies. We study the following process known as the birth process , also known as the Yule process . The colony starts with \(n_0\) cells at time \(t = 0\). Assume that the probability that any individual cell divides in the time interval \((t, t + \delta t)\) is proportional to \(\delta t\) for small \(\delta t\). Further assume that each cell division is independent of others. Let \(\lambda\) be the birth rate. The probability of a cell division for a population of \(n\) cells during \(\delta t\) is \(\lambda n \delta t\). We assume that the probability that two or more births take place in the time interval \(\delta t\) is \(o(\delta t)\). That is, it can be ignored. Consequently, the probability that no cell divides during \(\delta t\) is \(1 - \lambda n \delta t - o(\delta t)\). Note that this process is an example of the Markov chain with states \({n_0}, {n_0 + 1}, {n_0 + 2}...
In mathematics, we must prove (almost) everything and the proofs must be done logically and rigorously. Therefore, we need some understanding of basic logic. Here, I will informally explain some rudimentary formal logic. Definitions (Proposition): A proposition is a statement that is either true or false. "True" and "false" are called the truth values, and are often denoted \(\top\) and \(\bot\). Here is an example. "Dr. Akira teaches at UBD." is a statement that is either true or false (we understand the existence of Dr. Akira and UBD), hence a proposition. The following statement is also a proposition, although we don't know if it's true or false (yet): Any even number greater than or equal to 4 is equal to a sum of two primes. See also: Goldbach's conjecture Next, we define several operations on propositions. Note that propositions combined with these operations are again propositions. (Conjunction, logical "and"): Let \(P\)...
Taylor's theorem As we have seen in a previous post, the essence of differentiation is a linear approximation. See also: Partial and total differentiation of multivariate functions Given a totally differentiable function \(z = f(x,y)\), we have in the neighbor of \(P = (a,b)\), \[f(x,y) = f(a,b) + f_x(a,b)(x - a) + f_y(a,b)(y - b) + o(\|X - P\|)\] where \(X = (x,y).\) By discarding the higher order terms \(o(\|X - P\|)\), we have the equation of the tangent plane at \(P\): \[z = f(a,b) + f_x(a,b)(x - a) + f_y(a,b)(y - b)\] that gives the linear approximation of \(z = f(x,y)\) around the point \(P\). But we may obtain better approximations by looking into the details of \(o(\|X - P\|)\), that is, by incorporating higher-order derivatives. Just as in the univariate case, we have a multivariate version of Taylor's theorem, which can be stated concisely and proved easily by using differential operators. See also: Differential operators Theorem (Taylor's theorem) Let \(f(x,y)...
Comments
Post a Comment