Lecture 5

In this lecture we discuss another notion of convergence than pointwise convergence.

Banach and Hilbert spaces

Banach spaces

Normed vector spaces

Let VV be a complex vector space. A norm on VV is given by a non-negative real valued function \|\cdot\| on VV such that

  • x0\|x\|\geq 0 and x=0\|x\| = 0 if and only if x=0x= 0 (i.e. is the zero element);
  • x+yx+y\|x+y\|\leq \|x\|+ \|y\|;
  • λx=λx\|\lambda x\| = |\lambda| \|x\|.

Completeness

VV is called complete if for every sequence xnx_n such that

limm,nxnxm=0\lim_{m,n\to \infty}\|x_n - x_m\| = 0

there exists some element xVx\in V such that limnxxn=0\lim_{n\to \infty}\|x - x_n\| = 0.

A sequence xnx_n in VV is called Cauchy if it satisfies (1). By definition, it means for every ϵ>0\epsilon > 0, there exists NN (depending on ε) such that xnxm<ϵ\|x_n - x_m\| < \epsilon for every m,n>Nm,n> N. Thus

Lemma. Let xnx_n be a Cauchy sequence, then xn\|x_n\| is uniformly bounded, i.e. there exists C>0C>0 such that xnC\|x_n\|\leq C for every nn.

Proof. Choose any ϵ>0\epsilon > 0, then for some NN we have xnxmϵ\|x_n - x_m\|\leq \epsilon whenever m,nNm,n \geq N. In particular, xnxNϵ\|x_n - x_N\|\leq \epsilon for all n>Nn> N, so xnxN+ϵ\|x_n\|\leq \|x_N\| + \epsilon. Take C=max1mNxm+ϵC = \max_{1\leq m \leq N}\|x_m\|+\epsilon.

Pre-Hilbert spaces

Let VV be a vector space over C\mathbb{C}. An inner product on VV is a map (,):V×VC(,):V\times V\to \mathbb{C} satisfying

  • Conjugate symmetry: (x,y)=(y,x)(x,y) = \overline{(y,x)}.
  • (αx,y)=α(x,y)(\alpha x,y) =\alpha (x,y).
    • By the conjugate symmetry, (x,αy)=αˉ(x,y)(x,\alpha y) = \bar{\alpha}(x,y).
  • (x,x)0(x,x)\geq 0.

If VV is finite-dimensional, an inner product corresponds to a Hermitian matrix via a basis. Let A=(ei,ej)ijA=(e_i,e_j)_{ij}, then (x,y)=ytAx(x,y) = \overline{\mathbf{y}^t}A x. If VV is a real vector space, an inner product is bilinear symmetric. In the finite dimensional case, it corresponds to a positive definite matrix.

The Cauchy-Schwartz inequality

Let (,)(,) be an inner product, then (x,y)2(x,x)(y,y)|(x,y)|^2\leq (x,x)(y,y).

Proof. If y=0y=0, it trivially holds; If y0y\neq 0, let (x,y)yyy=(x,y)y(y,y)\frac{(x,y)}{\|y\|}\frac{y}{\|y\|} = \frac{(x,y)y}{(y,y)} be the projection of xx to yy direction. Then let x:=x(x,y)y(y,y)x^{\perp}:=x - \frac{(x,y)y}{(y,y)}. Now (x,x)0(x^{\perp},x^{\perp})\geq 0 gives the desired inequality.

Corollary. Let x:=(x,x)\|x\|:=\sqrt{(x,x)}, then x\|x\| is a semi-norm.

Proof. (x+y)2=x2+y2+2xyx2+y2+2(x,y)x2+y2+2Re(x,y)=(x+y,x+y)=x+y2(\|x\|+\|y\|)^2 = \|x\|^2 + \|y\|^2 + 2\|x\|\|y\|\geq \|x\|^2+\|y\|^2 +2|(x,y)| \geq \|x\|^2+\|y\|^2 +2Re(x,y) = (x+y,x+y) = \|x+y\|^2.

The inner product space VV is called a Hilbert space provided that the induced semi-norm is a norm and is complete. We’ll call an inner product space a pre-Hilbert space.

Example. The space of continuous functions on T\mathbb{T}, denoted by C0(T)\mathscr{C}^0(\mathbb{T}), equipped with the inner product

(f,g)ππf(θ)g(θ)dθ(f,g)\mapsto\int_{-\pi}^\pi f(\theta)\overline{g(\theta)}d\theta

is a pre-Hilbert space. The norm induced by the inner product is the L2L^2-norm. Its completion is the space of square Lebesgue integrable functions (see form example this note), denoted by L2(T)L^2(\mathbb{T}), which is a Hilbert space.

Orthogonality

x,yVx,y\in V is called orthogonal provided that (x,y)=0(x,y) = 0, we denote this by xyx\perp y. Then we have the analogy of Pythagoren theorem

xy2=(xy,xy)=x2+y2+0\|x - y\|^2 = (x - y,x-y) = \|x\|^2 + \|y\|^2 + 0

Lemma (Best approximation)

If ff0f - f_0 is orthogonal to VV. Then for every gVg \in V, ff0fg\|f - f_0\|\leq \|f - g\|.

Proof. Since f0gVf_0 - g\in V, ff0f0gf-f_0 \perp f_0 - g. Then by Pythagorean theorem (3)

fg=ff0+f0g=ff0+f0gff0+0\|f - g\|= \|f - f_0 + f_0 - g\| = \|f - f_0 \|+ \|f_0 - g\| \geq \|f - f_0\|+0.

Mean-Square convergence of Fourier series

Now we apply the functional analysis result to Fourier series. Let VNV_N be the complex vector space spanned by trignomic polynomials of degree NN. Then for integrable ff, fSNff - S_N f is orthogonal to VNV_N. Apply the best approximation lemma we can derive

Theorem

For fL2(T)f\in L^2(\mathbb{T}), SNffL20.\|S_N f - f\|_{L^2}\to 0.

Proof. Let ff be a continuous function on T\mathbb{T}. Let σN\sigma_N be the Cesaro sum of ff. Then σNVN\sigma_N\in V_N. By Fejer’s theorem, σNf\sigma_N\to f uniformly, this implies fσNL20\|f - \sigma_N\|_{L^2}\to 0. By best approximation lemma, fSNL2fσNL20 \|f - S_N\|_{L^2} \leq\|f-\sigma_N\|_{L^2}\to 0.

To conclude the L2L^2 case. We need the fact that continuous functions are dense in L2L^2. This means for fL2(T)f\in L^2(\mathbb{T}) there exists some hh in C0(T)\mathscr{C}^0(\mathbb{T}) such that fgL2<ϵ\|f - g\|_{L^2}<\epsilon. The mean square convergence implies that trignomic polynomials are L2L^2- dense in C0(T)\mathscr{C}^0(\mathbb{T}). Hence trignomic polynomials are dense in the L2L^2 space. It follows that for any fL2(T)f\in L^2(\mathbb{T}), there exists a trignomic polynomial gg such that fgL2ϵ\|f-g\|_{L^2}\leq \epsilon. Then gVNg\in V_N for some NN. But SNfS_N f is the orthogonal projection of ff on VNV_N so fSNffg<ϵ\|f - S_N f\|\leq \|f - g\|< \epsilon. This shows fSNf0\|f - S_N f \|\to 0.