Student’s Book Numerical Functional Analysis Editor: Stefan Engblom Fall 2014 Preface This little book collects some of the student’s output during the course Numerical Functional Analysis, which was given for the first time in the fall 2014 at the Department of Information technology, Uppsala university. The course consisted of a total of five lectures, the first three of which went over Metric spaces, Normed spaces, and Inner product spaces, thus following closely the first three chapters of Kreyszig’s book Introductory functional analysis with applications. Two more lectures were devoted to the five ‘Big’ theorems of functional analysis as presented by the students themselves: the Hahn-Banach theorem, the Uniform boundedness theorem, the Open mapping theorem, the Closed graph theorem, and the Banach fixed point theorem. The final part of the course consisted of short essays on various topics, typically connecting Numerical analysis and Functional analysis in one way or the other. The essays were improved by double open review among the students themselves after which the final version entered this collection. A selection of student’s solution to book exercises has also been included: the selection was very loosely based on the difficulty of the exercise itself and on the solution quality. Stefan Engblom Uppsala, January 2015 Contents I Student’s essays 3 Dual-consistent approximations of PDEs and super-convergent functional output Martin Almquist 4 Lax equivalence theorem Siyang Wang 15 The Schauder fixed point theorem Hannes Frenander 20 Existence of a weak solution to the Stokes equations Hanna Holmgren 24 The Lax-Milgram theorem Andrea Alessandro Ruggiu 31 A time dependent mapping from Cartesian coordinate systems into curvilinear coordinate systems and the geometric conservation law Samira Nikkar 37 The Babuˇ ska inf-sup condition Simon Sticko 42 Uniformly best wavenumber approximations by spatial central difference operators Viktor Linders 47 The Ekeland variational principle with applications Markus Wahlsten 54 Construction of Sobolev spaces Cristina La Cognata 59 1 The Arzel´ a -Ascoli theorem Tomas Lundquist 65 A completeness theorem for non-self-adjoint eigenvalue problems Saleh RezaeiRavesh 69 The metric Lax and applications Cheng Gong 77 Fixed-point proof of Lax-Milgram theorem Fatemeh Ghasemi 82 II 87 Solutions to exercises 1 Metric spaces 88 2 Normed spaces 96 3 Inner product spaces 112 2 Part I Student’s essays 3 Dual-consistent approximations of PDEs and super-convergent functional output Martin Almquist∗ December 22, 2014 Abstract Finite-difference operators satisfying the summation-by-parts (SBP) rule can be used to obtain high-order accurate time-stable schemes for partial differential equations (PDEs), when the boundary and interface conditions are imposed weakly by the simultaneous approximation term method (SAT). The most common SBP-SAT discretizations are accurate of order 2p in the interior and order p close to the boundaries, which yields a global accuracy of p + 1 for first order PDEs. However, Berg and Nordstr¨om [2012] and Hicken and Zingg [2014] have shown that any linear functional computed from the time-dependent numerical solution will be accurate of order 2p if the spatial discretization is dual-consistent. This text discusses the concepts of dual problem and dual consistency. We walk through some of the definitions and results in Berg and Nordstr¨om [2012] and Hicken and Zingg [2014] while emphasizing the underlying functional analysis considerations. 1 Dual problems and dual consistency We introduce the concepts of dual problem and dual consistency by considering linear partial differential equations with homogeneous boundary conditions. Let L be a linear differential operator of order m on a domain Ω and consider the problem Lu − f = 0, x ∈ Ω, (1.1) ∗ Division of Scientific Computing, Department of Information Technology, Uppsala university, SE-751 05 Uppsala, Sweden. martin.almquist@it.uu.se 1 subject to homogeneous boundary conditions. f ∈ L2 (Ω) is independent of the solution u ∈ H k (Ω) ⊂ L2 (Ω). Here H k (Ω) = W k,2 (Ω) is the Sobolev space consisting of functions whose weak derivatives up to order k (for some integer k) exist and are square-integrable on Ω. We assume that m ≤ k. For v, w ∈ L2 (Ω) we use the notation Z (v, w) := vw dx. (1.2) Ω for the standard L2 inner product. We now introduce the formal adjoint of the operator L. Definition 1.1. Let u ∈ H k (Ω). The formal adjoint of a linear differential operator L is the operator L† such that (φ, Lu) = (L† φ, u) The word formal in the definition emphasizes that we are not yet specifying the domain D(L† ) of the adjoint operator. The reason for this definition is that L is not bounded on L2 (Ω), in general. Hence we need to distinguish between the formal adjoint and the Hilbert-adjoint (which concerns bounded operators) even though L2 (Ω) is a Hilbert space. Note also that the definition of L† is abstract; an explicit expression for L† φ can be obtained via (repeated) integration by parts on (φ, Lu). We now consider a bounded linear functional of the solution u, J (u) = (z, u) (1.3) for some z ∈ L2 (Ω). Note that since L2 (Ω) is a Hilbert space, the representation (1.3) includes all linear functionals, by Riesz’s representation theorem. To derive the adjoint equation of (1.1) with respect to the functional J , we take the inner product of (1.1) with a function φ ∈ D(L† ) and express J as J (u) = (z, u) = (z, u) − (φ, Lu − f ) = (φ, f ) − (L† φ − z, u). (1.4) We now obtain the adjoint or dual equation by seeking φ such that J becomes independent of u, which requires L† φ − z = 0, yielding J = (φ, f ). 2 (1.5) Remark. To simplify the analysis we have disregarded the boundary conditions. In general, the dual problem depends on the type of boundary conditions in the primal problem, but not on the particular boundary data. Boundary conditions for the dual problem can be constructed as the minimal set of homogeneous conditions which, together with the homogeneous primal boundary conditions, make all boundary terms resulting from the integration by parts procedure vanish. We outline the procedure for a model problem in Section 2.3. Definition 1.2. The continuous dual problem associated with the primal problem (1.1) and the functional J is L† φ − z = 0 subject to the dual boundary conditions. The solution φ is called the dual variable. Note that the forcing function f and functional representation z in the primal problem have switched roles in the dual problem. We now turn to the discrete case. Let Lh uh − f = 0 (1.6) be a consistent discretization of (1.1) including boundary conditions, where uh ∈ Rn may hold basis function coefficients and/or collocation values. Lh is a bounded linear operator which can be represented by matrix multiplication, Lh : Rn −→ Rn . Let (·, ·)h denote a discrete inner product on Rn × Rn such that (Rn , (·, ·)h ) is a Hilbert space. We can then use the Hilbert-adjoint: Definition 1.3. The Hilbert-adjoint of Lh is the operator L†h : Rn −→ Rn such that (φh , Lh uh )h = (L†h φh , uh )h for all φh , uh ∈ Rn . Let Jh (uh ) = (z, uh )h be an approximation of the functional J . We now derive the discrete dual problem by taking the inner product of the discretization (1.6) with φh ∈ Rn and expressing Jh as Jh (uh ) = (z, uh )h − (φh , Lh uh − fh )h = (φh , fh )h − (L†h φh − z, uh )h . (1.7) We obtain the discrete dual problem by seeking φh such that Jh is independent of uh , which requires L†h φh − z = 0 . (1.8) 3 Definition 1.4. The discrete dual problem associated with (1.6) and the functional Jh is L†h φh − z = 0 Definition 1.5. A discretization (1.6) of a continuous problem (1.1) is called dual consistent if the discrete dual problem (1.8) is a consistent discretization of the continuous dual problem (1.5). 1.1 Time-dependent problems We now extend the concepts of dual problems and dual consistency to timedependent problems. Consider the problem ut + Lu − f = 0, x ∈ Ω, t > 0 (1.9) subject to homogeneous boundary and initial conditions. Let d u dt h + Lh uh − f = 0, t > 0 (1.10) be a semi-descretization of (1.9), including the boundary conditions. Definition 1.6. The semi-discretization (1.10) is called spatially dual-consistent if Lh uh − f = 0 is a dual-consistent discretization of the steady adjoint problem, i.e., the dual problem corresponding to (1.9) with ut = 0. 2 SBP-SAT discretizations We begin this section by introducing notation and some basic concepts related to SBP-SAT discretizations, and then proceed to consider stable and dual-consistent SBP-SAT discretizations. 2.1 Definitions Consider as an example the advection equation on 0 ≤ x ≤ 1, ut + ux = f u(0, t) = g(t). 4 (2.1) We introduce the N + 1 equidistant grid points xj = jh, j = 0, 1, . . . , N, h= 1 N (2.2) and denote the semi-discrete solution vector corresponding to grid size h by uh , where uh = [uh,0 , uh,1 , . . . , uh,N ]T ∈ RN +1 and uh,j (t) approximates u(xj , t). The restriction [f (x0 ), f (x1 ), . . . , f (xN )]T of f onto the grid will also be denoted by f , with no risk for confusion. Spatial derivatives will be approximated by first-derivative SBP operators: Definition 2.1. A matrix D is called a first-derivative SBP operator if D can be written as D = H −1 Q where H = H T is positive definite, and thus defines a norm, and Q satisfies Q + QT = diag[−1, 0, . . . , 0, 1]. In this text we will restrict ourselves to diagonal norm matrices H. In that case, D consists of a 2pth-order accurate centered difference stencil in the interior and pth-order accurate one-sided stencils near the boundaries. The global accuracy for first order problems can then be shown to be p + 1 Sv¨ard and Nordstr¨om [2006]. Let u, v ∈ RN +1 . We define the discrete inner product and norm as kuk2h = (u, u)h . (u, v)h = uT Hv, (2.3) The inner product space (RN +1 , (·, ·)h ) is complete and thus a Hilbert space. Now consider a discretization matrix Lh . We can compute the Hilbert-adjoint L†h with respect to the inner product (2.3) explicitly: Lemma 2.1. Let Lh be an (N + 1) × (N + 1) matrix with real entries. Then the Hilbert adjoint is L†h = H −1 LTh H (2.4) Proof. Starting from Definition 1.3, we have (φh , Lh uh )h = (L†h φh , uh )h ∀φh , uh ∈ Rn φTh HLh uh = (L†h φh )T Huh ∀φh , uh ∈ Rn φTh HLh uh = φTh (L†h )T Huh ∀φh , uh ∈ Rn HLh = (L†h )T H (L†h )T = HLh H −1 L†h = H −1 LTh H 5 ⇐⇒ ⇐⇒ ⇐⇒ ⇐⇒ ⇐⇒ 2.2 Stable SBP-SAT discretizations The SBP-SAT semi-discretization of (2.1) can be written d uh + Duh = f + τ H −1 e0 (eT0 uh − g) dt (2.5) where e0 = [1, 0, . . . , 0]T and τ is a penalty parameter, which will be chosen such that the scheme (2.5) is stable. Multiplying (2.5) by uTh H from the left and setting f = 0, g = 0 yields d kuh k2h = (1 + 2τ )u2h,0 − u2h,N , dt (2.6) from which it is clear that τ ≤ − 21 leads to a stable scheme. Thus, the penalty parameter τ is allowed to vary in a semi-infinite range. This flexibility creates a possibility to impose additional conditions on the scheme, such as dual consistency. 2.3 Dual-consistent SBP-SAT discretizations Consider again the advection equation (2.1) on 0 ≤ x ≤ 1, ut + ux = f u(0, t) = 0. (2.7) with some initial condition and an associated linear functional Z 1 J (u) = zu dx. (2.8) 0 Note that J is a time-dependent functional. The dual problem is obtained by setting ut = 0 and finding φ such that J = (φ, f ). Using integration by parts and inserting the boundary condition in (2.7), we obtain Z 1 Z 1 J (u) = J (u) − φ(ux − f ) dx = −φ(1, t)u(1, t) + (z + φx )u dx + (φ, f ), 0 0 and hence the dual problem is −φx = z φ(1, t) = 0 . (2.9) Note that the sign has changed and the boundary condition is located at the opposite boundary, compared to the primal problem. We are now ready to state the main result in this text: 6 Theorem 2.2. Let d uh + Lh uh = f (2.10) dt be a stable and spatially dual consistent SBP-SAT approximation of the continuous problem ut + Lu = f. (2.11) A linear functional J (u) = (z, u) (2.12) Jh (uh ) = (z, uh )h (2.13) Jh (uh ) = J (u) + O(h2p ) (2.14) approximated by satisfies Proof. The proof can be found in Berg and Nordstr¨om [2012]. We now investigate whether the scheme (2.5) can be made spatially dualconsistent. Assuming homogeneous boundary conditions (g = 0) and neglecting the time-derivative, the scheme can be written as Lh uh = f (2.15) Lh = D − τ H −1 e0 eT0 . (2.16) where By definition the scheme is spatially dual-consistent if the discrete dual problem L†h φh − z = 0 (2.17) is a consistent discretization of (2.9). Using Lemma 2.1 and the SBP property, we have L†h φh −z = H −1 LTh Hφh −z = H −1 −Q + eN eTN − (τ + 1)e0 eT0 φh −z (2.18) which is a consistent discretization of (2.9) if and only if τ = −1. For τ 6= −1 the scheme imposes a boundary condition at x = 0 which does not exist in the continuous equation. Note that the requirement τ = −1 does not contradict the stability requirement τ ≤ − 21 . Thus, we expect convergence of order 2p for any linear functional of the solution if we choose τ = −1. 7 3 Numerical experiments For the interested reader, we include a set of numerical experiments which validate the theoretical results. We solve the advection equation using the scheme (2.5) and consider the functional Z 1 cos(4πx)u dx. (3.1) J (u) = 0 Using the technique of manufactured solutions, we choose the exact solution u = cos(4πx) sin(2πt), (3.2) which corresponds to the forcing function f = 2π cos(4πx) cos(2πt) − 4π sin(4πx) sin(2πt). (3.3) The analytic solution is used as initial and boundary data. With the solution (3.2), the exact time-dependent functional becomes J = 1 sin 2πt. 2 (3.4) We use an SBP-SAT discretization of internal order 6 (2p = 6) and hence expect 4th order convergence in the l2 -norm for any τ ≤ − 21 . We expect the functional error to converge as (at least) h4 for τ ≤ − 12 . For the dualconsistent choice τ = −1, we expect 6th order convergence in the functional. This is verified by Figures 3.1-3.3, which show the convergence rates for τ = −0.5, τ = −1 and τ = −1.5, respectively. The l2 -error converges as h4 regardless of τ , whereas the functional error converges as h6 for τ = −1 and as h4 otherwise. Figure 3.4 shows both the l2 -error and the functional error as functions of τ , using N = 256. As expected, the functional error is minimized by the dual-consistent choice τ = −1. It is interesting to note that also the l2 -error is minimized by τ ≈ −1 in this case, although none of the theory in this text leads us to expect that. However, the l2 -error is much less sensitive to the choice of τ than the functional error, which decreases by approximately 4 magnitudes in the vicinity of τ = −1. References J. Berg and J. Nordstr¨om. Superconvergent functional output for time-dependent problems using finite differences on summation-by-parts 8 10 0 10 -2 10 -4 10 -6 10 -8 l 2 -error 4th order reference functional error 6th order reference 10 -10 1 10 10 2 10 3 N Figure 3.1: Convergence in k·kl2 and functional for τ = −0.5; dualinconsistent discretization. 10 0 10 -5 10 -10 l 2 -error 4th order reference functional error 6th order reference 10 -15 1 10 10 2 10 3 N Figure 3.2: Convergence in k·kl2 and functional for τ = −1; dual-consistent discretization. 9 10 0 10 -2 10 -4 10 -6 10 -8 l 2 -error 4th order reference functional error 6th order reference 10 -10 1 10 10 2 10 3 N Figure 3.3: Convergence in k·kl2 and functional for τ = −1.5; dualinconsistent discretization. 10 -6 10 functional error l 2 -error 10 -4 -5 10 -6 -2 -1.5 -1 10 -8 10 -10 10 -12 -2 -0.5 τ -1.5 -1 -0.5 τ (a) (b) Figure 3.4: The (a) l2 -error ku−uh kl2 and (b) functional error, as functions of the penalty parameter τ . Only τ = −1 yields a dual-consistent discretization. 10 form. Journal of Computational Physics, 231(20):6846–6860, Aug. 2012. ISSN 00219991. doi: 10.1016/j.jcp.2012.06.032. URL http://www. sciencedirect.com/science/article/pii/S0021999112003476. J. Hicken and D. Zingg. Dual consistency and functional accuracy: a finitedifference perspective. Journal of Computational Physics, 256:161–182, Jan. 2014. ISSN 00219991. doi: 10.1016/j.jcp.2013.08.014. URL http:// www.sciencedirect.com/science/article/pii/S0021999113005524. M. Sv¨ard and J. Nordstr¨om. On the order of accuracy for difference approximations of initial-boundary value problems. J. Comput. Physics, 218: 333–352, October 2006. 11 Lax Equivalence Theorem Siyang Wang∗ December 26, 2014 Abstract Convergence of numerical methods for solving differential equations is of great importance, but difficult to prove in a direct way. The Lax equivalence theorem links consistency, stability and convergence. This enables us to prove convergence by consistency and stability analysis. In this essay, the Lax equivalence theorem is proved from the functional analysis point of view. A differential equation example is also presented. 1 Introduction The Lax equivalence theorem is originally presented in Lax and Richtmyer [1956]. It applies to an initial value problem ut = f, 0 ≤ t ≤ tf , u = u0 , t = 0. (1.1) (1.2) The theorem reads Given a properly posed initial value problem 1.1, 1.2 and a finite difference approximation C(∆t) to it that satisfies the consistency condition, stability is a necessary and sufficient condition that C(∆t) be a convergent approximation.1 Remark. The term properly posed is equivalent to wellposed, which is used more often nowadays. That is, a unique solution U exists, and a small perturbation of the data leads to a small perturbation of the solution. ∗ Division of Scientific Computing, Department of Information Technology, Uppsala University, SE-751 05 Uppsala, Sweden. siyang.wang@it.uu.se 1 This theorem is copied entirely from Lax and Richtmyer [1956]. 1 2 The theorem We prove the Lax equivalence theorem from the functional analysis point of view. We reformulate the mathematical problem to a more general one. Find x ∈ X such that T x = y, (2.1) where y ∈ Y , T : X → Y is a linear operator from a normed space X onto a normed space Y . We assume that the above problem is wellposed, that is, T −1 : Y → D(T ) is continuous. This ensures that a small perturbation in the data y leads to a small perturbation in the solution x. Since T is linear, T −1 is also linear. The spaces X and Y are equipped with appropriate norms. The mathematical problem 2.1 is solved numerically as Tn xn = yn , (2.2) where Tn : Xn → Yn , xn ∈ Xn and yn ∈ Yn . The spaces Xn and Yn are equipped with appropriate norms. Here, n can be considered as the resolution of 2.2. As n → ∞, we hope for xn → x. In practice (think of 2.1 as a differential equation), the true solution x is a continuous function, while the numerical solution xn is discrete. In order to compare xn with x to check convergence, we define the following two operators RnX : X → Xn , RnY : Y → Yn . These two operators project x and y (in X and Y ) to RnX x and RnY y (in Xn and Yn ). We make the following definition • The numerical method 2.2 is consistent with 2.1 if Tn RnX z → RnY T z, for any z ∈ D(T ). • The numerical method 2.2 is stable if kTn−1 k ≤ K, for all n, where K is a constant independent of n. • The numerical method is convergent if Tn−1 RnY y → RnX T −1 y, for every y ∈ Y . 2 The Lax equivalence theorem from the functional analysis point of view reads: If the numerical method is consistent, it is convergent if and only if it is stable. Proof. 1. consistent + stable ⇒ convergent: kTn−1 RnY y − RnX T −1 yk =kTn−1 RnY T x − RnX xk =kTn−1 (RnY T x − Tn RnX x)k ≤kTn−1 kkRnY T x − Tn RnX xk ≤K kRnY T x − Tn RnX xk | {z } →0 by consistency Tn−1 RnY y RnX T −1 y. Therefore, → The method is convergent. Note that this proof does not require deep functional analysis knowledge. 2. consistent + convergent ⇒ stable: From convergence, we know that the sequence (Tn−1 RnY y) is bounded [Kreyszig, 1978, Lemma 1.4–2], that is kTn−1 RnY yk ≤ cy , for every y ∈ Y and n = 1, 2, .... By uniform boundedness theorem [Kreyszig, 1978, Lemma 4.7–3], the sequence of the norms kTn−1 RnY k is bounded independent of n. That is, there exists a constant β such that kTn−1 RnY k ≤ β. In order to proceed, we have to make the following assumption (*): For every wn ∈ Yn , kwn k ≤ 1, there exists w ∈ Y, kwk ≤ α, such that RnY w = wn , where α is a constant independent of n. This is a plausible assumption in most cases in practice, since RnY is just a restriction operator bringing a continuous function to its corresponding discrete one. It then follows kTn−1 RnY wk kTn−1 wn k kTn−1 RnY k = sup = sup ≤β kwk kwk kwk6=0 kwk6=0 The norm of Tn−1 can be written as kTn−1 k = supkzk=1 kTn−1 zk. Choose wn such that kwn k = 1 and kTn−1 k = kTn−1 wn k. With this wn , we still have kwk ≤ α. Therefore, kTn−1 k ≤ αβ. This proves that the numerical method is stable. 3 Remark. There are interesting examples Sanz-Serna and Palencia [1985] where the assumption (*) does not hold, so that a consistent and convergent method is not stable. Also, the uniform boundedness theorem requires that Yn is complete, i.e. a Banach space. This is not assumed or guaranteed, and might be a strong assumption in practice. Rosinger even claims that the Lax equivalence theorem is wrong Rosinger [2005]. 3 Application We choose the operator T in 2.1 as a differential operator d/dt. Then we have a differential equation of the type xt = y, 0 ≤ t ≤ tf , (3.1) with the initial condition x(0) = c. Comparing with the mathematical problem 2.1, the space X and Y can be chosen as {x ∈ C 2 [0, tf ] | x(0) = c} equipped with the maximum norm in [0, tf ]. We introduce the time discretization tj = jk, j = 0, 1, · · · , N, where the step size k = tf /N . The corresponding discrete solutions are x0 , x1 , · · · , xN . We equip the discrete space with the maximum norm. The finite difference scheme can be written as xn+1 = P (k)xn , n = 0, 1, · · · , N − 1, where P (k) is the finite difference operator taking the discrete solution from the current time step to the next time step. The recursion relation can be written as x1 P (k)x0 1 −P (k) x2 0 1 x3 0 −P (k) 1 . = .. .. .. .. . . . . xN 0 −P (k) 1 {z } | Q(k) Stability requires that kQ−1 (k)k ≤ cQ , where cQ is independent of k. Due to 4 the special structure of Q(k), its inverse can be computed exactly, yielding 1 P (k) 1 2 1 Q−1 (k) = P (k) P (k) . .. . . . . . . . . . . N −1 2 P (k) · · · P (k) P (k) 1 Therefore, kQ−1 (k)k = |P (k)| < 1. PN −1 i=0 |P i (k)| = 1−|P (k)|N . 1−|P (k)| Stability is ensured if Remark. By taking y = −λx, we get the familiar test equation, which is often used to analyze the stability properties of numerical methods for time integration. For forward Euler method, P (k) = 1 − kλ, and stability requires |1 − kλ| < 1. This is the same result as if we would do the standard stability analysis. References E. Kreyszig. Introductory functional analysis with applications. John Wiley & Sons, 1978. P. D. Lax and R. D. Richtmyer. Survey of the stability of linear finite difference equations. Comm. Pure Appl. Math., IX:267–293, 1956. E. E. Rosinger. What is wrong with the Lax-Richtmyer fundamental theorem of linear numerical analysis? Technical report, 2005. J. M. Sanz-Serna and C. Palencia. A general equivalence theorem in the theory of discretization methods. Math. Comp., 45(171):143–152, 1985. 5 The Schauder fixed point theorem Hannes Frenander∗ January 5, 2015 Abstract The Brouwer and Schauder fixed point theorems are presented; the latter with a complete proof. An application of the Schauder fixed point theorem regarding solutions to partial differential equations has also been investigated. 1 Brouwer fixed point theorem For completeness, we start by introducing the Brouwer fixed point theorem (BFPT) since it is essential for understanding the proof of our main theorem: the Schauder fixed point theorem (SFPT). However, we neglect the proof since it would obscure the main points of this essay. The theorem is typically used in numerical analysis to prove existence of initial boundary value problems, which will be demonstrated in the end of this essay. In general, the BFPT states that every continous mapping that maps a convex set into itself has at least one fixed point. That is: Theorem 1.1 (Brouwer fixed point theorem). Let X be a compact and convex set and f : X → X a continuous map. Then there is a x0 ∈ X s.t f (x0 ) = x0 . To illustrate, lets take a simple example by considering the domain X = [0, 1] ⊂ R, i.e. the set of all real numbers between zero and one. This set is compact and convex. BFPT now states that every mapping f (x) that remains in this domain has at least one fixed point. For example, f (x) = x2 has a fixed point at x0 = 0, 1, f (x) = sin(x) has a fixed at x0 = 0 and so on. ∗ Division of Computational Mathematics, Department of Mathematics, Link¨oping university, SE-58183 Link¨oping, Sweden. hannes.frenander@liu.se 1 2 Shauder fixed point theorem The SFPT was proved 1930 by Juliusz Schauder, but only for special cases. It remained for Robert Cauty to prove it in its final form in 2001. The theorem is a generalization of BFPT into vector spaces and claims that every continuous mapping from a convex vector space into itself has at least one fixed point. That is: Theorem 2.1 (Shauder fixed point theorem). Let X be a normed vector space and K ⊂ X a convex and compact subset of X. Then every continuous map f : K → K has a fixed point. Proof. Since K is compact, it has a finite subcover. That is: there is a finite sequence xi such that the open balls Bϵ (xi ) cover the set K, where ϵ > 0 is the radius of the balls. Lets define the continuous functions gi (x) as { ϵ − ||x − xi || if ||x − xi || ≤ ϵ gi (x) = 0 if ||x − xi || ≥ ϵ. ∑ Since gi (x) ≥ 0 and i gi (x) > 0 for all x ∈ K, we can define another continuous function in K: ∑ gi (x)xi g(x) = ∑i . i gi (x) 1 We can easily∑verify that g is a map from ∑ K into the convex hull K0 of K. That is, g = i αi xi where αi ≥ 0 and i αi = 1. One can also observe that ||g(x) − x|| ≤ ϵ for all x ∈ K. The continuous map H = g ◦ f maps K0 into itself. Since K0 is compact, the BFPT tells us that there is a z ∈ K0 such that H(z) = g(f (z)) = z and therefore ||f (z) − z|| = ||f (z) − g(f (z))|| ≤ ϵ. Since ϵ can be arbitrarily small, there is for all m ∈ N a zm ∈ K such that ||f (zm ) − zm || ≤ 1 1 . m The ∑ convex hull K0∑of K is the smallest convex set that contains K. In this case, the set { i αi xi |αi > 0, i αi = 1} is a convex hull of K. 2 Remember that the theorem demands K to be compact, so there is a sub sequence (zm )k such that f ((zm )k ) → x0 for some x0 ∈ K. Hence, we have ||(zm )k − x0 || = ||(zm )k − f ((zm )k ) + f ((zm )k ) − x0 || ≤ ||(zm )k − f ((zm )k )|| + ||f ((zm )k ) − x0 || 1 ≤ + ||f ((zm )k ) − x0 || → 0 mk which means (zm )k → x0 . Since f is continuous, we also have f ((zm )k ) → f (x0 ). This means f (x0 ) = x0 , i.e. x0 ∈ K is a fixed point. 3 An application The Schauder fixed point theorem is commonly used to prove existence of solutions to partial differential equations (Pouso [2012],Cao et al. [2014],Agarwal et al. [2013]), and we will in this section consider an example given by Pouso [2012]. For clarity, we skip some details of the proof and the reader is referred to Pouso [2012] for the full version. Consider the problem xtt = f (t, x) x(0, t) = x(1, t) = 0 t ∈ I = [0, 1] (3.1) under the Carath´eodory’s conditions Pouso [2012]. Hence, in (3.1), the function f (t, x) is allowed to be discontinuous. We now prove that (3.1) has a solution. First, we demand that there is a M , such that |f (t, x)| ≤ M (t), and consider the domain ∫ t 1 K = {x ∈ C (I), x(0) = x(1) = 0, |xt (t) − xt (s)| ≤ M (r)dr} s which can be proved to be compact and convex subset of the Banach space C 1 (I) if we define the norm ||x|| = maxt∈I |x(t)| + maxt∈I |xt (t)|. Consider now the operator T , defined by ∫ t T x(t) = G(t, s)f (s, x(s))ds 0 3 where G(t, s) is the Green’s function to (3.1). T maps K into itself if all the above conditions are fulfilled. Therefore, according the SFPT, T has a fixed point: T x(t) = x(t) which is then the solution to (3.1). References R. Agarwal, S. Arshad, and D. O’regan. A Schauder fixed point theorem in semilinear spaces and applications. Fixed Point Theory and Applications, 306, 2013. Z. Cao, C. Yuan, and X. Li. Applications of Schauders fixed point theorem to semipositone singular differential equations. Journal of Applied Mathematics, 2014, 2014. H. Pouso. Schauder’s fixed-point theorem: new applications and a new version for discontinous operators. Boundary Value Problems, 92, 2012. 4 Existence of a weak solution to the Stokes equations Application of the Babuška-Brezzi inf-sup theorem to the Stokes equations Hanna Holmgren∗ January 5, 2015 Abstract The dynamics of incompressible viscous fluids are described by the Stokes equations if stationary laminar flow conditions with low Reynolds numbers are assumed, see Figure 0.1 for an example. The existence of a unique solution to the weak formulation of the Stokes equations is proved in this essay. First the Two Hilbert Spaces H10 (Ω) and L20 (Ω) are introduced and the weak formulation of the Stokes equations is derived. Then, the BabuškaBrezzi inf-sup theorem is applied to the corresponding variational problem to show existence and uniqueness of a weak solution. Figure 0.1: Laminar flow conditions on the River Derwent in North Yorkshire. 1 Division of Scientific Computing, Department of Information Technology, Uppsala university, SE-751 05 Uppsala, Sweden. hanna.holmgren@it.uu.se ∗ 1 1 The Stokes Equations The dynamics of incompressible fluids are generally described by the Navier-Stokes equations. These equations are derived by considering the conservation of mass, momentum and energy of a fluid flow. If the flow is “slow and calm” it is called a laminar flow and an example is illustrated in Figure 0.1. For steady laminar flows which have a small Reynolds number (intertial forces are small compared with viscous forces) the Navier-Stokes equations are reduced to the Stokes equations. In the Stokes equations all time dependent terms as well as the term describing advection have been neglected. This essay aims at proving the existence of a unique solution to the weak formulation of the Stokes equations. Assume an incompressible viscous fluid with kinematic viscosity ν > 0 is filling up a domain Ω ⊂ R3 with boundary Γ, then the Stokes system of equations is given by −ν∆u + ∇p = f, in Ω ∇ · u = 0, in Ω u = 0, on Γ (1.1) (1.2) (1.3) where f is a given volume force, u is the unknown vector describing the velocity of the fluid and p is the unknown pressure. The Two Hilbert Spaces H10(Ω) and L20(Ω) 2 Before deriving the weak formulation of the Stokes system (1.1)−(1.3) we first need to introduce two Hilbert spaces related to the two unknowns u and p. We assume the velocity u to be in the Hilbert space H10 (Ω) (here Ω ⊂ R3 ), which is the space of all vector fields v = (v1 , v2 , v3 ) with components vi in the space H01 defined by H01 (Ω) = {v ∈ H 1 (Ω) : v|Γ = 0}. (2.1) Remember that the space H 1 (Ω) is equipped with the norm defined by kvk2H 1 (Ω) = kvk2L2 (Ω) + k∇vk2L2 (Ω) . The Hilbert space(s) H10 (Ω) is equipped with the inner product and seminorm defined by (u,v) H10 (Ω) = Z Ω ∇u : ∇v dx |v|2H10 (Ω) = (v,v)H10 (Ω) 1 Picture from http://evidence.environment-agency.gov.uk 2 (2.2) (2.3) where Z Ω ∇u : ∇v dx = Z Ω ∇u1 · ∇v1 dx + Z Ω ∇u2 · ∇v2 dx + Z Ω ∇u3 · ∇v3 dx. Further, we assume the pressure p to be in the Hilbert space L20 (Ω) defined by L20 (Ω) 2 = {q ∈ L (Ω) : Z Ω q dx = 0} (2.4) and equipped with the usual inner product and norm of the space L2 (Ω). 3 The Weak Form of Stokes Equations Now, the weak formulation is derived by first multiplying the momentum equation (1.1) by a test vector v ∈ H10 (Ω) and then integrating by parts which gives the following weak formulation of the momentum equation ν Z Ω ∇u : ∇v dx − Z Ω (∇ · v)p dx = Z Ω f · v dx, where all boundary terms from the integration by parts vanishes since v|Γ = 0. Further, multiplying the incompressibility constraint (1.2) by a test function q ∈ L20 (Ω) and integrating yields Z Ω (∇ · u)q dx = 0. Summarizing, the variational formulation derived from the Stokes system of equations reads: Find the solution pair (u,p) ∈ H10 (Ω) × L20 (Ω) such that a(u,v) + b(v,p) = l(v) ∀v ∈ H10 (Ω) b(u,q) = 0 ∀q ∈ L20 (Ω), (3.1) (3.2) where the following linear forms have been introduced: a(u,v) = ν Z b(u,q) = − l(v) = Z ZΩ Ω ∇u : ∇v dx, (3.3) (∇ · u)q dx, (3.4) Ω f · v dx. 3 (3.5) 4 The Babuška-Brezzi inf-sup theorem For the abstract variational problem: Find u ∈ V , such that a(u,v) = l(v), ∀v ∈ V, where V is a Hilbert space with inner product (·,·), a(u,v) is a coercive continuous bilinear form on V and l(v) is a continuous linear form on V , the Lax-Milgram lemma can be used to prove the existence and uniqueness of a solution [1] . Remember that a bilinear form a(·,·) is said to be V -coercive if there is a positive constant m such that mkvk2V ≤ a(v,v) ∀v ∈ V. However, in the case of the weak formulation of the Stokes equations (3.1)-(3.2), the existence of a unique solution does not follow from the Lax-Milgram lemma alone, since it is impossible to establish coercivity for the bilinear form b(·, ·) in (3.4) [2]. In this case we need the Babuška-Brezzi inf-sup theorem [1]: Theorem 1: The Babuška-Brezzi inf-sup theorem. Let V and M be two Hilbert spaces, and let a(·,·) : V × V −→ R and b(·,·) : V × M −→ R be two continuous bilinear forms with the following properties: There exists a positive constant α such that a(v,v) ≥ αkvk2V for all v ∈ U0 = {v ∈ V : b(v,q) = 0 ∀q ∈ M }, i.e. a(·,·) is U0 -coercive, and there exists a positive constant β such that |b(v,q)| ≥ β, q∈M,q6=0 v∈V,v6=0 kvkV kqkM inf sup where the latter is called the Babuška-Brezzi inf-sup condition. Finally, let l : V −→ R and g : M −→ R be two continuous linear forms. Then the variational problem: Find (u,p) ∈ V × M such that a(u,v) + b(v,p) = l(v) b(u,q) = g(q) has one and only one solution. 4 ∀v ∈ V, ∀q ∈ M, 5 Existence and uniqueness of a weak solution We use the Babuška-Brezzi inf-sup theorem to prove existence and uniqueness of a solution to the variational problem (3.1)-(3.2) derived from the Stokes equations. Let V = H10 (Ω) and M = L20 (Ω), where V and M are the two spaces from the Babuška-Brezzi inf-sup theorem. First, we prove that a(·,·) is U0 -coercive: 1. a(·,·) is H10 (Ω)-coercive since a(v,v) = ν|v|2H10 (Ω) ∀v ∈ H10 (Ω). 2. Then a(·,·) is also coercive on U0 = {v ∈ H10 (Ω) : b(v,q) = 0 ∀q ∈ L20 (Ω)} since the above equality holds for all v ∈ H10 (Ω). Then, for proving the second part of the Babuška-Brezzi inf-sup theorem (i.e. that b(·,·) fulfills the Babuška-Brezzi inf-sup condition) we need the following theorem [1]: Theorem 2. Let Ω be a domain in RN . Then the injective continuous linear operator div : (ker div)⊥ −→ L20 (Ω) is surjective and has a continuous inverse. Here “div” is another notation used for the divergence, i.e. divv = ∇ · v for any v. The space (ker div) is defined by (ker div) = {v ∈ H10 (Ω) : divv = 0 in L20 (Ω)}. Now, we prove that b(·,·) fulfills the Babuška-Brezzi inf-sup condition: 1. Since div : (ker div)⊥ −→ L20 (Ω) is surjective (by Theorem 2), there exists a unique vector field w ∈(ker div)⊥ ⊂ H10 (Ω) such that divw = q in L20 (Ω), and besides, since div : (ker div)⊥ −→ L20 (Ω) has a continuous inverse (also by Theorem 2), there exists a constant C such that |w|H10 (Ω) ≤ CkqkL20 (Ω) 5 ∀q ∈ L20 (Ω). 2. Now we can prove the Babuška-Brezzi inf-sup condition holds for b(·,·), for each non-zero q ∈ L20 (Ω): R |b(v,q)| | Ω (∇ · v)q dx| sup = sup v∈H10 (Ω),v6=0 |v|H10 (Ω) kqkL20 (Ω) v∈H10 (Ω),v6=0 |v|H10 (Ω) kqkL20 (Ω) h ∇ · v = divv i R | Ω (divv)q dx| = sup v∈H10 (Ω),v6=0 |v|H10 (Ω) kqkL20 (Ω) h i since w ∈ (ker div)⊥ ⊂ H10 (Ω) we can choose v = w R | (divw)q dx| ≥ Ω |w|H10 (Ω) kqkL20 (Ω) h divw = q i R q 2 dx| = |w|H10 (Ω) kqkL20 (Ω) | Ω hR = q 2 dx = kqk2L2 (Ω) 0 i kqkL20 (Ω) |w|H10 (Ω) h ≥ Ω |w|H10 (Ω) ≤ CkqkL20 (Ω) i 1 =β C So, by the Babuška-Brezzi inf-sup theorem, there exists a unique solution pair (u,p) ∈ H10 (Ω) × L20 (Ω) to the variational problem (3.1)-(3.2) derived from the Stokes equations. 6 References [1] Philippe G. Ciarlet, Linear and Nonlinear Functional Analysis with Applications. Society for Industrial and Applied Mathematics, Philadelphia, 2013. [2] Mats G. Larson, Fredrik Bengtzon, The Finite Element Method - Theory, Implementation and Applications. Springer, Umeå, Sweden, 2012. 7 The Lax–Milgram Theorem Andrea Alessandro Ruggiu∗ January 7, 2015 Abstract In this essay we state and prove the Lax–Milgram Theorem, a result which claims well–posedness of variational problems. This theorem is here proven through the classical approach, that is to say in this work the Banach Fixed Point Theorem is not taken into account. 1 Introduction Various problems in science can be formulated through PDEs, which necessarly don’t have a classic solution. Solving a problem weakly means find a solution for the problem, called weak solution, whose derivatives could also not exist in a classical sense. Not rarely they are the only solutions that is possible to find. In particular, it’s possible to formulate any differential problem as a variational problem of the following kind given F ∈ H ∗ , with H ∗ dual space of H, find u ∈ H such that a (u, v) = F (v) , ∀v ∈ H. (1) A common example is the Poisson problem −∆u = f, x ∈ Ω u = 0, x ∈ ∂Ω (2) that can be formulated as a variational problem through the integration by parts. ∗ Computational Mathematics, Department of Mathematics, Link¨ oping university, SE581 83 Link¨oping, Sweden. andrea.ruggiu@liu.se 1 Indeed, multiplying the equation by v ∈ H01 (Ω) and integrating by parts leads to Z Z (3) ∇u · ∇v dx = f v dx, ∀v ∈ H01 (Ω) . Ω Ω The solution of (3), u ∈ 2 H01 (Ω), is said weak solution for (2). Background theorems In this section we recall two important theorems that we want to use in order to prove the Lax–Milgram Theorem. Both of them were studied and analysed in the Numerical Functional Analysis course. The first one is the Hilbert Projection Theorem. Theorem 2.1. Let us consider a Hilbert space H, a non–empty subspace V ⊂ H closed in H and x ∈ H. Then ∃!v = PV (x) ∈ V such that kx − vk = inf v∈V kx − vk = d ≥ 0. Moreover • PV (x) = x ⇐⇒ x∈V, • QV (x) = x − PV (x) ∈ V ⊥ . The second theorem that we want to use is the Riesz Representation Theorem Theorem 2.2. Let H be a Hilbert space. Then ∀L ∈ H ∗ ∃!u ∈ H s.t. L (v) = (u, v) ∀v ∈ H and moreover kukH = kLkH ∗ . 3 Lax–Milgram Theorem In this section we prove the following statement Theorem 3.1. Let H be a Hilbert space. If a (·, ·) is a bilinear form such that • ∃M > 0 such that |a (u, v)| ≤ M kuk kvk, • ∃α > 0 such that a (u, u) ≥ α kuk2 , 2 and F ∈ H ∗ is linear and bounded, then ∃!u solution for (1) and kuk ≤ 1 kF k∗ , α (4) where k·k∗ is the norm associated to H ∗ . Proof. By Theorem 2.2 ∃!z ∈ H s.t. F (v) = (z, v) , ∀v ∈ H and kF k∗ = kzk. But a (u, ·) : V 7→ a (u, v) is a linear operator in H ∗ , then, again by Theorem 2.2 we know that ∃!h ∈ H s.t. a (u, v) = (h, v) , ∀v ∈ H with u fixed. Thus we can call h = A (u). Notice that if u is fixed, then h exists and it is unique. Therefore (1) can be rewritten as (A (u) , v) = (z, v) , ∀v ∈ H, which implies A (u) = z, A : H → H. In order to show the uniqueness of u we have to prove that A is invertible. This proof requires 5 steps • A is linear. As stated before, by Theorem 2.2 (A (u) , v) = a (u, v) , ∀v ∈ H. Then we can use the bilinearity of a (·, ·) in order to prove the linearity of A (A (λu1 + µu2 ) , v) = a (λu1 + µu2 , v) = λa (u1 , v) + µa (u2 , v) = = λ (A (u1 ) , v) + µ (A (u2 ) , v) = = (λA (u1 ) + µA (u2 ) , v) , ∀v ∈ H. This implies that A (λu1 + µu2 ) = λA (u1 ) + µA (u2 ), that is to say A is linear. For this reason from now on we write Au instead of A (u). • A is bounded. We can write kAuk2 = (Au, Au) = a (u, Au) ≤ M kuk kAuk . (5) If kAuk = 0, then the boundedness follows trivially. Therefore we assume kAuk = 6 0: dividing by kAuk both the left and the right–hand side of (5), the boundedness of A follows. 3 • A is injective. This statement is equivalent to nul (A) = {0}, where nul is the null space of A. Using the hypotheses and Cauchy–Schwarz inequality in a (u, u) = (Au, u) leads to α kuk2 ≤ a (u, u) = (Au, u) ≤ kAuk kuk . This implies that if u 6= 0 (trivially an element of nul (A)) then kAuk ≥ α kuk, that is to say kuk ≤ 1 kAuk , α ∀u ∈ H. (6) Therefore if we consider u ∈ nul (A) (kAuk = 0), then u = 0. This means that the only element in the nullspace of A is 0, then A is injective. • The image of A is closed. Let yn ∈ Im (A) an arbitrary convergent sequence in H, let’s say yn → y. The statement is proven if, and only if, y ∈ Im (A). But since yn ∈ Im (A) ⇐⇒ ∃xn s.t. yn = Axn we can also prove the statement by showing the existence of x s.t. y = Ax. As a first step we can prove that xn is convergent. Using (6) and the linearity of A we can write 0 ≤ kxn − xm k ≤ 1 1 1 kA (xn − xm )k = kAxn − Axm k = kyn − ym k . α α α But since yn is convergent in H, then it is also Cauchy. In particular, this implies by the above equation that xn is Cauchy in H. The completeness of H leads to the convergence of xn to a certain x ∈ H. By hypothesis it is known that Axn = yn → y. (7) But, on the other hand, we know by the continuity of A (it is linear and bounded) that Axn → Ax. (8) By (7), (8) and the uniqueness of the limit we have y = Ax, which prove the closedness of Im (A) since x is such that y = Ax ∈ Im (A). 4 • A is surjective, that is to say Im (A) = H. We want to prove this statement by contradiction: let us assume that Im (A) ( H. Then ∃q0 ∈ / Im (A) and applying Theorem 2.1 (notice that Im (A) is closed by the previous point) leads to QIm(A) (q0 ) = q0 − PIm(A) (q0 ) ∈ Im (A)⊥ . If we define q0 − PIm(A) (q0 ) q= q0 − PIm(A) (q0 ) , then q ∈ Im (A)⊥ and kqk = 1. Then we can write 0 = (Aq, q) = a (q, q) ≥ α kqk2 = α > 0. Since we have a contradiction, then is impossible to have Im (A) ( H and therefore A is surjective. Finally, we have proven that A is invertible by proving that A is injective and surjective. In particular Au = z has the unique solution u = A−1 z and this means that there exists a unique solution for (1). In order to prove the remaining part of the statement we use (6) and Theorem 2.2 1 1 1 kuk ≤ kAuk = kzk = kF k∗ . α α α 4 Example of application Let’s take into account the problem (2) and its weak formulation given by (3). We can write Z Z a (u, v) ∇u · ∇v dx, F (v) = f v dx. (9) Ω Ω With this notation we have written (3) as a problem of the same kind of (1) with H ≡ H01 (Ω). It can be shown that a suitable norm for H01 (Ω) is given by k·kH 1 (Ω) = k∇·kL2 (Ω) 0 and, equivalently, we can use as inner product in H01 (Ω) (u, v)H 1 (Ω) = (∇u, ∇v)L2 (Ω) . 0 5 We can easily check that |a (u, v)| ≤ k∇ukL2 (Ω) k∇vkL2 (Ω) = kukH 1 (Ω) kvkH 1 (Ω) 0 and 0 a (u, u) = k∇uk2L2 (Ω) = kuk2H 1 (Ω) . 0 Moreover F is linear and it is bounded, indeed Z f v dx ≤ kf kL2 (Ω) kvkL2 (Ω) ≤ cp kf kL2 (Ω) kvkH 1 (Ω) , F (v) = 0 Ω where we have used the Cauchy–Schwarz and the Poincar´ e inequalities with constant cp . Then the Lax–Milgram theorem holds with α = M = 1 and therefore (2) admits a unique solution u ∈ H01 (Ω) that is such that kuk ≤ cp kf kL2 (Ω) . References [1] Evans L.C., Partial Differential Equations, American Mathematical Society, PP.297-299, 1997. [2] Salsa S., Equazioni a derivate parziali. Metodi, modelli e applicazioni, Springer, Chapter 6, 2nd edition, 2010. 6 A time dependent mapping from Cartesian coordinate systems into curvilinear coordinate systems and the geometric conservation law Samira Nikkar∗ January 9, 2015 Abstract A time-dependent transformation that maps a moving domain in the physical space (in terms of the Cartesian coordinates) into a fixed domain in the computational space (in terms of the curvilinear coordinates) is considered. The continuous transformation is usually non-singular, which means that the Geometric Conservation Law (GCL) holds exactly. In the numerical set up, the differential operators and the metric formulations must satisfy some certain criteria in order for the Numerical GCL (NGCL) to be fulfilled. With the use of the Summation-by-Parts (SBP) operators for the space and time discretization, we can prove that even for the discrete mapping, the NGCL is also guaranteed. 1 The continuous problem Consider a time-dependent transformation from the Cartesian coordinates into curvilinear coordinates, which results in a fixed spatial domain, as x = x(τ, ξ, η), y = y(τ, ξ, η), t = τ, ξ = ξ(t, x, y), η = η(t, x, y), τ = t. (1.1) such that 0 ≤ ξ ≤ 1, 0 ≤ η ≤ 1, 0 ≤ τ ≤ T , see Figure 1. The Jacobian ∗ Division of scientific computing, Department of Mathematics, Linkoping university, Sweden. samira.nikkar@liu.se 1 y η c b c' d' d Ω n a x ab bc cd da a' b' ξ a'b' : South (s) b'c' : East (e) c'd' : North (n) d'a' : West (w) Figure 1.1: A schematic of the moving and fixed domains and boundary definitions. matrix of the transformation is xξ yξ 0 [J] = xη yη 0 , xτ y τ 1 (1.2) Jξt = xη yτ − xτ yη , Jξx = yη , Jξy = −xη Jηt = yξ xτ − xξ yτ , Jηx = −yξ , Jηy = xξ , (1.3) The relation between [J], and its inverse, which transforms the derivatives back to the Cartesian coordinates leads to the metric relations in which J = xξ yη −xη yξ > 0 is the determinant of [J]. If we want to preserve any conservation property while transforming the physical domain into the computational domain, the transformation and the metric relations must satisfy the GCL Farhat et al. (2001); Sj¨ogreen et al. (2014), summarized as the following Jτ + (Jξt )ξ + (Jηt )η = 0, (Jξx )ξ + (Jηx )η = 0, (Jξy )ξ + (Jηy )η = 0. (1.4) In practice (1.4) holds for any non-singular transformation. Now let us investigate the requirements for the discrete transformation such that the numerical version of (4) also holds. 2 2 The discrete problem The spatial computational domain Ω is a square in ξ, η coordinates, see Figure 1.1, and discretized using N and M nodes in the direction of ξ and η respectively. In time we use L time levels from 0 to T. The fully-discrete grid vector is a column vector of size LM N organized as follows Z0 .. . Z= [Zk ] .. . ZL Z0 Z0 .. .. . . ; [Zk ] = [Zi ] ; [Zi ] = [Zj ] , k .. .. . . ZM ki ZN k (2.1) where Zkij = Z(τk , ξi , ηj ) and Z ∈ {x, y}. The first derivative xξ is approximated by Dξ x, where Dξ is a so-called SBP operator of the form Dξ = Pξ−1 Qξ , (2.2) and x = [x0 , x1 , · · · , xN ]T is the x coordinate of the grid points. Pξ is a symmetric positive definite matrix, and Q is an almost skew-symmetric matrix that satisfies Qξ + QTξ = E1 −E0 = B = diag(−1, 0, ..., 0, 1). (2.3) In (2.3), E0 = diag(1, 0, ..., 0) and E1 = diag(0, ..., 0, 1). The η and τ directions are discretized in the same way. A finite difference approximation including the time discretization Nordstr¨om and Lundquist (2013), on SBP-SAT form, is constructed by extending the one-dimensional SBP operators in a tensor product fashion as Dτ = Pτ−1 Qτ ⊗ Iξ ⊗ Iη , Dξ = Iτ ⊗ Pξ−1 Qξ ⊗ Iη , Dη = Iτ ⊗ Iξ ⊗ Pη−1 Qη (2.4) where ⊗ represents the Kronecker product Loan (2000). Note that in (2.4) we have used the same names for the differential operators in multi dimensional cases as the operators in the one dimensional case. All matrices in the first position are of size L×L, the second position N×N , the third position M×M . I denotes the identity matrix with a size consistent with its position in the Kronecker product. 3 The Kronecker product is bilinear and associative. For square matrices the following rules exist Nordstr¨om and Berg (2013), (A⊗B)(C ⊗D) = (AC ⊗BD), (A⊗B)−1 = A−1 ⊗B −1 , (A⊗B)T = AT ⊗B T . (2.5) For later reference we need Lemma 2.1. The difference operators in (2.4) commute. Proof. The properties (2.5) of the Kronecker product lead to Dτ Dξ = (Pτ−1 Qτ ⊗ Iξ ⊗ Iη )(Iτ ⊗ Pξ−1 Qξ ⊗ Iη ) = Pτ−1 Qτ ⊗ Pξ−1 Qξ ⊗ Iη = (Iτ ⊗ Pξ−1 Qξ ⊗ Iη )(Pτ−1 Qτ ⊗ Iξ ⊗ Iη ) = Dξ Dτ . The proof is analogous for the other coordinate combinations. We show Lemma 2.2. The Numerical Geometric Conservation Law (NGCL) holds: Jτ + (Jξt )ξ + (Jηt )η = 0, (Jξx )ξ + (Jηx )η = 0, (Jξy )ξ + (Jηy )η = 0. (2.6) Proof. Consider the following definitions, Jτ (Jξ t )ξ (Jη t )η (Jξx )ξ (Jξy )ξ (Jηx )η (Jηy )η = = = = = = = diag[Dτ (Dη M (1) − Dξ M (2) )] diag[Dξ (Dτ M (2) − Dη M (3) )] diag[Dη (Dξ M (3) − Dτ M (1) )] diag[Dξ (Dη y)] diag[−Dξ (Dη x)] diag[−Dη (Dξ y)] diag[Dη (Dξ x)] (2.7) in which x and y are the discrete Cartesian coordinates defined in (2.1). Also M (1) = diag(y)(Dξ x), M (2) = diag(y)(Dη x) and M (3) = diag(y)(Dτ x). By Lemma 2.1 we find that the NGCL holds exactly. 3 Summary and conclusions We have considered a time-dependent curvilinear coordinates. By using SBP operators in space and time and a special definitions for the metric coefficients the Geometric Conservation Law is proven to hold numerically. 4 References C. Farhat, P. Geuzaine, C. Grandmont, The discrete geometric conservation law and the nonlinear stability of ALE schemes for the solution of flow problems on moving grids, Journal of Computational Physics 174 (2001) 669–694. B. Sj¨ogreen, H. C. Yee, M. Vinokur, On high order finite-difference metric discretizations satisfying GCL on moving and deforming grids, Journal of Computational Physics 265 (2014) 211–220. J. Nordstr¨om, T. Lundquist, Summation-by-parts in time, Journal of Computational Physics 251 (2013) 487–499. C. F. V. Loan, The ubiquitous Kronecker product, Journal of Computational and Applied Mathematics 123 (2000) 85–100. J. Nordstr¨om, J. Berg, Conjugate heat transfer for the unsteady compressible Navier-Stokes equations using a multi-block coupling, Computers & Fluids 72 (2013) 20–29. 5 The Babuˇska Inf-Sup Condition Simon Sticko∗ January 12, 2015 Abstract A general weak formulation, typically arising from a linear timeindependent partial differential equation, is discussed. The Babuˇska inf-sup condition is shown to guarantee that the weak formulation is a well-posed problem. 1 Introduction The finite element method applied to a linear time-independent partial differential equation typically leads to a weak formulation of the following form: find u ∈ U such that a(u, v) = l(v) ∀v ∈ V. (1.1) Here U and V are Hilbert spaces, a(·, ·) : U × V → R a bounded bilinear form: |a(u, v)| ≤ αC kukU kvkV , αC ∈ (0, ∞) and l(v) : V → R is a bounded linear functional. Our goal is to find a condition for when (1.1) is a well-posed problem. That is, the problem should fulfill the following requirements: (E) There exist a solution u. (U) The solution is unique. (C) The solution depends continuously on initial data. ∗ Division of Scientific Computing, Department of Information Technology, Uppsala university, SE-751 05 Uppsala, Sweden. simon.sticko@it.uu.se 1 2 Theory In order to understand if the problem in (1.1) is well posed or not we would like to reformulate it into something which is a bit more familiar. In the following we denote the scalar product in V and U by h·, ·iV and h·, ·iU respectively. Since U and V are Hilbert spaces a(·, ·) has a Ritz-representation A : U → V such that: a(u, v) = hAu, viV , where A is a bounded linear operator with the same norm as a(·, ·). Also l(·) has a Ritz-representations f ∈ V : l(v) = hf, viV . Using this we can now express (1.1) as hAu − f, viV = 0 ∀v ∈ V. Since this should hold for all v ∈ V it holds in particular for v = Au − f . So this leads to the equation Au = f. (2.1) This formulation is equivalent to (1.1), but it’s easier to understand what is required for the problem to be well posed. In order for a solution of (2.1) to exist for an arbitrary f ∈ V we must require that there is at least one u ∈ U such that Au = f . Thus A must be a surjective mapping. If we want uniqueness we can not have two different elements u1 , u2 ∈ U such that Au1 = f and Au2 = f . Thus we must also require A to be injective. Finally, consider two different right-hand sides f1 , f2 , such that Au1 = f1 and Au2 = f2 . Since A is a linear operator we obtain A(u1 − u2 ) = f1 − f2 , which leads to ku1 − u2 kU ≤ A−1 V kf1 − f2 kV . Thus if we want a small change in right hand side to generate a small change in the solution we must require that A−1 is a bounded operator. So in summary we have the following requirements for our problem to be well posed: 2 (E) A is surjective. (U) A is injective. (C) A−1 is a bounded operator. Our goal is to show that A fulfills these three requirements. This was shown by Babuˇska [1971] and can be summarized as: Theorem 2.1 (Babuˇska-Lax-Milgram). The weak formulation in (1.1) is well posed if there exists αV , αU > 0 such that a(·, ·) fulfills: αV kukU ≤ sup v∈V αU kvkV ≤ sup u∈U |a(u, v)| ∀u ∈ U kvkV |a(u, v)| ∀v ∈ V. kukU (2.2) (2.3) Proof. The three requirements (E), (U ) and (C) are proved separately in Lemma 2.2, 2.3 and 2.4. Remark. The conditions (2.2) and (2.3) are frequently referred to as inf-sup conditions. The reason for this is that they can be formulated as: there exist αU , αV > 0 such that |a(u, v)| ≥ αV inf sup u∈U v∈V kuk kvk U V inf sup v∈V u∈U |a(u, v)| ≥ αU . kukU kvkV Lemma 2.2 ((U ) Injectivity). If a(·, ·) satisfies (2.2) its Ritz-representation is injective. Proof. By the definition of the Ritz-representation and the Cauchy-Schwartz inequality we have | hAu, viV | kAukV kvkV |a(u, v)| = ≤ = kAukV . kvkV kvkV kvkV By taking the supremum and utilizing (2.2) we obtain: αV kukU ≤ sup v∈V |a(u, v)| ≤ kAukV . kvkV 3 (2.4) This gives us that Au = 0 =⇒ u = 0 which implies injectivity since Au1 − Au2 = 0 =⇒ A(u1 − u2 ) = 0 =⇒ u1 − u2 = 0. Lemma 2.3 ((E) Surjectivity). If a(·, ·) satisfies (2.2) and (2.3) its Ritzrepresentation is surjective. Proof. Consider the range of A: R(A), this is a subspace of V . In order to show that this is a complete subspace let {yj }∞ j=1 be a Cauchy sequence in R(A). For every > 0 there exists N such that kyi − yj kV ≤ ∀i, j ≥ N. But since yi ∈ R(A) we have yi = Axi with xi ∈ U so we get kAxi − Axj kV ≤ ∀i, j ≥ N, which by linearity and (2.4) gives us: αV kxi − xj kU ≤ kA (xi − xj )kV ≤ ∀i, j ≥ N. This shows us that {xj }∞ j=1 is a Cauchy sequence, which converges since U is a Hilbert space. Denote the limit by x and let y = Ax. Since we originally assumed that A is a bounded operator we now see that yj converges to y: ky − yj kV = kAx − Axj kV ≤ αC kx − xj kU → 0 j → ∞. Thus the conclusion is that R(A) is a complete subspace of V . This can only happen if R(A) is also closed. Since R(A) is a closed subspace of V we can decomposed V as a direct product space: M V = R(A) R(A)⊥ . (2.5) But consider now a point v ∈ R(A)⊥ . From (2.3) we get αU kvkV ≤ sup u∈U | hAu, vi | = 0, kukU since Au ∈ R(A) and v ∈ R(A)⊥ . So this implies v = 0, which means that R(A)⊥ = {0} since v was arbitrary. By (2.5) we must then have R(A) = V , so that A is surjective. 4 Lemma 2.4 ((C) Bounded Ritz-Inverse). Given that a(·, ·) fulfills (2.2), the inverse of the Ritz-projection fulfills −1 A ≤ 1 , U αV and thus is a bounded operator. Proof. By rewriting (2.4) we obtain kukU ≤ 1 kAukV . αV Since we now know that A is a bijection we knot that for each u ∈ U there exist an unique f ∈ V such that u = A−1 f . Inserting this gives us −1 A f ≤ 1 kf k . V U αV Since this holds for all f ∈ V we obtain −1 −1 A = sup kA f kU ≤ 1 . V kf kV αV f ∈V References I. Babuˇska. Error-bounds for finite element method. Numerische Mathematik, 16:322–333, 1971. ISSN 0029599X. doi: 10.1007/BF02165003. 5 Uniformly best wavenumber approximations by spatial central difference operators Viktor Linders∗ January 12, 2015 Abstract We show that the problem of finding uniformly best wavenumber approximations by central difference schemes is equivalent to approximating a continuous function from a finite dimensional subspace. A characterisation theorem for best approximations is proven. 1 Introduction and motivation Consider a central difference approximation of a first derivative ux of some function u(x, t) at the point x = xi , ∂u ∂x i p + O(∆x2p ) = 1 X (p) c (ui+k − ui−k ). ∆x k=1 k (1.1) (p) Here ∆x is the spatial step size of the discretisation and ck are the coefficients of the 2pth order classical central difference scheme. It is well known that the numerical dispersion relation of this scheme is ξ¯c = 2 p X (p) ck sin (kξ). (1.2) k=1 where ξ = κ∆x is the normalised wavenumber. The subscript c serves as a reminder that (1.2) corresponds to a classical stencil. Here we have introduced an overbar to signify that the wavenumber is a numerical approximation. ∗ Division of Computational Mathematics, Department of Mathematics, Link¨ oping University, SE-581 83 Link¨oping, Sweden. viktor.linders@liu.se 1 Noting that the smallest wavelength the scheme can resolve is λmin = 2∆x, the largest wavenumber is κmax = 2π/λmin = π/∆x so that we have |ξ| ≤ π. Typically, waves with high frequencies and wavenumbers require small spatial increments, ∆x, in order to be properly resolved. Over sizeable intervals, ¯ may come to dominate the error in the the so called dispersion error, |ξ − ξ|, approximation, which gravely restricts the ∆x that may be used. Problems of this type are commonly encountered in computational fluid dynamics, aeroacoustics, electromagnetism, elasticity, seismology, and other fields where energy is propagated. Wave properties encoded within the dispersion relation include phase velocity, group velocity and dissipation. It is therefore of interest to develop difference schemes that preserve the analytic dispersion relation of the governing equations for a wide range of spatial increments. 2 Problem formulation Let us perturb the stencil (1.1) by adding n new coefficients, without increasing the accuracy. Thus, the scheme we consider has the form p+n 1 X ∂u 2p ak (ui+k − ui−k ). (2.1) + O(∆x ) = ∂x i ∆x k=1 Such a scheme uses (p + n) points on either side of xi to approximate the derivative, however the formal order of accuracy remains O(∆x2p ). In view of (1.2) the numerical wavenumber corresponding to this scheme is ξ¯ = 2 p+n X ak sin (kξ). (2.2) k=1 Our goal may thus be formulated in terms of the following problem: Choose ak , k = 1, . . . , p + n, such that the error function ¯∞ kEk∞ := kξ − ξk is minimised, ensuring that the stencil has formal accuracy O(∆x2p ). (p) (2.3) Note that if ap+j = 0, ∀j ≥ 1, we have ak = ck . By linearity we may thus write n X (p) (j) ak = c k + αk ap+j , k = 1, . . . , p (2.4) j=1 2 (j) where the parameters αk are independent of ap+j . We summarise in the following proposition: Proposition 2.1. The numerical wavenumber of the 2pth order central difference stencil (2.1) can be written as ξ¯ = ξ¯c + n X ap+j φj (ξ) j=1 where ξ¯c is the wavenumber apprpximation of the classical central difference stencil and p X (j) φj (ξ) = 2 sin ((p + j)ξ) + 2 αk sin (kξ). k=1 (j) The coefficients αk are independent of ap+j for all k = 1, . . . , p and j = 1, . . . , n. Proof. Inserting (2.4) into (2.2) and collecting terms multiplying ap+j immediately gives the desired result. The set {φj } is linearly independent and thus span some n-dimensional vector space, which we will denote by Ξn . We may thus consider each φj as a basis function of Ξn . P Let ψ(ξ, a) = nj=1 ap+j φj (ξ). We see from Proposition 2.1 that the problem (2.3) may equivalently be written find a ∈ Rn that minimises kE(ξ, a)k∞ = kEc (ξ) − ψ(ξ, a)k∞ , (2.5) where Ec (ξ) = ξ − ξ¯c is the error of the classical stencil and ψ(ξ, a) ∈ Ξn . 3 Approximation theory For convenience we use the following definition: Definition 3.1. If a vector a solves problem (2.5), then ψ(ξ, a) is called a best approximation. Finding and characterising best approximations are central problems in the field of approximation theory. It is well-known that a solution to problem (2.5) exists Riesz [1918]. It is in place to introduce some relevant concepts: 3 Definition 3.2. A set of functions, {θi }, i = 1, . . . , n forms a Chebyshev set on the interval [0, ξmax ] if any non-trivial linear combination has at most (n − 1) zeros in [0, ξmax ]. We will also need the following definition: Definition 3.3. Let M (a) denote the set of points {ξ1 , ξ2 , . . . } at which |E(ξ, a)| = kE(ξ, a)k∞ . Then E(ξ, a) is said to alternate n times on [0, ξmax ] if M (a) contains (n + 1) points 0 ≤ ξ1 < ξ2 < · · · < ξn+1 ≤ ξmax such that E(ξi , a) = −E(ξi+1 , a), i = 1, 2, . . . , n. The set {ξi } is referred to as an alternating set. The usefulness of the above definitions becomes apparent when we consider the following well-known extension of the Chebyshev alternation theorem: Theorem 3.1. Let {θi (ξ)}, i = 1, . . . , n be a Chebyshev set on [0, ξmax ]. Then a solves problem (2.5) if and only if there is an alternating set of (n + 1) points in [0, ξmax ]. Such a solution exists and is unique. For a proof, see e.g. Watson [1980]. Theorem 3.1 identifies best approximations as those whose error oscillates between a positive and a negative fixed value. The main task left is thus to show that the set {φj } constitutes a Chebyshev set. 4 Characterisation of best approximations The following observation is useful. The proof is omitted for brevity. Lemma 4.1. Consider the function f (p) (x, a) = 1 − 2 p+n X ak kTk (x) k=1 where ak is defined as before and Tk (x) is the k th order Chebyshev polynomial of the first kind, uniquely defined through the relation Tk (cos (φ)) = cos (kφ). Then f (p) has a root at x = 1 of multiplicity p. We are in position to prove our main result: Theorem 4.2. The set {φj }, j = 1, . . . , n defined in Proposition 2.1 forms a Chebyshev set on the semi-open interval (0, ξmax ]. 4 Proof. Recall the form of the basis functions, φj (ξ) = 2 sin ((p + j)ξ) + 2 p X (j) αk sin (kξ). k=1 Note first that φj (ξ) is a trigonometric polynomial. Thus the number of zeros of any non-trivial combination, ψ(ξ, a) is finite in a bounded interval. The function f (p) (x, a) is a polynomial in x of degree p + n with a root of multiplicity p at x = 1, i.e. f (p) (x, a) = (1 − x)p Pn (x, a), where Pn (x, a) is some polynomial of degree n that depends linearly on a. Let x = cos (ξ). We have f (p) (cos (ξ), a) = 1 − 2 p+n X k=1 ak kTk (cos (ξ)) = 1 − 2 p+n X ak k cos (kξ) = k=1 dE . dξ Thus, E(ξ, a) has an extreme point at ξ = 0 and at most n other extreme points ξ1 ≤ ξ2 ≤ · · · ≤ ξn in [0, ξmax ]. If ξ1 = 0 the stencil would be of higher order than O(∆x2p ) so we may assume that 0 < ξ1 . Consider now another error function, E(ξ, b). Linearity gives ∂ (E(ξ, a) − E(ξ, b)) = (1 − cos (ξ))p [Pn (cos (ξ), a) − Pn (cos (ξ), b)] ∂ξ = (1 − cos (ξ))p Pn (cos (ξ), a − b) but also ∂ (E(ξ, a) − E(ξ, b)) ∂ = [(Ec (ξ) − ψ(ξ, a)) − (Ec (ξ) − ψ(ξ, b))] ∂ξ ∂ξ ∂ = ψ(ξ, b − a), ∂ξ where, as before, ψ is a linear combination of the basis functions, {φj }. The vectors a and b are general so we may write c = b − a for any vector c ∈ Rn . Thus ψ(ξ, c) has at most n extrema in the open interval (0, ξmax ]. Obviously there is at most one zero between any two consecutive extreme points. Thus ψ(ξ, c) has at most n − 1 zeros in (0, ξmax ]. This completes the proof. 5 Corollary 4.3. Let {φj (ξ)}, j = 1, . . . , n be defined as in Proposition 2.1. Then a solves problem (2.5) if and only if there exists an alternating set of (n + 1) points in (0, ξmax ]. Such a solution exists and is unique. Proof. Theorem 4.2 shows that for any δ satisfying 0 < δ < ξmax , {φj } is a Chebyshev set on [δ, ξmax ]. Thus Theorem 3.1 applies. Letting δ → 0 and noting that at ξ = 0 the approximation is exact yields the desired result. 5 Example We conclude with a brief example: Figure 5.1 shows the error of a stencil optimised with respect to the max-norm. The stencil is second order accurate and is 31 points wide. Here we have set ξmax = π/2. In this region the dispersion error is bounded from above by kEk∞ ∼ O(10−12 ). Clearly the alternation property holds as expected from Theorem 3.1. 1.5 kEk∞ = 1.337520e − 12 ×10 -12 1 E 0.5 0 -0.5 -1 -1.5 0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 ξ Figure 5.1: Optimised stencil for the case p = 1 and n = 14, i.e. the stencil has formal accuracy O(∆x2 ) and is 31 points wide. 6 References ¨ F. Riesz. Uber lineare functionalgleichungen. Acta Mathematica, 41:71–98, 1918. G. A. Watson. Approximation Theory and Numerical Methods. John Wiley & Sons Ltd., 1980. 7 The Ekeland Variational Principle With Applications Markus Wahlsten ∗ January 12, 2015 Abstract We state and prove the Ekeland Variational Principle and exemplify its usefulness with some applications. 1 Introduction The Ekeland Variational Principle is an important tool in areas such as nonlinear analysis and optimization theory. The principle is often used to establish existence of solutions for optimization problems where the Weierstrass’ theorem does not apply. Further, it’s proven to be a valuable tool when studying partial differential equations Ekeland [1974]. 2 The Ekeland Variational Principle Before we state The Ekeland Variational Principle we recall the Cantor’s intersection theorem and some definitions. Theorem 2.1. (Cantor’s intersection theorem) Let {Sk } be a sequence of non-empty, closed and bounded sets satisfying S0 ⊇ S1 ⊇ · · · Sk ⊇ Sk+1 · · · , then ∞ \ k Sk ! ∗ 6= ∅. (2.1) (2.2) Department of Mathematics, Computational Mathematics, Link¨ oping University, SE581 83 Link¨oping, Sweden markus.wahlsten@liu.se 1 Definition 2.1. (Upper and lower semicontinuity) Let f be a function and x0 a point such that f : R → R and x0 ∈ R. The functiom f is said to be upper semicontinuous at the point x0 if f (x0 ) ≥ lim sup f (x) (2.3) f (x0 ) ≤ lim inf f (x) (2.4) x→x0 or lower semicontinuous if x→x0 Theorem 2.2. (The Ekeland Variational Principle) Let X be a complete metric space and φ : X → R an upper semi-continuous map that is bounded from above. Then, for any > 0 there exists a y ∈ X such that, φ(y) > φ(x) − d(x, y), ∀x ∈ X\{y} (2.5) Remark. The theorem can be modified by replacing the words ”upper” with ”lower” when also ”above” is replaced by ”below” and finally in (2.5) the ”−” and ”>” is replaced by ”+” and ”<” respectively. Proof. By the fact that sup φ(X) < ∞, ∃z0 ∈ X such that, φ(z0 ) > sup φ(X) − . (2.6) ∞ If we now construct two sequences {zi }∞ i=0 and {Si }i=0 inductively in the following manner 1. Assume zi is given 2. Define Si := {x ∈ X : φ(zi ) ≤ φ(x) − d(x, zi )} 3. Pick any element zi+1 ∈ Si such that 1 φ(zi+1 ) ≥ sup φ(Si ) − i+1 (2.7) Since φ is upper semi-continuous, the map φ(·) − d(·, zi ) is also upper semicontinuous, which leads to the fact that Si is closed. Next, we need to prove that Si+1 ⊆ Si for i = 0, 1, . . . . First, we pick an arbitrary i ∈ Z+ and let x ∈ Si+1 . Then, by using the triangle inequality, and the facts that x ∈ Si+1 and zi+1 ∈ Si , we obtain the relation φ(x) − d(x, zi ) ≥ φ(x) − d(x, zi+1 ) − d(zi+1 , zi ) ≥ φ(zi+1 ) − d(zi+1 , zi ) ≥ φ(zi ) which means x ∈ Si and since x was arbitrarily chosen Si+1 ⊆ Si . 2 (2.8) Further, in order to apply the Cantor Frechet Intersection Theorem we also need to prove that δ(Si ) = sup d(x, y) ≤ x,y∈Si 2 . i (2.9) By picking an i ∈ Z+ , and an x ∈ Si+1 , which also is x ∈ Si which we know from the proof above. By using this fact and the inequality in (2.7) we get, d(x, zi ) ≤ φ(x) − φ(zi ) ≤ sup φ(Si−1 ) − φ(zi ) ≤ 1i (2.10) which proves the desired result. By applyingT the Cantor Frechet Intersection Theorem we conclude that the intersection +∞ i=0 Si = {y} for some y ∈ X. To conclude our proof we need to show that the element y satisfies (2.7). This is done by assuming the contrary, let x ∈ X\{y}, hence we have φ(x) − d(x, y) ≥ φ(y), (2.11) φ(x) − d(x, y) ≥ φ(y) ≥ φ(zi ) + d(x, zi ) (2.12) then since (2.12) is satisfied for all i = 0, 1, . . . we have φ(x) ≥ φ(zi ) + (d(x, y) + d(y, zi )) ≥ φ(zi ) + d(x, zi ) (2.13) is true for all i = 0, 1, . . . , hence x ∈ hence a contradiction. 3 T+∞ i=0 (2.13) Si , which means x = y, Applications In this section several useful applications of the Ekeland Variational Principle is presented. Theorem 3.1. Caristi’s Fixed Point Theorem Let ψ be a self-map on a complete metric space X. If d(x, ψ(x)) ≤ φ(x) − φ(ψ(x)), ∀x ∈ X (3.1) for some lower semi-continuous φ ∈ RX that is bounded from below, then ψ has a fixed point in X. 3 Proof. From the Ekeland Variational Principle, we know that there exists some y ∈ X such that, φ(y) < φ(x) + d(x, y), ∀x ∈ X\{y}. (3.2) However, by (3.1) we have that φ(y) ≥ φ(ψ(y)) + d(y, ψ(y)). (3.3) So ψ(y) ∈ / X\{y}, that is ψ(y) = y. Definition 3.1. (Palais-Smale condition) We say that a C 1 functional φ : X → R satisfies the Palais-Smale condition if every sequence {un }∞ n=1 ∈ X which satisfies kψ(un )k ≤ const, and ψ(un ) → 0 in X possesses a convergent subsequence. (3.4) Theorem 3.2. Let ψ : X → R be a C 1 functional where X is a Hilbert space. Let S be a closed and convex subset of X. Suppose that • K ≡ I − ψ 0 which maps S into S • ψ is bounded from below in S • ψ satisfies definition 3.1 in S Then there exists a u0 ∈ S such that ψ 0 (u0 ) = 0 and inf ψ(u) = ψ(u0 ) u∈S Proof. We start by applying the Ekeland Variational Principle to ψ. Given there exists a u ∈ C such that ψ(u ) ≤ inf ψ(u) + and u∈S ψ(u ) ≤ ψ(u) + ku − u k , ∀u ∈ S (3.5) If we now let u = (1 + t)u + tKu , where 0 < t < 1. Next, we use Taylors formula to expand ψ(u + t(Ku − u )) about u and obtain, 2 t kψ 0 (u )k ≤ kψ 0 (u )k + O(t), (3.6) which implies kψ 0 (u )k < . Finally we use the property (3.1). 4 Summary The Ekeland Variational Principle has been stated and proved in section 2 followed by some useful applications of the theorem in section 3. 4 References I. Ekeland. On the variational principle. Journal of Mathematical Analysis and applications, 47:324–353, 1974. 5 Construction of Sobolev spaces Cristina La Cognata∗ January 15, 2015 Abstract Sobolev spaces are Banach spaces that are characterized by the definition of ”weak” derivatives. The functions in Sobolev spaces and their weak derivatives lie in Lp spaces. Introduction This essay is an introduction to the theory of “Sobolev” spaces. They are powerful tools from Functional Analysis for defining the solutions of Partial Differential Equations (PDE). In particular, they are used in those classes of problems in which smoothness and continuity are too strong requirements. In a certain sense, functions that belong to Sobolev spaces represent a good compromise as they have some, but not all, smoothness properties [Evans, 2002]. We start with the definition of the concept of weak derivatives and then we will apply it to the construction of the Sobolev spaces. We conclude with an overview of some of the most interesting properties concerning these spaces. 1 Weak derivatives A kth-order PDE can be symbolically represented by the following expression F (Dk u(x), Dk−1 u(x), ..., Du(x), u(x), x) = 0, x ∈ U, (1.1) where k ≥ 1 is a fixed integer, U is an open set in Rn and u : U → R is the unknown. The equation (1.1) involves an unknown function u : U → R, ∗ Division of Computational Mathematics, MAI, Link¨ oping University, SE-581 83 Link¨oping, Sweden. cristina.la.cognata@liu.se 1 of variable x ∈ U , and some of its partial derivatives. By solving a PDE of type (1.1) we mean (ideally) finding a simple (in the best case), explicit expression for such a u, or at least guaranteeing its existence together with some properties. Let us ignore for the moment the strict concept of a wellposed PDE problem. What do we require for u to be a “solution” of (1.1)? It is natural to desire having a solution to be at least k times continuously differentiable (u ∈ C k ). A solution of this type is called “classical” solution. Unfortunately there exist certain PDEs (or most of them) which cannot be solved in a classical sense. For instance, in general the conservation law of the form ut + F (u)x = 0, x ∈ U, (1.2) has no classical solutions Evans [2002],R.LeVeque [1992]. Then we need to allow a more general or weaker class of functions to be our solutions. Let us denote by Cc∞ (U ) the spaces of infinitely many times differentiable functions φ : U → R, with compact support on U . We will refer to a function in Cc∞ (U ) as a test function. Furthermore, we will widely use Lp spaces and their norms. For a a reader which is not familiar with these type of spaces we remand to Evans [2002], R.LeVeque [1992]. Definition. Suppose that u, v ∈ L1loc (U )1 and α is a multi-index. We say that v is the αth-weak derivative of u if Z Z α |α| uD φ dx = (−1) vφ dx (1.3) U U for all test functions φ ∈ Cc∞ (U ). In other words, a weak derivative is a generalization of the concept of the derivative of a function (strong derivative) for functions not assumed differentiable, but only integrable in the Lebesgue sense. Note that, since u is not necessarily a differentiable function, the locally summability assumption is needed to ensure the existence of the right-hand side of (1.3). The notion of weak derivatives originated with the works of J. Leray (19061998) and later from S. Sobolev (1908-1989) when he came to deal with non-continuous functions with almost everywhere existing derivatives. Lemma 1.1. The αth-weak derivative of u is uniquely defined almost everywhere2 1 2 see Appendix see Appendix 2 Proof. Let us assume that v, w ∈ L1loc (U ) both satisfy Z Z Z α |α| |α| uD φ dx = (−1) vφ dx = (−1) wφ dx, U for all test functions φ ∈ U Cc∞ (U ). Z U U Then (v − w)φ dx = 0, for all test functions φ ∈ Cc∞ (U ). Hence, (v − w) = 0 a.e. Example. Let n = 1, U = (a, b) and consider u, v defined as follows x a < x ≤ (a + b)/2 u(x) = 1 (a + b)/2 ≤ x < b 1 a < x ≤ (a + b)/2 v(x) = 0 (a + b)/2 ≤ x < b. We can show that v is the weak derivative of u by choosing any test function φ ∈ Cc∞ (U ) and calculate Rb 0 R (a+b)/2 0 Rb uφ dx = a xφ dx + (a+b)/2 φ0 dx a R (a+b)/2 = − a φ dx + [φ((a + b)/2) − φ((a + b)/2)] ((a + b)/2) = Rb = − a vφ dx. as required. 2 Sobolev spaces We now define a function space which contains elements with weak derivatives of various orders from Lp spaces. Let us fix 1 ≤ p ≤ ∞ and let k be a nonnegative integer. Definition (Sobolev spaces). The Sobolev space W k,p (U ) consists of all locally summable functions u : U → R such that for each multi-index α with |α| ≤ k, the weak derivative Dα u exists and lie in Lp (U ). Definition (Sobolev norms). W k,p (U ) is a normed space equipped with the norm defined as 1/p R P α p (1 ≤ p ≤ ∞) |α|≤k U |D u| dx kukW k,p (U ) = , (2.1) P α ess sup |D u| (p = ∞) U |α|≤k where u ∈ W k,p (U ). For the definition of ess sup f see 3 3 We finally introduce the following notation for convergence: k,p Definition (Sobolev convergence). We say that {um }∞ (U ) m=1 , um ∈ W k,p converge to u ∈ W (U ) provided that lim kum − ukW k,p (U ) = 0. m→∞ Next we list some elementary properties Sobolev functions. The proofs can be found in Evans [2002]. Note that the following properties are trivially true for functions in C k (U ) spaces. For Sobolev functions, instead, we must think about these rules as in the weak definition. Theorem 2.1. Assume u, v ∈ W k,p (U ), |α| ≤ k. Then 1. Dβ (Dα u) = Dα (Dβ u) = Dα+β u for all |α| + |β| ≤ k. 2. For each λ, µ ∈ R, λu + µv ∈ W k,p (U ) and λDα u + µDα , |α| ≤ k. 3. if V is an open set in U , then U ∈ W k,p (V ). Some of these properties are needed to prove that the Sobolev spaces are Banach spaces. Theorem 2.2 (Sobolev spaces as a function spaces). For each k = 1, ..., and 1 ≤ p ≤ ∞, the Sobolev space W k,p (U ) is a Banach space. Proof. First we check that (2.1) is a norm. It is easy to check from the definition that kλukW k,p (U ) = |λ|kukW k,p (U ) kukW k,p (U ) = 0 and iff u=0 a.e. (2.2) Next we assume that u, v ∈ W k,p (U ) with 1 ≤ p < ∞. Then ku + vkW k,p (U ) = P |α|≤k kD u+D α vkpLp (U ) 1/p p 1/p α α p (U ) + kD vkLp (U ) kD uk L |α|≤k P 1/p P 1/p p p α α ≤ kD uk + kD vk p p |α|≤k |α|≤k L (U ) L (U ) = kukW k,p (U ) + kvkW k,p (U ) . ≤ P α Note that in the second inequality we have used Minkowski’s relation defined in 3. Next we prove that W k,p (U ) is complete by showing that any Cauchy sequence um ∈ W k,p (U ) converges. 4 1. For any |α| ≤ k, Dα um is a Cauchy sequence in Lp (U ). Since Lp (U ) is complete, the existence of u and uα , ∀α, such that um → u and Dα um → uα in Lp (U ), (2.3) is guaranteed. 2. Consider any test function φ ∈ Cc∞ (U ). Then we have R R α uD φ dx = lim u Dα φ dx m→∞ U U m R = limm→∞R(−1)|α| U Dα um φ dx = (−1)|α| U uα φ dx. Thus, uα = Dα u, |α| ≤ k. 3. From the two previous steps we find that um → u in Lp (U ) and the uα are the weak derivatives of u. Therefore um → u in W k,p (U ). Thus W k,p (U ) is a Banach space. We conclude by noting that in the special case p = 2, the Sobolev space has a more interesting structure. Definition (Hilbert-Sobolev spaces). We define the Hilbert-Sobolev space as H k (U ) = W k,2 (U ). The space H k (U ) is a Hilbert space endowed with the inner product XZ (u, v)H k (U ) = Dα uDα v dx. |α|≤k U Note that H 0 (U ) = L2 (U ). For the curious reader, more details and results can be found in Brezis [1983]. 3 Appendix Definition (Almost everywhere). If a property holds everywhere in Rn , except for a measurable set of Lebesgue measure zero, we say that the property is true almost everywhere (a.e.). Definition (ess sup). If a real function f is measurable, the essential supremum of f is defined as ess sup f = inf {µ ∈ R s.t. |f > µ| = 0} . 5 Definition (L1loc functions). L1loc = {u : U → R s.t. u ∈ Lp (V ) for each V ⊂⊂ U } . Definition (Minkowski’s inequality). Assume 1 ≤ p ≤ ∞ and u, v ∈ Lp (U ). Then ku + vkLp (U ) ≤ kukLp (U ) + kvkLp (U ) . The Minkowski inequality is the triangle inequality in Lp (U ). References H. Brezis. Analyse Fonctionnelle. Masson, 1983. L. Evans. Partial Differential Equations. AMS, 2002. R.LeVeque. Numerical Methods for Conservation Laws. Birkh¨auser, 1992. 6 The Arzel´a-Ascoli theorem Tomas Lundquist∗ January 15, 2015 Abstract We discuss the Arzel´ a-Ascola theorem from a functional analysis point of view, and give examples of applications within the field, in particular relating to the theory of compact embeddings in function spaces. 1 Introduction The Arzel´a-Ascola theorem is of fundamental importance in both real analysis and functional analysis, as well as in topology. In functional analysis, it is particularly useful in function space theory, providing a constructive criterion for compactness of subsets in C(X), the set of complex valued continuous functions on a metric space X. In its original form, the theorem was formulated for sequences of continuous functions on compact intervals of the real line, giving a criterion for the existence of uniformly bounded subsequences. Later it was extended to compact metric spaces then further to compact Hausdorff topological spaces. Here we will focus on the functional analysis version of the theorem. In section 2 we present the theorem and discuss some of its implications. Section 3 is a brief introduction to the theory of compact embeddings in continuous function spaces, to which the theorem is closely related. 2 Statement of the theorem The Arzel´a-Ascoli theorem gives a necessary and sufficient condition for compactness of continuous function spaces. Before we proceed with stating the theorem, we need the following definition. ∗ Division of Computational Mathematics, Department of Mathematics, Link¨oping university, SE-581 83 Link¨oping, Sweden. tomas.lunqduist@liu.se 1 Definition 2.1. A set F of functions between two metric spaces X and Y is said to be equicontinuous if for each ϵ > 0, ∃ a δ > 0 such that d(f (x0 ), f (x1 )) < ϵ for all f ∈ F , and x0 , x1 ∈ X with d(x0 , x1 ) < δ. The parameter δ may depend on x0 but not on f . Note the contrast between equicontinuity and uniform continuity. In the former case, ϵ is not allowed to depend on f , whereas in the latter case, ϵ is not allowed to depend on x0 . The metric space version of the Arzel´a-Ascoli theorem now follows below. Theorem 2.1. Let X be a compact metric space, and denote with C(X) the space of real (or complex) valued continuous functions on X with respect to the supremum norm. Then a subset M of C(X) is relatively compact in C(X) if and only if M is equicontinuous and bounded. Proof. A proof can be found e.g. in Nagy [2006]. ¯ of We recall that relatively compact in this case means that the closure M ¯ M is compact in C(X); that is, each sequence in M has a convergent subsequence. The Arzela-Ascola theorem is used in the proof of many important theoretical results. One example is Peano’s theorem of local existence for continuous ordinary differential equations. Another is to prove that an operator between Banach spaces is compact if and only if the adjoint operator is compact (see e.g. Garrett [2012]). 3 Application: compact embeddings of continuous functions The concept of compact embeddings between function spaces is important not the least in the study of numerical methods for partial differential equations. For example, the famous Sobolev embedding theorems are examples of results relating to compact embeddings. These results are important when applying Sobolev spaces to partial differential equations. For a discussion of this, see e.g. Dlotko [2014]. An embedding E : X ← Y from a subset X ⊂ Y into Y is simply the identity operator. As an example, consider the space C 1 (Ω) of C 1 −differentiable functions on the domain Ω ∈ Cn , equipped with the norm ∥f ∥C 1 = sup|f | + sup|∇f |. With this definition, the embedding E(C 1 (Ω)) into C(Ω) is the space formed by the set C 1 (Ω) and standard supremum norm. As we shall later see, this embedding is indeed compact. 2 Definition 3.1. Let (X, ∥·∥X ) and (Y, ∥·∥Y ) be two normed spaces, where X is a subset of Y . We say that X is compactly embedded in Y if the following holds: 1. ∃ c > 0 : ∥x∥Y ≤ c∥x∥X ∀x ∈ X. (uniform continuity) 2. The embedding of X into Y is a compact operator. (compactness) We recall that an operator is compact if it maps bounded sets to relatively compact sets. The Arzel´a-Ascoli theorem provides a constructive way to show compactness of embeddings. Too see this, we continue with the example above. Thus, consider a bounded subset M of the space C 1 (Ω). The boundedness of sup|f | and sup|∇f | then implies equicontinuity of the set (this part is left as an exercise). From the Arzel´a-Ascoli theorem, it also follows that M is relatively compact in C(Ω). i.e. the embedding E(M) into C(Ω) is compact. Moreover, uniform continuity follows directly from the definitions of the norms involved: ∥f ∥C(Ω) = sup|f | ≤ sup|f | + sup|∇f | = ∥f ∥C 1 (Ω) . Thus the space C 1 (Ω) is compactly embedded in C(Ω) with respect to the norm given above. Analogous results to the one given above can be derived using the Arzol´aAscoli theorem for a variety of continuous function spaces. In particular the Sobolev embedding theorems, of fundamental importance for Sobolev space theory, can be proven using similar arguments. The most famous of these result states that the embedding H m+1 (Ω) in H m (Ω) is compact, where H m (Ω) is the Sobolev space with square integrable weak derivatives up to order m. Note that H 0 (Ω) = L2 (Ω), so that all Sobolev spaces are compactly embedded in the space of square integrable functions. A full proof of this, generalized to Lp (Ω), is given e.g. in Adams [1975]. 4 Summary We have reviewed the metric space formulation of the Arzol´a-Ascoli theorem, and shown its usefulness as a constructive criterion for compactness of continuous function spaces. An example of application is given by the famous Sobolev embedding theorems. References R. Adams. Sobolev spaces. Academic Press, New York, 1975. 3 T. Dlotko. Sobolev spaces and embedding theorems. www.icmc.usp.br/~ andcarva/sobolew.pdf, 2014. P. Technical report, Garrett. Compact operators on banach spaces: Fredholm-riesz. Technical report, http://www.math.umn.edu/~ garrett/m/fun/fredholm-riesz.pdf, 2012. G. Nagy. A functional analysis point of view of the arzela-ascoli theorem. Real Analysis Exchange, 32:583–586, 2006. 4 A Completeness Theorem for Non-Self-Adjoint Eigenvalue Problems Saleh RezaeiRavesh∗ January 16, 2015 Abstract An important completeness theorem for a general class of nonself-adjoint eigenvalue problems is stated. The corresponding proof is reproduced in connection with basic material in spectral theory and functional analysis. 1 Introduction In many applications in hydrodynamic stability analysis we are confronted with eigenvalue problems in the form (L − λM )ϕ = 0 (1.1) where L and M are differential operators and λ ∈ σ(L, M ). The associated boundary conditions for (1.1) are ϕ(0) = Dϕ(0) = ϕ(1) = Dϕ(1) = 0, where, D := d/dy is the differentiation operator. One of the famous examples is the Orr-Sommerfeld equation which is derived in linearized stability analysis of parallel flows (e.g., cf. Drazin and Reid [1981]). The general idea for studying the spectrum of (1.1) is to write L := Ls +B where Ls and M are self-adjoint operators and B is called perturbation. In Section 2, some of the definitions and theorems required in this document are expressed. Then, theorem 3.1 shows how we can find a self-adjoint extension operator G for M −1 Ls in such a way that two problems Gϕ = λϕ and Ls ϕ = λM ϕ become equivalent. Subsequently, theorem 3.2 proves the completeness of the generalized eigenvectors of (1.1). Finally, an application of the completeness theorem is briefly explained. ∗ Division of Scientific Computing, Department of Information Technology, Uppsala university, SE-751 05 Uppsala, Sweden. saleh.rezaeiravesh@it.uu.se 1 2 Spectral Theory Definition 2.1 (Resolvent Set & Spectrum of T ) Let X 6= {0} be a complex normed space and T : D(T ) → X 1 a linear operator with D(T ) ⊂ X. Then, (a) The resolvent set ρ(T ) of T is the set of all regular values λ of T . If T : X → X is a bounded linear operator and X is a Banach space, then, ρ(T ) = {λ ∈ |(T − λI) is one-to-one and onto} Evans [2010] (Any λ ∈ σ(T ) is called a spectral value of T (cf. 7.2-1 Kreyszig [1989])), (b) The spectrum of T is the complement of ρ(T ); i.e., σ(T ) = − ρ(T ) (cf. 7.2-1 Kreyszig [1989]). Definition 2.2 (Resolvent Operator) If Tλ := (T − λI) has an inverse, denoted by Rλ (T ) := Tλ−1 = (T − λI)−1 , we call it the resolvent operator of T (cf. 7.2 Kreyszig [1989]). Definition 2.3 (Point, Continuous, Residual Spectra) The point or discrete spectrum σp (T ) is the set such that Rλ (T ) does not exist. λ ∈ σp (T ) is called an eigenvalue of T . The continuous spectrum σc (T ) is the set such that Rλ (T ), exists and has a domain dense in X, but is unbounded. The residual spectrum σr (T ) is the set such that Rλ (T ), (a) exists (may be bounded or not) but its domain is not dense in X (cf. 7.2-1 Kreyszig [1989]). C C Theorem 2.1. (Distance from σ(T )) If T ∈ B(X, X), where X is complex Banach space and λ ∈ ρ(T ), then (cf. 7.5-3 Kreyszig [1989]) kRλ (T )k ≥ 1 , d(λ, σ(T )) d(λ, σ(T )) := inf |λ − s| s∈σ(T ) (2.1) Definition 2.4 (Eigenvalue & Eigenvector) We say that λ ∈ σ(T ) is an eigenvalue of T provided N (T − λI) 6= {0}. We write σp (T ) to denote the collection of eigenvalues of T ; clearly, σp (T ) is a point spectrum. If λ is an eigenvalue and ϕ 6= 0 satisfies T ϕ = λϕ, we say ϕ is an associated eigenvector (cf. D.3 Evans [2010]). Definition 2.5 (Generalized Eigenvector of (L, M )) Let L and M be linear operators with D(L) ⊂ D(M ) dense in a Hilbert space H. We say that ψ ∈ H is a generalized eigenvector of (L, M ) if and only if for some λ and some integer p > 1, there exist p non-zero vectors ψ1 , ψ2 , · · · , ψp = ψ such that Lψi = λM ψi + M ψi−1 , i = 1, · · · , p, where ψ0 = 0. By allowing p = 1, eigenvectors become included in generalized eigenvectors (DiPrima and Habetler [1969]). 1 In this manuscript, D(T ), R(T ), and N (T ) respectively denote domain, range and null space of operator T . 2 Definition 2.6 (Completeness) The generalized eigenvectors of a linear operator with domain dense in H are said to span H if the following is true. Let ϕk1 , ϕk2 , · · · , ϕknk be the eigenvectors (and possibly generalized eigenvectors) corresponding to the eigenvalues in rk−1 < |λ| < rk for k > 1 and |λ| < r1 for k = 1; then each ϕ ∈ H has a representation ϕ= ni ∞ X X aij ϕij (2.2) i=1 j=1 with convergence in the norm of H (DiPrima and Habetler [1969]). Theorem 2.2. (Friedrichs Extension) A positive-bounded below (hence positive definite) operator with domain dense in a Hilbert space can be extended to a self-adjoint operator which possesses an inverse defined on the entire space (cf. p 11 Mikhlin [1965]). Theorem 2.3. (Naimark’s Theorem) If T is a closed linear operator on a dense subspace of a Hilbert space H with norm k.k whose resolvent (T − λI)−1 is compact for some λ = λ0 , and if there exists a sequence of circles Ck , |λ| = rk such that (a) there are no eigenvalues of T on any Ck , (b) limk→∞ rk = ∞, and (c) limk→∞ supλ∈Ck k(T − λI)−1 k = 0, then the generalized eigenvectors of T span H in the sense of (2.2) (e.g., see DiPrima and Habetler [1969]). 3 Completeness Theorem Theorem 3.1. (Equivalent EVP’s) Let Ls and M be linear operators with D(Ls ) ⊂ D(M ) dense in a Hilbert space H and such that, 1. M and Ls are positive-bounded below with R(M ) = R(Ls ), 2. G−1 is compact and R(G−1 ) ⊂ D(M ) where G−1 is the inverse of the self-adjoint extension of M −1 Ls in HM .2 Then, the two problems Gϕ = λϕ and Ls ϕ = λM ϕ are equivalent (DiPrima and Habetler [1969]). Proof: I Step 1. (Construct G−1 ): M is psoitive-bounded below, i.e., ∀x ∈ D(M ) ∃α > 0 : hM x, xi ≥ αkxk2 2 A new Hilbert space HM which is embedded in H can be constructed by completing the pre-Hilbert space HM (D(M ), [., .]), where [f, g] = hf, M gi for f, g ∈ D(M ) is the 1 corresponding inner product and the norm is kf kHM = [f, f ] 2 (cf. 3 Mikhlin [1965]). 3 Combining this with the Schwarz inequality results in, αkxk2 ≤ hM x, xi ≤ kM xkkxk Note that y = M x for x and y in D(M ) and R(M ) of M , respectively; hence, kM −1 yk ≤ 1 kyk α Therefore, M −1 exists and is bounded. Furthermore, M −1 Ls is positivebounded below in HM ; since ∀x ∈ D(Ls ) ⊂ D(M ), [x, M −1 Ls x] = hx, M M −1 Ls xi = hx, Ls xi ≥ βkxk2 , β>0 where the last inequality holds since Ls is positive-bounded below in H. According to Theorem 2.2, M −1 Ls has a self-adjoint extension (call it G) that has an inverse G−1 as mentioned in hypothesis 2. I Step 2. (Find the Spectrum of G): Since G−1 is compact, then it has a point spectrum, σp (G) (cf. 8.3-1, 8.3-3 Kreyszig [1989]). Furthermore, the set of eigenvalues G is countable (may be finite and even empty) and λ = 0 is the only possible point of accumulation of that set. In general, on D(G) ⊂ X which is a complex normed space two cases can be considered; if X is finite dimensional, then G has a matrix representation and clearly 0 may or may not belong to σ(G) = σp (G); that is, if dimX < ∞ we may have 0 ∈ / σ(G); then 0 ∈ ρ(G). However, if dimX = ∞, then we must have 0 ∈ σ(G) (cf. pp. 420 and 433 Kreyszig [1989]). I Step 3. (Equivalent Spectra): We show that σ(G) = σp (G) = σp (Ls , M ). 3.a) Suppose λ ∈ σ(G); then, there exists ϕ ∈ D(G) such that Gϕ = λϕ. Since from hypothesis 2, D(G) = R(G−1 ) ⊂ D(M ), then we have ϕ ∈ D(M ) and hence Gϕ ∈ D(M ). But, since R(M ) = R(Ls ), M −1 Ls is onto D(M ). This means that there exists ψ ∈ D(Ls ) such that M −1 Ls ψ = Gϕ, and since G = M −1 Ls on D(Ls ) and G−1 exists, we have ψ = ϕ. As a result, M −1 Ls ψ = M −1 Ls ϕ = λϕ ⇒ Ls ϕ = λM ϕ ⇒ λ ∈ σp (Ls , M ) 3.b) Conversely, consider λ ∈ σp (Ls , M ) and correspondingly, Ls ϕ = λM ϕ, ϕ 6= 0. Then immediately we have M −1 Ls ϕ = λϕ an hence Gϕ = λϕ and λ ∈ σ(G). Consequently, σp (Ls , M ) ∈ σ(G) and an eigenvector of G is an eigenvector of (Ls , M ). This completes the proof of equivalency of the Gϕ = λϕ and Ls ϕ = λM ϕ . Theorem 3.2. (Completeness) Let L = Ls + B, and let Ls , B, and M be linear operators with D(Ls ) = D(B) ⊂ D(M ) dense in a Hilbert space H, such that, 4 1. Ls and M satisfy conditions of theorem 3.1, 2. M −1 B is bounded in HM and D(Ls ) is dense in HM , 3. for some λ, R(Ls + B + λM ) = R(M ) and (G + M −1 B + λI) has an inverse, 4. there exists a sequence of concentric circles {Ck } with the radii {rk } such that, (a) limk→∞ rk = ∞, (b) d(Ck , σp (Ls , M )) > 0 for all k, (c) limk→∞ d(Ck , σp (Ls , M )) = ∞. Then, 1. the eigenvalues of (L, M ) lie within circles of radii kM −1 BkHM about those of (Ls , M ), and, 2. the generalized eigenvectors of (L, M ) span HM in the sense of (2.2) (DiPrima and Habetler [1969]). Proof: I Step 1. (Define T and Show it is Closed) Consider T := G + M −1 B where M −1 B is a unique bounded extension from D(B) to D(G). Now, we show that T satisfies the condition of Naimark’s theorem 2.3. According to the proof of theorem 3.1, G is self-adjoint and therefore closed. In addition by hypothesis 2., M −1 B is bounded and when we add it to the closed operator G, we end up with closed operator T . I Step 2. (Show that ∃λ0 : Rλ (T ) = (T − λ0 I)−1 is Compact): Let’s choose λ0 such that d(λ0 , σ(G)) > kM −1 BkHM . This is always possible; since according to hypothesis 4.(c), it is enough to choose λ0 on Ck for sufficiently large k. G−1 is compact and G is closed, so (G − λI)−1 is compact for all λ ∈ ρ(G) (cf. 5.7 Engel and Nagel [2006]), and hence for λ = λ0 . Now, we can show that (T − λ0 I)−1 is compact. we have, (T − λ0 I)−1 = (G + M −1 B − λ0 I)−1 = [I + (G − λ0 I)−1 M −1 B]−1 (G − λ0 I)−1 By using theorem 2.1, we can show that [I +(G−λ0 I)−1 M −1 B]−1 is bounded: kI+(G−λ0 I)−1 M −1 BkHM ≥ 1−k(G−λ0 I)−1 kHM kM −1 BkHM ≥ 1− kM −1 BkHM >0 d(λ0 , σ(G)) Therefore, (T − λ0 I)−1 which is the product of a bounded operator and a compact operator, is compact (cf. 8.3-2 Kreyszig [1989]). I Step 3. (Obtain Conclusion 1.): 3.a) Since (T − λ0 I)−1 is compact and T is closed, then (T − λI)−1 is compact for all λ ∈ ρ(T ) (see 5.7 Engel and Nagel [2006]). Therefore, σ(T ) is a point spectrum such that each λ ∈ σ(T ) = σp (T ) has finite multiplicity (cf. 8.3-1 Kreyszig [1989]). On the other hand, if λ is such that d(λ, σ(G)) > kM −1 BkHM then necessarily λ ∈ ρ(T ). Consequently, for λ ∈ σ(T ), we have d(λ, σ(G)) < kM −1 BkHM . In other 5 words, σ(T ) lies within the union of a set of circles of radii kM −1 BkHM and center σ(G) = σp (Ls , M ). 3.b) By a similar discussion made in Step 3. of the proof of theorem 3.1, it can be shown that σ(T ) = σp (M −1 L). For this, replace Ls by L + λM ; similarly, existence of G−1 is replaced by existence of (G + M −1 B + λI)−1 for λ mentioned in hypothesis 3. Since it is possible that G + M −1 B does not have an inverse, it is required that such λ exists. Hence, σp (M −1 L) lies within the union of a set of circles of radii kM −1 BkHM centered about σp (M −1 Ls ). I Step 4. Obtain Conclusion 2.: According to hypothesis 4.(c) we can choose a k0 such that for all integers k ≥ k0 , d(Ck , σp (Ls , M )) = d(Ck , σp (G)) > kM −1 BkHM . Then for such a k and λ ∈ Ck , we have k(T − λI)ϕkHM = k(G − λI)ϕ + M −1 BϕkHM ≥ k(G − λI)ϕkHM − kM −1 BϕkHM ≥ {d(λ, σp (Ls , M )) − kM −1 BkHM }kϕkHM Furthermore, k(T − λI)ϕkHM ≤ k(T − λI)kHM kϕkHM . Combining this with the latest inequality and then taking supremum of both sides results in, sup k(T − λI)−1 kHM ≤ λ∈Ck 1 d(λ, σp (Ls , M )) − kM −1 BkHM Clearly, by hypothesis 4.(c), limk→∞ supλ∈Ck k(T − λI)−1 kHM = 0. Therefore all the conditions of Naimark’s theorem are satisfied by T for circles Ck with k ≥ k0 . Consequently, the generalized eigenvectors of T , and hence those of (L, M ) span HM in the sense of (2.2) . 4 Application To show an application of theorem 3.2, consider the Orr-Sommerfeld equation rewritten in the form of equation (1.1); then λ = iαcRe Ls = (D − α ) , B = −iαRe[U (D2 − α2 ) − U 00 ] M = −(D2 − α2 ) 2 2 2 (4.1) (4.2) (4.3) Here, U (y) is the base flow profile and U 00 = d2 U/dy 2 . Moreover, α and Re are wavenumber and Reynolds number, respectively. In hydrodynamic temporal stability analysis we are interested in finding the phase speed c appearing in λ for a set of input parameters. In association with the completeness theorem mentioned above, consider H = L2 (Ω) = H 0 (Ω), HM = H01 (Ω), 6 where Ω = [0, 1] Then, through a detailed process, one can show that differential operators in the Orr-Sommerfeld equation satisfy all hypotheses of theorems 3.1 and 3.2. Here as a step of that process, the construction of concentric circles is presented. Consider σp (Ls , M ) = {µn } ∪ {ϑn } where the µn are the ordered positive eigenvalues of the problem Ls ϕ1 = µM ϕ1 , ϕ1 (0) = ϕ01 (0) = ϕ1 (1) = ϕ01 (1) = 0 (repeated eigenvalues, if any occurs, are listed once) and the numbers νn are the ordered positive eigenvalues of the problem M ϕ2 = ϑϕ2 , ϕ2 (0) = ϕ2 (1) = 0 For the operators in the Orr-Sommerfeld equation, it can be shown that as n → ∞, µn+1 − µn → ∞ and ϑn+1 − ϑn → ∞. Thus, the sequence {Ck }∞ k=1 is a set of concentric circles with center at the origin and with radii rk as follows. For n ≥ 1 consider the interval (µn , µn+1 ). Then, (a) if there is no ϑl ∈ (µn , µn+1 ), take rn = 21 (µn + µn+1 ); (b) if there is only one ϑl ∈ (µn , µn+1 ), then when ϑl ∈ (µn , 21 (µn +µn+1 )) take 3 1 rn = 14 µn + 34 µn+1 and when ϑl ∈ ( 21 (µn + µn+1 ), µn+1 ) take rP n = 4 µn + 4 µn+1 ; m 1 (c) if there are several (say m) ϑl in (µn , µn+1 ), take rn = m i=1 ϑi (DiPrima and Habetler [1969]). As a result of theorem 3.2, for a bounded flow, the spectrum of the OrrSommerfeld equation consists of an infinite number of discrete eigenvalues and the associated eigenvectors span HM = H01 (Ω) in the sense of (2.2). By this completeness, an arbitrary initial disturbance which satisfies the physical boundary conditions can be expanded in terms of these eigenfunctions. This technique is very useful in studying different hydrodynamic instability mechanisms. References R. C. DiPrima and G. J. Habetler. The completeness theorem for nonselfadjoint eigenvalue problems in hydrodynamic stability. Archieve for Rational Mechanics and Analysis, 34(3):218–227, 1969. P. G. Drazin and W. H. Reid. Hydrodynamic Stability. Cambridge Univ. Press, 1981. K.-J. Engel and R. Nagel. A Short Course on Operator Semigroups. Springer, 2006. 7 L. C. Evans. Partial Differential Equations. American Mathematical Society, 2nd edition, 2010. E. Kreyszig. Indtroductory Functional Analysis with Applications. John Wiley & Sons, 1989. S. G. Mikhlin. The Problem of the Minimum of a Quadrature Functional. Holden-Day INC., 1965. 8 The Metric Lax and Applications∗ Cheng Gong† January 16, 2015 Abstract We define consistency, convergence, and stability of numerical methods in the setting of metric spaces. The Lax equivalence theorem is formulated and a proof is given in this setting. We also indicate its use in some applications. 1 Introduction The Lax (or Lax-Richtmyer) equivalent theorem Lax and Richtmyer [1956] is the fundamental theorem in numerical analysis. It was established by Peter Lax and Robert D. Richtmyer in 1956. The theorem states that “a consistent finite difference method for a well-posed linear initial value problem is convergent if and only if it is stable”. The convergence of a numerical method is a measurement of the differences between the numerical solution and the analytic solution from the mathematical problem. In general, the analytic solutions are usually not accessible and the convergence is difficult to be observed directly. However, the consistency and stability can be evaluated from the properties of the numerical method, e.g. error estimate, Fourier analysis and Von Neumann analysis Gustafsson et al. [2013]. Thus, the Lax equivalent theorem provides a practical way to prove the convergence of a numerical method from stability. This principle can be formulated and proved in metric spaces. ∗ This essay is based on the notes from Stefan Engblom. stefan.engblom@it.uu.se Division of Scientific Computing, Department of Information Technology, Uppsala university, SE-751 05 Uppsala, Sweden. cheng.gong@it.uu.se † 1 2 Definition ˜ be two metric spaces and we define the Let X = (X, d) and Y = (Y, d) mathematical problem as: find x ∈ X, such that T x = y, (2.1) where T : X −→ Y is an operator defined by the problem and y ∈ Y is given. Definition 2.1. The mathematical problem in (2.1) is well-posed if T −1 exists and is continuous in some neighborhood containing y. Then, the solution of a well-posed problem with the same form as in (2.1) is given by the inverse operator of T as x = T −1 y. (2.2) The problem in (2.1) is approximated by a sequence of numerical problems: find xn ∈ X such that Tn xn = y, (2.3) where Tn : X −→ Y is an operator of the corresponding numerical method and y ∈ Y is the same right-hand-side as in (2.1). The family of equations in (2.3) is called the numerical approximations for the problem in (2.1) [Hunter and Nachtergaele, 2001] and n refers to the indices of the sequence such that Tn −→ T as n −→ ∞. Definition 2.2. Let D(T ) be the domain of T . The numerical method in (2.3) is consistent, if ∀x ∈ D(T ), where D(T ) ⊂ X is the domain of T in X, then Tn x −→ T x as n −→ ∞. Definition 2.3. The numerical method in (2.3) is stable if ∀n ∈ N, Tn−1 exists and is continuous in some neighborhood containing y. Definition 2.4. The numerical method in (2.3) is convergent if xn −→ x as n −→ ∞. For a stable numerical method, the solution is given as xn = Tn−1 y by the inverse operator of Tn−1 . If the numerical method is convergent, the numerical solution will converge to the analytic solution of the mathematical problem. 2 3 The Lax Equivalent Theorem A general form of Lax equivalent theorem is given in Theorem 3.1. Theorem 3.1. A consistent method applied to a well-posed problem is convergent if and only if it is stable. By introducing the definitions from Section 2, the Lax equivalent theorem is formulated in Theorem 3.2. Theorem 3.2. Define a mathematical problem as (2.1) and the numerical approximation as (2.3). Suppose T −1 is continuous and Tn x −→ T x for any x in the domain of T . Then, Tn−1 y −→ x as n −→ ∞ if and only if Tn−1 is continuous for all n. Proof. [Engblom, 2014] ‘=⇒’ By the consistency of Tn , for any given δ, ∃N ∈ N such that, for all n > N ˜ n x, T x) < δ. d(T (3.1) From the assumption in (2.1) and (2.3), T x = y = Tn xn , (3.2) together with the well-posedness of the problem, ˜ n x, Tn xn ) < δ. d(T (3.3) By the stability of the numerical method, Tn−1 is continuous at y for all n. Denote y1 = Tn x, then ∀ > 0 and for all y1 satisfying (3.3), ∃δ > 0, such that d(Tn−1 y1 , Tn−1 y) < , (3.4) which is equivalent to d(x, xn ) < . (3.5) Therefore, xn −→ x as n −→ ∞ and the numerical method is convergent. ‘⇐=’ If the numerical method in (2.3) is not stable, then ∃ > 0, for ∀δ > 0, ˜ 1 , y) < δ, we have ∀y1 ∈ Y satisfying d(y d(Tn−1 y1 , Tn−1 y) > . (3.6) From the consistency, there exists an N ∈ N, ∀n > N , ˜ n x, T x) < δ. d(T 3 (3.7) If we take y1 = Tn x, then (3.6) is written as d(Tn−1 Tn x, Tn−1 Tn xn ) > , (3.8) d(x, xn ) > , (3.9) by (3.2) and consider the well-posedness of the problem that T −1 exists in some neighborhood of y, (3.10) d(T −1 y, Tn−1 y) > , which indicates that the numerical method is not convergent. Therefore, a consistent and convergent numerical method is stable. 4 Applications In a metric space, the metric can be used to define convergence, continuity and compactness. In section 3, the metric form of Lax equivalent theorem is shown in Theorem 3.2. Usually, this theorem is applied to the problems defined in normed spaces, or even in Banach spaces, to prove the numerical method is convergent by using the stability. There are many linear and nonlinear examples for this forward direction of Lax equivalent theorem, e.g. [Sanz-Serna and Palencia, 1985, Palencia and Sanz-Serna, 1984]. On the other hand, to get stability from convergence is not very practical in some finite difference method since the convergence always relates to the analytic solution of the problem which is not accessible in many cases. However, this direction of the Lax equivalent theorem can be use to show a consistent numerical method does not converge from the condition that it is unstable. An interesting example is to show that Newton-Cotes formulas for numerical integration do not converge [Engels, 1980] by knowing that Newton-Cotes formulas are numerically unstable [Krommer and Ueberhuber, 1998]. The importance of the Lax equivalent theorem is that the convergence of a numerical method is difficult to be established while it is usually desired. The consistency is straightforward to be verified and the stability which concerns the ‘local’ properties of numerical method is much easier to show than the convergence. Moreover, the metric Lax equivalent theorem only requires the operator for the numerical method to be defined on a metric space which may provide a wider use than the classical one for numerical analysis. However, the operators Tn and T are defined on the same metric space which could cause some troubles for the applications in the future. 4 References S. Engblom. Notes of lax equivalent theorem, 2014. H. Engels. Numerical quadrature and cubature. 1980. B. Gustafsson, H.-O. Kreiss, and J. Oliger. Time-dependent problems and difference methods, volume 121. John Wiley & Sons, 2013. J. K. Hunter and B. Nachtergaele. Applied analysis. World Scientific, 2001. A. R. Krommer and C. W. Ueberhuber. Computational integration, volume 53. Siam, 1998. P. D. Lax and R. D. Richtmyer. Survey of the stability of linear finite difference equations. Communications on Pure and Applied Mathematics, 9(2): 267–293, 1956. C. Palencia and J. Sanz-Serna. An extension of the lax-richtmyer theory. Numerische Mathematik, 44(2):279–283, 1984. J. Sanz-Serna and C. Palencia. A general equivalence theorem in the theory of discretization methods. Mathematics of computation, 45(171):143–152, 1985. 5 Fixed-point proof of Lax-Milgram theorem Fatemeh Ghasemi∗ January 16, 2015 Abstract Lax-Milgram theorem gives conditions under which a bilinear form can be inverted to show existence and uniqueness of weak solution of a given partial differential equation (PDE). There are different ways for proving this theorem. Here we use the well-known Banach fixed-point theorem. 1 Introduction Riesz representation theorem is of the fundamental theorem about Hilbert space that is the basis for the existence and uniqueness for (symmetric) variational problems. For using this theorem, the variational problems needs to be symmetric. But a wide variety of variational problems are not symmetric and this means that this theorem is not applicable. Fortunately, there is an analogous result, the Lax-Milgram theorem, that does apply to nonsymmetric problems. The Lax-Milgram lemma has played a critical role over the last half century by establishing existence and uniqueness of weak solutions of operator equations Au = f where A is a continuous and coercive linear operator from a Hilbert space H into its topological dual H 0 . The result has had far-reaching impact on the analytic and numerical study of elliptic and parabolic partial differential equations; boundedness and coercivity of the weak operators are all that need be verified. While in 1954 the paper (1) of Lax and Milgram provided a nonconstructive proof of the result, constructive proofs appeared over the next several ∗ Division of Computational Mathematics, Department of Mathematics, Link¨ oping University, SE-581 83 Link¨oping, Sweden . fatemeh.ghasemi@liu.se 1 years. In 1960, Zaran- tanello provided a generalization of Lax-Milgram to strongly monotone nonlinear operators via Banach fixed points. A similar fixed point proof in the linear case was given by Lions and Stampacchia in a 1967 paper (2) on variational inequalities. In both cases, the mapping shown to be contractive involves applying the Riesz map for H to a residual Av − f ∈ H for some v ∈ H. Also in 1967, Petryshyn gave a constructive proof of the result based on so-called upper and lower semi-orthogonal bases for subspaces of H. In this essay, after introducing some preliminaries in the first section, we prove Lax-Milgram theorem by using Banach fixed-point theorem. 2 Preliminaries Let H denote a real Hilbert space equipped with inner product (., .)H and 0 0 associated norm k.k. H denotes the topological dual with h., .i the H × H duality pairing. In this essay we use Banach fixed-point theorem and Riesz Representation theorem in order to prove Lax-Milgram Theorem. In the following we review these two theorems. Theorem 2.1. (Banach fixed-point theorem) Let (X, d) be a non-empty complete metric space, and T : X → X a map such that d(T x, T y) ≤ kd(x, y) for some real positive k < 1. Then T has a unique fixed point, i.e. there exists a unique x ∈ X such that T x = x; furthermore for any initial x0 , the iterates T n x0 converge to the fixed point x. Theorem 2.2. (Riesz Representation theorem) For every bounded linear functional f ∈ H there exists a unique element u ∈ H such that f (v) = (u, v), 3 ∀v ∈ H. Lax-Milgram Theorem Theorem 3.1. Suppose that B : H × H → R is a bilinear form that is 1. bounded, i.e. there exists an α > 0 such that |B(u, v)| ≤ αkukkvk, and 2 ∀u, v ∈ H, 2. coercive, i.e. there exists a β such that B(u, u) ≥ βkuk2 , ∀u ∈ H. 0 Then for any f ∈ H there exists a unique uf ∈ H such that B(uf , v) = f (v), ∀v ∈ H. (3.1) Furthermore kuf k ≤ β −1 kf kH 0 , (3.2) where kf kH 0 = sup v∈H,kvk=1 |hf, vi|. in particular uf depends continuously on f , i.e. kuf − ug k ≤ β −1 kf − gk. Proof. The uniqueness is straightforward. Suppose there exist two solutions u, u¯ then B(u, v) = B(¯ u, v) = f (v), ∀v ∈ H. Since B is bilinear, we have B(u − u¯, v) = 0 for all v ∈ H. Let v = u − u¯. Using the coercivity of B leads to βku − u¯k2 ≤ B(u − u¯, u − u¯) = 0, i.e. u = u¯. Let v = uf in (3.1) to get B(uf , uf ) = f (uf ) By definition of k.kH 0 we know that f (uf ) ≤ kf kH 0 . By using this inequality and coercivity of B, the bound (3.2) is derived. The continuity result follows by considering B(uf , v) − B(ug , v) = B(uf − ug , v) = (f − g, v), and setting v = uf − ug . So only existence requires any work. Fix u ∈ H and consider the map v → B(u, v). We claim that this defines a bounded linear functional on H. Since B is bilinear and bounded, it is clearly linear and bounded. It follows from the Riesz Representation Theorem that there exists a w ∈ H such that (w, u) = B(u, v), 3 ∀v ∈ H. We define Au = w. By this definition we have (Au, u) = B(u, v), ∀v ∈ H. and claim that this definition yields a bounded linear operator from H into itself. Indeed, for every v ∈ H (A(α1 u1 + α2 u2 ), v) =B(α1 u1 + α2 u2 , v) =α1 B(u1 , v) + α2 B(u2 , v) =α1 (Au1 , v) + α2 (Au2 , v) =(α1 Au1 + α2 Au2 , v) Since this holds for every v ∈ H, it follows that A(α1 u1 + α2 u2 ) = α1 Au1 + α2 Au2 , and this implies that A is linear. To show that A is bounded, note that kAuk2 = (Au, Au) = B(u, Au) ≤ αkukkAuk, hence kAuk ≤ αkuk and A is bounded. Using the Riesz Representation, we know that there exists a ϕ ∈ H such that (ϕ, v) = f (v), ∀v ∈ H. We can rewrite our equation in the following way (Au, v) = (ϕ, v), ∀v ∈ H. This means that u satisfies (3.1) if and only if Au = ϕ. Now we have to show that this equation has a solution. In this step of proof we use Banach fixed-point theorem. Clearly, for any λ > 0 Au = ϕ ⇔ u = u − (Au − ϕ). Let T u = u − λ(Au − ϕ). So we are investigating a fixed point for T. Since H is a Hilbert space, it is enough to show that T is a contraction. We have kT u − T vk2 =k(u − v) − λA(u − v)k2 =ku − vk2 − 2λ(A(u − v), u − v) + λ2 kA(u − v)k2 =ku − vk2 − 2λB(u − v, u − v) + λ2 kA(u − v)k2 ≤ku − vk2 − 2λβku − vk2 + λ2 α2 ku − vk2 =(1 − 2λβ + λ2 α2 )ku − vk2 , where we have used the coercivity of B and the fact that A is bounded. It follows that if we choose sufficiently small λ, T is a contraction. It therefore has a unique fixed point, which provides our solution. 4 References [1] P. D. Lax and A. N. Milgram, Parabolic equations , in Contributions to the theory of partial differential equations, Annals of Mathematics Studies, no. 33, Princeton University Press, Princeton, N. J., 1954, pp. 167-190. [2] J.-L. Lions and G. Stampacchia, Variational inequalities ,Comm. Pure Appl. Math. Comm. Pure Appl. Math., 20 (1967), pp. 493-519. 5 Part II Solutions to exercises 87 Chapter 1 Metric spaces 88 Ex. 1.2.4 Consider the sequence 1 , n ≥ 2. log (n) Clearly this sequence converges to zero. However, log (n) grows slower than any polynomial sequence so for each 1 ≤ p < ∞ there exists an Np ∈ N such that ξnp ≥ 1/n for all n ≥ Np . Therefore, since the harmonic series diverges, the sequence ξn is not in any space lp . ξn = Ex. 1.3.8 Show that the closure B(x0 ; r) of an open all B(x0 ; r) in a metric space can differ from the ˜ 0 ; r). closed ball B(x Solution: Let x0 = (0, 0) and r = 1. Define a metric space (X, d), where X = {(x1 , x2 ) | x21 + x22 ≤ 1, x1 and x2 are real and nonnegative} ∪ {(x1 , x2 ) = (cos(1.25π), sin(1.25π))}, and d is the Euclidean distance. The points in set X are shown in Figure 1.1. According 1.5 1 0.5 0 −0.5 −1 −1.5 −1.5 −1 −0.5 0 0.5 1 1.5 Figure 1.1: Points in X to the definitions, B(x0 ; r) = {(x1 , x2 ) ∈ X | x21 + x22 < 1}, ˜ 0 ; r) = {x ∈ X | d(x, x0 ) ≤ 1} = X, B(x B(x0 ; r) = {(x1 , x2 ) ∈ X | (x1 , x2 ) 6= (cos(1.25π), sin(1.25π))}. 89 ˜ 0 ; r). Therefore, B(x0 ; r) can differ from B(x Ex. 1.3.8 Let X be a set with at least two elements and let the metric space (X, d) have the discrete metric ( 0 x=y d(x, y) = . 1 x 6= y ˜ 0 , 1) be the closed ball around x0 ∈ X Let B(x ˜ 0 , 1) = {x ∈ X : d(x, x0 ) ≤ 1}. B(x By the definition of the metric this means ˜ 0 , 1) = X. B(x Let now B(x0 , 1) = {x ∈ X : d(x, x0 ) < 1} be the open ball around x0 . Which means B(x0 , 1) = {x0 }. Let now x˜ ∈ X be an accumulation point of B(x0 , 1). This means that d(˜ x, x0 ) < , ∀ > 0 which implies x˜ = x0 thus the closure of the open ball is B(x0 , 1) = {x0 }, which is not equal to the closed ball. Ex. 1.3.8 Consider the discrete metric on R2 , X = (R2 , d) where B(x0 ; 1) = X and B(x0 ; 1) = {x0 }, clearly differing. Ex. 1.3.12 Let us consider the family of characteristic functions X = x (t) = χ(a,c) (t) , c ∈ (a, b) 90 that is non–countable because (a, b) is a non–countable set. Notice that d (x, y) = 1, x 6= y, x, y ∈ X. Now, let us assume that there exists a dense set in B [a, b], call it G: then there exist as 1 many non–empty balls B x0 , 2 , x0 ∈ X as the elements of X. On the other hand, we know that 1 1 0 0 B x0 , ∩ B x0 , = ∅, x0 6= x0 2 2 and therefore G is non–countable as well. Thus, it does not exist a countable dense set in B [a, b], since we have proved that every dense set in X is non–countable. Ex. 1.3.12 Consider the subset M ⊂ B[a, b] consisting of the functions δx = 1 at x and zero everywhere else on [a, b]. Then M is uncountable, and sup |δx − δy | = 1 whenever x 6= y. Consider now all open balls B(δx ; 1/2). These are disjoint, and so if a subset S ⊂ B[a, b] is dense, then it must intersect each ball in a different point. This renders S uncountable. Since S is general this shows that B[a, b] is not separable. Ex. 1.4.2 If xnk → x, then for every ε > 0 there is an N1 = N1 (ε) such that d (xnk , x) < ε 2 ∀nk > N1 and, since xn is Cauchy we know that there is an N2 = N2 (ε) such that d (xn , xm ) < ε 2 ∀n, m > N2 . Hence by the triangular inequality we obtain for n, nk > max {N1 , N2 } d (xn , x) ≤ d (xn , xnk ) + d (xnk , x) < and the claim follows. 91 ε ε + =ε 2 2 Ex. 1.4.2 For the subsequence we have d(xnk , x) −→ 0, and since (xn ) is Cauchy, Using the triangle inequality we get d(xn , xnk ) −→ 0. d(xn , x) ≤ d(xn , xnk ) + d(xnk , x) −→ 0 which proves that (xn ) is convergent with the limit x. Ex. 1.4.8 If d1 and d2 are metrics on the same set X and there are positive numbers a and b such that for all x, y ∈ X, ad1 (x, y) ≤ d2 (x, y) ≤ bd1 (x, y), (1.0.1) show that the Cauchy sequences in (X, d1 ) and (X, d2 ) are the same. Solution: We use the relation (1.0.1) to show (xn ) is Cauchy in (X, d1 ) ⇐⇒ (xn ) is Cauchy in (X, d2 ) . First assume that (xn ) is Cauchy in (X, d1 ). That is, for any ε > 0 there is an N such that d1 (xn , xm ) < ε b ∀m, n > N. Using (1.0.1) we obtain d2 (xn , xm ) ≤ bd1 (xn , xm ) < ε ∀m, n > N. which proves that (xn ) is Cauchy also in (X, d2 ). Now assume that (xn ) is Cauchy in (X, d2 ). That is, for any ε > 0 there is an N such that d2 (xn , xm ) < aε ∀m, n > N. Using (1.0.1) we obtain 1 d1 (xn , xm ) ≤ d2 (xn , xm ) < ε ∀m, n > N. a which proves that (xn ) is Cauchy also in (X, d1 ). Thus, (xn ) is Cauchy in (X, d1 ) iff (xn ) is Cauchy in (X, d2 ), i.e., the Cauchy sequences in (X, d1 ) and (X, d2 ) are the same. 92 Ex. 1.5.6 Show that the set of all real numbers constitutes an incomplete metric space if we choose d(x, y) = |arctan x − arctan y|. Solution: Let X = R. We show that (X, d) is incomplete by constructing a Cauchy sequence in X which does not converge. Consider the sequence (xn ), where xn = n. Clearly, (xn ) does not converge. Since π , 2 lim arctan x = x→∞ for any ε > 0 we can find N = N (ε) such that π ε arctan n − < 2 2 ∀n > N. Hence, π π d(xn , xm ) = |arctan xn − arctan xm | = arctan n − − arctan m + 2 2 π π ε ε ≤ arctan n − + arctan m − < + = ε ∀n, m > N 2 2 2 2 which shows that (xn ) is Cauchy. Ex. 1.5.8 Let y1 , y2 ∈ Y . Since Y is a subspace it is enough to show that Y is closed. Let {yn }∞ n=1 be a sequence in Y with yn → y ∈ C[a, b], this means ∀ > 0 ∃N : sup |y(t) − yn (t)| < |. t∈[a,b] But by definition |y(a) − yn (a)| ≤ sup |y(t) − yn (t)| t∈[a,b] so we get and by the same argument also yn (a) → y(a), yn (b) → y(b). But since each yn ∈ Y we have yn (a) = yn (b) so we get |y(a) − y(b)| ≤ |y(a) − yn (a)| + |y(b) − yn (b)| → 0 which shows that y(a) = y(b). This means that y ∈ Y and thus is a closed subspace and complete. 93 Ex. 1.5.9 Assume: that (xm ) is the sequence of continuous functions on [a, b] that uniformly converges on [a, b]; Prove: that the limit function x is continuous on [a, b]. Solution: Since (xm (t)) is converging to x, ∀ > 0 ∃N (), ∀t ∈ [a, b], |xm (t) − x(t)| < 3 (1.0.2) On the other hand, (xm (t)) is continuous at any to ∈ [a, b]; so, we have, ∀ > 0 ∃δ > 0, |t − to | < δ ⇒ |xm (t) − xm (to )| < 3 (1.0.3) Now, we want to show that x(t) is continuous on [a, b]; therefore, ∀ > 0 ∃δ1 > 0, |t − to | < δ1 ⇒ |x(t) − x(to )| < (1.0.4) For this purpose, we can find the following relation, using (1) and (2): |x(t) − x(to )| = |x(t) ± xm (t) ± xm (to )| = = |(x(t) − xm (t)) + (xm (t) − xm (to )) + (xm (to ) − x(to ))| ≤ |x(t) − xm (t)| + |xm (t) − xm (to )| + |xm (to ) − x(to )| + + = ≤ 3 3 3 which is equivalent to (3). Ex. 1.6.14 The only question here is whether d(x, y) = 0 ⇒ x = y or not. i) Let d(x, y) = 0 and let h(t) = x(t) − y(t) which implies Z a b |h(t)|dt = 0. Assume that there exist t0 ∈ [a, b] such that |h(t0 )| = 6 0. Since x and y are continuous h is continuous. So let 0 < = |h(t0 )| and let δ be so small that if |t − t0 | < δ we have |h(t) − h(t0 )| < . 94 The reverse triangle inequality gives us: |h(t0 )| − |h(t)| ≤ |h(t) − h(t0 )| < . Since = |h(t0 )| this now implies |h(t)| > 0, ∀t ∈ [t0 − δ, t0 + δ]. But since: Z a this implies b |h(t)|dt ≥ Z a Z t0 +δ t0 −δ |h(t)|dt b |h(t)|dt > 0 which is a contradiction. So no t0 with |h(t0 )| = 6 0 can exist, thus x(t) = y(t). ii) For Riemann-integrable functions d(x, y) = 0 does not imply x = y. An example is: let tm = (a + b)/2 and ( 0 t < tm x(t) = 1 t ≥ tm ( 0 t ≤ tm y(t) = 1 t > tm We have x(tm ) 6= y(tm ) but d(x, y) = 0. 95 Chapter 2 Normed spaces 96 Ex. 2.3.10 Suppose that the Banach space (X, k · k) has a Schauder basis {en }∞ n=1 so that for any x ∈ X there is a unique sequence of scalars, α1 , α2 , . . . such that x= ∞ X αk ek . k=1 We will for simplicity assume that αk ∈ R, though generalisation to complex numbers is straight forward. Given any > 0 there is an N such that kx − N X αk ek k < . 2 k=1 Now, recall that Q is dense in R, meaning that for any αk ∈ R we can find a βk ∈ Q such that |αk − βk | < /2k+1 . By the triangle inequality we have N X N X N X N X N X kx − βk ek k ≤ kx − αk e k k + k αk ek − βk ek k < + < . 2 k=1 2k+1 k=1 k=1 k=1 k=1 Thus the Schauder basis constitutes a basis for a dense subset of X, which is countable since βk ∈ Q for all k. Consequently, X is separable. Ex. 2.3.10 Let Y be the set, N X Y ={ αj ej : αj ∈ R, N ∈ N+ }. j=1 Let x ∈ X, since we have a Schauder-basis there exist {αj }∞ j=1 such that kx − xn k → 0 where xn := n X αj ej . j=0 Thus for every x ∈ X we have a sequence in Y converging to x. This shows that Y¯ = X. 97 Let now instead xn = set Pn j=1 αj ej ∈ Y be an arbitrary element in Y and let M be the N X M ={ βj ej : bj ∈ Q, N ∈ N+ }, j=1 k which is countable. Since Q is dense in R there exist a sequence {βjk }∞ k=1 with βj ∈ Q such that |αj − βjk | ≤ , ∀k ≥ K, Cn where C = max kej k . j∈{1,...,n}] Let xkn = Pn j=1 βjk ej ∈ M , we get: n n X X xn − xkn = (αj − βjk )ej |αj − βjk | kej k , ≤ j=1 j=1 so for k ≥ K we get n n X X xn − xkn ≤ kej k ≤ 1 ≤ . Cn n j=1 j=1 Showing that xkn → xn when k → ∞. Thus for every xn ∈ Y we can find a sequence in M ¯ = Y and in total converging to xn . So M ¯) = M ¯ X = Y¯ = (M showing that M is dense in X so that X is separable. Ex. 2.4.4 Show that equivalent norms on a vector space X induce the same topology for X. Solution: Assume that k·k1 and k·k2 are equivalent norms on X. Let T1 be the topology for X induced by k·k1 and let T2 be the topology for X induced by k·k2 . In other words, T1 and T2 are the collections of all open subsets of X in the sense of k·k1 and k·k2 , respectively. We show that T1 = T2 by showing that any subset of X is open in the sense of k·k1 iff it is open in the sense of k·k2 . Since k·k1 and k·k2 are equivalent, we have a k·k1 ≤ k·k2 ≤ b k·k1 (2.0.1) for some a, b > 0. Let Bi (x; r) denote the open ball of radius r around x in the sense of k·ki . Now assume that M is an open subset of X in the sense of k·k1 . By the definition of an open set, M contains an open ball around each of its points, i.e., for any x ∈ M ∃ε > 0 98 such that B1 (x; ε) ⊂ M . Because of (2.0.1) we have B2 (x; aε) ⊂ B1 (x; ε), which means that M contains an open ball in the sense of k·k2 around each of its points. This proves that any M ⊂ X is open in the sense of k·k2 if it is open in the sense of k·k1 . By reversing the above argument, we can show that any M ⊂ X is open in the sense of k·k1 if it is open in the sense of k·k2 . Ex. 2.4.4 Suppose that the norms kk0 and kk1 are equivalent. Then there are constants c1 and c2 with 0 ≤ c1 ≤ c2 such that c1 k x k0 ≤k x k1 ≤ c2 k x k0 for all x ∈ X. Let U be a subset of X which is open with respect to the topology on X induced by the norm kk1 . For u ∈ U there is an such that {x ∈ X k u − x k1 < } ⊂ U . But since the norms are equivalent {x ∈ X k u − x k1 < } ⊂ {x ∈ X k u − x k0 < /c2 }. This means that U is open with respect to the topology induced by norm kk0 . By the same argument, we can show that any open subset of X with respect to the topology induced by the norm kk0 is open with respect to the topology induced by kk1 . Ex. 2.5.4 Show that for an infinite subset M in the space s to be compact, it is necessary that there are numbers γ1 , γ2 , . . . such that for all x = (ξk (x)) ∈ M we have |ξk (x)| ≤ γk . Solution: The metric on the space s is d(x, y) = ∞ X 1 |ξk − ηk | 2k 1 + |ξk − ηk | k=1 where x = (ξj ), y = (ηj ). Assume that not all (ξk ) in M are bounded by |ξk | ≤ γk . Then there is at least one k0 such that |ξk0| is unbounded. Now consider a sequence of elements (m) (m) in M , x1 , x2 , . . ., where xm = ξk . We choose the sequence such that ξk0 = m. We then have, for m 6= n, ∞ (m) (n) X 1 |m − n| 1 1 1 |ξk − ξk | ≥ k0 ≥ k0 . |xm − xn | = (m) (n) k 2 1 + |ξk − ξk | 2 1 + |m − n| 2 2 k=1 Clearly (xm ) can not have a convergent subsequence. Thus, M is not compact. Ex. 2.5.10 Let X and Y be metric spaces, X compact, and T : X → Y bijective and continuous. Show that T is a homeomorphism. 99 In the proof, I use the result from Prob. 14 in Section 1.3: a mapping T : X → Y is continuous if and only if the inverse image of any closed set M ⊂ Y is a closed set in X. This result is equivalent to a mapping T −1 : Y → X is continuous if and only if for any closed set M ⊂ X, T (M ) is a closed set in Y . (*) Let M be a closed subset in X. Since X is a compact metric space, M is compact (Prob. 2.5-9). Since T is continuous, according to Theorem 2.5-6 we have that T (M ) is compact. By using Lemma 2.5-2, we have that T (M ) is closed and bounded. Since M is arbitrary and T is bijective, by (*) we have that T −1 is continuous. Therefore, T is a homeomorphism. Ex. 2.5.10 Let X and Y be metric spaces, X be compact, T : X → Y bijective and continuous and F = T −1 . • Let V ⊆ X and assume that V is closed and bounded so that it is compact. • Since T is continuous, it follows that T (V ) is compact. • This implies that T (V ) is closed and bounded. • But T (V ) = F −1 (V ) so F −1 (V ) is closed and bounded. Hence F is continuous, and thus T is a homeomorphism. Ex. 2.5.10 We study the continuous and bijective operator T : X → Y where X is compact. By theorem 2.5-6, Y is compact. Lets assume that T −1 is discontinuous, that is there is a y ∈ Y and x ∈ X such that dy (yn , y) → 0 dx (xn , x) > δ > 0 (2.0.2) (2.0.3) where the sequences xn , yn satisfy T xn = yn and T x = y. However, X is compact, so there is a subsequence of xn such that xk → x0 6= x. We have dy (T x0 , T x) ≤ dy (T xk , T x) + dy (T x0 , T xk ) = dy (yk , y) + dy (y0 , yk ). Since T is continuous, we have dy (y0 , yk ) = dy (T x0 , T xk ) → 0 as k → ∞. Therefore, we have dy (T x0 , T x) ≤ dy (yk , y) + dy (y0 , yk ) → 0 100 as k → ∞. However, according to (8) we should have for bijective T dy (T x, T x0 ) > > 0, which is a contradiction. T −1 must therefore be continuous and T is then a homeomorphism. Ex. 2.5.10 A Homeomorphism is a continuous bijective mapping whose inverse is continuous. Since we have a continuous and bijective mapping T : X −→ Y (X and Y metric spaces, X compact), we need to show that the inverse T −1 is continuous. Generally, a mapping T : X −→ Y is continuous if and only if, for any closed set K ⊂ Y , the inverse image T −1 (K) is a closed set in X (cf. Prob. 14, Sec. 1.3). So here, to show that T −1 : Y −→ X is continuous, we want to show that T (M ) is a closed set in Y , for any closed set M ⊂ X. 1. Let M be a closed subset in X-compact, then from Prob. 9, Sec. 2.5, M is also compact. 2. Since T is continuous, and M is compact ⇒ T (M ) is compact (Theorem 2.5-6) 3. Since T (M ) is a compact subspace of the metric space Y , T (M ) is closed and bounded (Lemma 2.5-2) Ex. 2.7.2 • Consider T bounded. If there exists M > 0 such that kxk ≤ M , then kT xk ≤ c kxk = cM and therefore T maps bounded sets in X into bounded sets in Y . • Let us consider T which maps bounded sets in X into bounded sets in Y . If x = 0, then T is trivially bounded since T (0) = T (0R z) = 0R T (z) = 0 ∀z ∈ X ⇒ kT 0k = 0. Let us consider x 6= 0. We know that exists C > 0 such that kT xk ≤ C, ∀x 6= 0 such that kxk = 1. Therefore x T ≤C kxk and for the homogeneity of the norm we have kT xk ≤ C kxk. 101 Ex. 2.7.2 Let X and Y be normed spaces. A linear operator T : X → Y is bounded iff T maps bounded sets in X into bounded sets in Y . If T is bounded: We want to show that for any bounded set M ⊂ D(T ) s.t. for any x, y ∈ M we have that kx − yk ≤ r < ∞ than also kT x − T yk < ∞. Since D(T ) is a vector space then x, y ∈ M ⊂ D(T ) → x − y = z ∈ D(T ). Furthermore kT zk ≤ ckzk since T is bounded. Hence kT x − T yk = kT (x − y)k = kT zk ≤ ckzk < ∞. If T maps bounded sets in X into bounded sets in Y : We know that for any x, y ∈ M bounded than kT x − T yk < ∞. Then we call δ(M ) = sup kx − yk and δ(T (M )) = x,y∈M sup T x,T y∈T (M ) kT x − T yk. For any zD(T ), z 6= 0 we define x=y+ δ(M ) z kzk ⇒ kx − yk = δ(M ). Then kT x − T yk = kT (x − y)k = kT which means that kT zk = kT x − T yk where c = δ(M ) δ(M ) z k= kT zk kzk kzk δ(T (M )) kzk ≤ kzk = ckzk. δ(M ) δ(M ) δ(T (M )) . δ(M ) Ex. 2.7.2 A linear operator T : D(T ) −→ Y , (where D(T ) ⊂ X and X and Y are normed spaces), is bounded if there is a real number c such that for all x ∈ D(T ), kT xk ≤ ckxk. (2.0.4) Further, a subset K in a normed space X is bounded if and only if there is a positive number a such that kxk ≤ a (2.0.5) for every x ∈ K. Let M be a bounded set in X. 102 1. Show that if T is bounded ⇒ T (M ) is a bounded set in Y : Let x ∈ M and T x ∈ T (M ), then from (2.0.4) and (2.0.5) we have kT xk ≤ ckxk ≤ ca (2.0.6) where ca is a positive number. Since x ∈ M is arbitrary this proves T (M ) is a bounded set in Y (according to the definition (2.0.5) of a bounded subset). 2. Show that if T (M ) is a bounded set in Y ⇒ T is bounded: Let x ∈ M and T x ∈ T (M ), then using the definition for the operator norm we have kT kkxk = kT xk kT xk kxk ≥ kxk = kT xk, kxk x∈M,x6=0 kxk sup x 6= 0. (2.0.7) Since kT k is a real number, call it c, we have kT xk ≤ ckxk, x 6= 0. (2.0.8) For the case when x = 0, since for linear operators T 0 = 0, we have kT xk = kT 0k = 0 ≤ ckxk, x = 0. (2.0.9) Since inequality (2.0.4) holds for all x ∈ M , T is bounded. Ex. 2.7.7 Given: T : X → Y is bounded and linear.Furthermore, ∃b > 0 3 ∀x ∈ X, kT xk ≥ bkxk ⇒ kxk ≤ kT xk b (2.0.10) We want to show that T −1 : Y → X exists and is bounded. According to [Theorem 2.6-10(a)], T −1 exists if and only if T x = 0 implies x = 0. In our case, T x = 0 ⇒ 0 ≥ bkxk Since b > 0 the only way for the latter inequality to be valid is that, x = 0. Hence, T −1 exists. Also, based on the definition of mapping we have, ∀x ∈ X ∃y ∈ Y 3 y = T x ∧ x = T −1 y kT xk kyk ⇒ kT −1 yk = kxk ≤(1) = b b 1 ⇒ kT −1 yk ≤ kyk b therefore, according to [Definition 2.7-1], T −1 is bounded. 103 (2.0.11) (2.0.12) (2.0.13) Ex. 2.7.8 If we consider T : l∞ → l∞ , x 7→ y defined by x = {αj } , y = {βj } , s.t. βj = αj j the bounded set y = {1} ⊂ R (T ) is mapped by T −1 into x = {j}j∈N , that is not bounded. Ex. 2.7.8 Let X be a normed space with points x = {ξk } being sequences of real or complex numbers that has only finitely many non-zero terms. Further let kxk = supk |ξk |. Now define the operator T : X → X as T x = (ξ1 , ξ2 /2, . . . ) = {ξk /k}. If y = {ηk } and α and β are numbers then T (αx + βy) = {(αξk + βηk )/k} = α{ξk } + β{ηk } = αT (x) + βT (y) so T is a linear operator. Also note that kT xk = sup |ξk |/k ≤ kxk sup 1/k ≤ kxk k k so T is bounded. Now, note that T −1 x = {kξk }. Consider the sequence x = {δk,n } ∈ X. Then kxk = 1 and kT −1 xk = n. Since n is an arbitrary integer and x and infinite sequence, we may let n → ∞, which shows that T −1 is not bounded. Ex. 2.8.6 The space C 1 [a, b] is the normed space of all continuously differentiable functions on J = [a, b] with norm defined by kxk = max |x(t)| + max |x0 (t)|. t∈J (2.0.14) t∈J (N1) Since |x(t)| ≥ 0 and |x0 (t)| ≥ 0, kxk ≥ 0. (N2) ”⇒”: kxk = 0 ⇔ max |x(t)| = − max |x0 (t)|⇔ x(t) = 0 and x0 (t) = 0. t∈J t∈J ”⇐”: x = 0 ⇒ kxk = max |0| + max |0| = 0 t∈J t∈J (N3) kαxk = max |αx(t)| + max |αx0 (t)| = |α| max |x(t)| + |α| max |x0 (t)| = |αk|xk t∈J t∈J t∈J t∈J (N4) kx + yk = {differentiation is linear} = max |x(t) + y(t)| + max |x0 (t) + y 0 (t)| t∈J t∈J ≤ max |x(t)| + max |y(t)| + max |x0 (t)| + max |y 0 (t)| = kxk + kyk. t∈J t∈J t∈J t∈J 104 Define f (x) = x0 (c), c = (a + b)/2. This functional on C 1 [a, b] is bounded since a + b a + b a + b 0 0 kxk = max |x(t)| + max |x0 (t)| ≥ x + x ≥ x t∈J t∈J 2 2 2 ⇔ |f (x)| ≤ kxk. (2.0.15) Further, f is linear since f (αx + βy) = {differentiation is linear} = αx0 = αf (x) + βf (y). a + b 2 + βx0 a + b 2 (2.0.16) If f is now considered as a functional on the subspace of C[a, b] which consists of all continuously differentiable functions, we want to show that f is not bounded. We consider the sequence of continuous and differentiable functions (xε ) = arctan( εt ), where ε is a real number. If we let ε −→ 0, then for c = 0, lim x0ε (c) = ∞, ε−→0 see Figure 2.1. So, given a constant C, there exists an ε such that |f (xε )| > Ckxε k, which proves that f is not bounded (since kxε k = max | arctan( εt )| ≤ t∈J ε). π 2 independent of Ex. 2.8.6 The space C 1 [a, b] is the normed space of all continuously differentiable functions on J = [a, b] with norm defined by kxk = max |x(t)| + max |x0 (t)|. t∈J t∈J (2.0.17) Show that the axioms of a norm are satisfied. Show that f (x) = x0 (c), c = (a + b)/2, defines a bounded linear functional on C 1 [a, b]. Show that f is not bounded, considered as a functional on the subspace of C[a, b] which consists of all continuously differentiable functions. Solution: First, we show that the four axioms of a norm are satisfied in this problem. (N1). By the definition of maximum and absolute function, kxk ≥ 0. (N2). If kxk = 0, as both max|x(t)| and max|x0 (t)| are nonnegative, |x(t)| = |x0 (t)| = 0. t∈J t∈J If x(t) = 0 is a constant function, then |x0 (t)| = 0. Therefore, kxk = 0. 105 2 1.5 arctan(x/eps) 1 0.5 0 −0.5 −1 −1.5 −2 −100 −80 −60 −40 −20 0 20 40 60 80 100 x Figure 2.1: The function arctan( xε ). Smaller and smaller values of ε leads to a steeper and steeper function. (N3). kαxk = max |αx(t)| + max |αx0 (t)| t∈J t∈J = |α|(max |x(t)| + max |x0 (t)|) t∈J = |α|kxk. 106 t∈J (2.0.18) (N4). The triangle inequality is satisfied kx + yk = max |x(t) + y(t)| + max |x0 (t) + y 0 (t)|. t∈J t∈J ≤ max |x(t)| + max |y(t)| + max |x0 (t)| + max |y 0 (t)|. t∈J t∈J t∈J t∈J = kxk + kyk. (2.0.19) Then, we show that functional f (x) is linear and bounded on C 1 [a, b]. As the space C 1 [a, b] is a normed space, it defines the algebraic operations. For x, y ∈ C 1 [a, b] and any scalar α, the operation f satisfies f (x + y) = (x(c) + y(c))0 = f (x) + f (y), (2.0.20) f (αx) = (x(c) + y(c))0 = f (x) + f (y). (2.0.21) and This shows that f is linear. Consider the norm in C 1 [a, b], |f (x)| = |x0 (c)| ≤ max |x0 (t)| ≤ max |x(t)| + max |x0 (t)| = kxk, t∈J t∈J t∈J (2.0.22) followeding the definition in 2.8-2, f is bounded on C 1 [a, b]. However, on the space C[a, b], where the norm is defined by kxk = max |x(t)|, t∈J (2.0.23) let xn (t) = sin(ntπ/c), where n ∈ N. Then kxn k ≤ 1 and nπ f (xn ) = x0n (c) = cos(nπ). (2.0.24) c So kf (x)k = nπ/c. As n ∈ N is arbitrary, this shows that there is no fixed number α such that kf (x)k ≤ αkxk, so f is unbounded. Ex. 2.10.4 Let X and Y be normed spaces and Tn : X → Y , n = 1, 2, 3, .., be bounded linear operators. Show that convergence Tn → T implies that for every > 0 there is an N such that for all n > N and all x in any given closed ball we have kTn x − T xk < . ˜ 0 ; r) ⊂ X. Then, for all x ∈ B(x ˜ 0 ; r), Solution: Denote the given closed ball by B(x there exists a positive number C such that kxk ≤ C. From the convergence of Tn , for a given > 0, there exists N ∈ N such that for all n > N , (2.0.25) kTn − T k ≤ . C ˜ 0 ; r) Therefore, for all x ∈ B(x kTn x − T xk ≤ kTn − T kkxk ≤ . 107 (2.0.26) Ex. 2.10.8 ∞ Since {ek }∞ k=1 with ek = (δkj )j=1 is a Schauder basis for c0 , for every x ∈ c0 we can write ∞ X x= λk e k , k=1 where x = (λk ) and limn→∞ λk = 0 and we can approximate x by sn = n X λk ek , k=1 then limn→∞ k sn − x k∞ = 0. Let f ∈ c00 , which is the dual space of c0 . This means that f : c0 → R is a bounded linear functional, so f (sn ) → f (x), as n → ∞. So f (x) = ∞ X λk f (ek ). k=1 n) for f (en ) 6= 0 and γn = 0 Now we show that f (ek ) is in l1 . Define the sequence γn = |ff (e (en )| for f (en ) = 0. Consider the sequence xn = (γ1 , γ2 , · · · , γn , 0, 0, · · · ) = γ1 e1 + · · · + γn en . It is obvious that k xn k∞ ≤ 1, so |f (xn )| = So Pn k=1 n X k=1 |f (ek )| ≤k xn k∞ k f k≤k f k . |f (ek )| ≤k f k for all n. Since this inequality holds for all n, we have ∞ X k=1 |f (ek )| ≤k f k . (2.0.27) This means that (f (ek )) ∈ l1 . Now we want to show that for every y = (λk ) ∈ l1 , we can obtain a corresponding bounded linear operator g : c0 → R. Let g(x) = ∞ X λk µk , x = (µk ). k=1 Clearly g is linear and |g(x)| = | ∞ X k=1 λk µk | ≤ sup |µj | j ∞ X k=1 |λk | ≤k x k∞ ∞ X k=1 |λk |. So g is bounded. Taking the supremum over all x with norm 1, we get |g(x)| ≤ ∞ X k=1 108 |λk |. This inequality and the inequality (2.0.27) gives ∞ X k=1 |f (ek )| =k f k . So the mapping f : c00 → l1 with f → (f (ek )) is bijective and preserves norms and hence isomorphism. Ex. 2.10.8 We consider f a bounded linear functional on c0 . Since a Schauder basis for c0 is {ek }k∈N , with ek = δkj , then every x = {αk } ∈ c0 has a unique representation x= ∞ X αk ek k=1 and, from the linearity of f , we know also that f (x) = ∞ X αk f (ek ) . k=1 We can write |f (x)| = ∞ X αk f (ek ) ≤ k=1 ∞ X k=1 |αk | |f (ek )| ≤ kxk∞ ∞ X k=1 |f (ek )| = c kxk∞ , where c = |f (ek )| < ∞ because f (x) is bounded. This implies that {f (ek )} ⊂ l1 and 0 therefore c0 ⊆ l1 . On the other hand, for every b = {βk }k∈N ∈ l1 , we can obtain a bounded linear functional g in c0 . In fact, we may define g on c0 by f (x) = ∞ X αk βk . k=1 Then g is • Linear g (λx + µy) = ∞ X (λαk + µγk ) βk = k=1 ∞ X =λ k=1 αk βk + µ ∞ X k=1 where y = {γk }k∈N ∈ c0 . 109 γk βk = λg (x) + µg (y) , λ, µ ∈ R • Bounded ∞ ∞ ∞ X X X |g (x)| = α k βk ≤ |αk βk | ≤ kxk∞ βk = kxk∞ kbk1 . k=1 0 k=1 k=1 0 Then g ∈ c0 and therefore l1 ⊆ c0 . P We finally prove that kgk = kbk1 . Indeed, since ∞ k=1 |βk | converges, then ∞ X ∀ε > 0, ∃N = N (ε) s.t. k=n+1 |βk | < ε, ∀n > N. Let us consider x s.t. αk = x ∈ c0 , and ( |βk | , βk 0, k < N, k≥N . ∞ ∞ ∞ X X X |g (x) − kbk1 | = αk βk − |βk | = |βk | < ε. k=1 k=1 k=N +1 0 This shows that the bijective linear mapping on l1 onto c0 defined by g is an isomorphism. Ex. 2.10.8 Show that the dual space of the space c0 is l1 . c0 is the space of all sequences of scalars converging to zero. The norm on c0 is k · k∞ . 0 0 Let c0 be the dual space of c0 . We want to show that c0 is isomorphic with l1 . The norm on l1 is k · k1 . We prove it step by step. Step 1: Let x ∈ c0 , then x has a unique representation x= ∞ X ξk ek , k=1 where (ek ) = (δkj ) is a Schauder basis for c0 , (δkj ) has one in the k th place P and zeros otherwise, and ξk → 0 as k → ∞. Since f is linear, we have f (x) = ∞ k=1 ξk f (ek ). Since f is bounded on c0 , we have kf k = sup x∈c0 110 |f (x)| < ∞. kxk∞ We construct s sequence (¯ xk ) such that ( x¯k = |f (ek )| f (ek ) 0 k ≤ n and f (ek ) 6= 0 k > n or f (ek ) = 0 |f (¯ x)| Obviously, (¯ xk ) ∈ c0 . Therefore, k¯ < ∞. Since the elements of (¯ xk ) can only take xk∞ values 0 , 1, or -1, we have k¯ xk∞ = 1. We notice that f (¯ x) = ∞ X x¯k f (ek ) = k=1 n X k=1 |f (ek )|, f (¯ x) ≤ kf kk¯ xk∞ = kf k Since n is arbitrarily large, we have ∞ X k=1 |f (ek )| ≤ kf k. (2.0.28) Step 2: Take any x ∈ c0 , we have |f (x)| = | ∞ X ξk f (ek )| k=1 ≤ kxk∞ ∞ X k=1 |f (ek )|. Taking all x such that kxk∞ = 1 gives kf k ≤ ∞ X k=1 |f (ek )|. (2.0.29) P 1 By (2.0.28) and (2.0.29), we get that kf k = ∞ k=1 |f (ek )|. Therefore, f is in l . This also 1 shows that given P∞ a sequence (ηk ) in l , there exists a linear bounded operator g on c0 such that g(x) = k=1 ξk ηk , where x = ξk ∈ c0 . g is linear by construction. Boundedness can be proved similar to (2.0.29). 0 We conclude that c0 is isomorphic with l1 . 111 Chapter 3 Inner product spaces 112 Ex. 3.3.4 A X is a Hilbert space (complete) and M ⊂ X is closed. A closed subset of a complete one is also complete, so M is complete. M is also a vector space, which is convex by definition. Hence, M has all the properties needed for Theorem 3.3-1. B Appolonius identity is 1 1 ||x − yn ||2 + ||x − ym ||2 = ||yn − ym ||2 + 2||x − (yn + ym )||2 . 2 2 (3.0.1) Since M is convex, we have 1 2||x − (yn + ym )||2 ≥ 4δ 2 2 where δ = ||x − y|| the minimum distance. Putting vn = yn − x and using this in (3.0.1), one obtains 1 ||vn ||2 + ||vm ||2 ≥ ||vn − vm ||2 + 2δ 2 . (3.0.2) 2 Using the parallelogram equality on the left-hand side of (3.0.2), one can conclude that ||vn − vm ||2 ≥ 4δ 2 which is the first part of the proof. The rest of the proof can then proceed as usual. Ex. 3.3.4 (a) Show that the conclusion of Theorem 3.3-1 also holds if X is a Hilbert space and M ⊂ X is a closed subspace. (b) How could we use Appolonius’ identity (Sec. 3.1, Prob. 5) in the proof of Theorem 3.3-1? Solution The problem states that: Let X be a Hilbert space and a closed subspace M ⊂ X. For every given x ∈ X there exists a unique y ∈ M , such that δ = inf kx − y˜k = y˜∈M kx − yk. 1) Existence. From theorem 3.2-4, we know that M is complete. From theorem 3.3-4, the Hilbert space X can be represented as X = M ⊕ M ⊥ . Then for any given x ∈ X there exists y ∈ M and z ∈ M ⊥ such that x = y + z. Then we prove such a y satisfies kx − yk ≤ kx − y˜k for any y˜ ∈ M . Since y − y˜ ∈ M , the norm kx − y˜k2 = = = = ≥ kx − y + y − y˜k2 kz + y − y˜k2 kzk2 + < z, y − y˜ > + < y − y˜, z > +ky − y˜k2 kzk2 + ky − y˜k2 kzk2 . 113 2) Uniqueness. Suppose there are two different elements in M satisfy the minimizing condition, denote as y1 and y2 such that δ = inf kx − y˜k = kx − y1 k = kx − y2 k. By using the Appolonius’ y˜∈M identity, ky1 − y2 k2 = 2kx − y1 k2 + 2kx − y2 k2 − 4kx − = 2δ 2 + 2δ 2 − 4kx − y1 + y2 2 k. 2 y1 + y2 2 k 2 (3.0.3) 2 2 2 As y1 6= y2 , we have kx − y1 +y k < δ 2 . Then, y = y1 +y is a better approximation for 2 2 the norm kx − yk in M ., contradiction. Therefore, there is a unique y ∈ M satisfies the minimizing condition. Ex. 3.3.4 (a) Show that the conclusion of Theorem 3.3-1 (minimizing vector) holds also if X is a Hilbert space and M ⊂ X is a closed subspace. (b) How could we use Appolonius’ identity in the proof of Theorem 3.3-1? Solution: (a) The assumptions in Theorem 3.3-1 are that X is an inner product space and M 6= ∅ is a convex subset which is complete. We now assume that X is a Hilbert space and M is a closed subspace. Since M is closed and X is a Hilbert space, M is complete. Since M is a subspace, M is also convex. Thus, all the assumptions in Theorem 3.3-1 are fulfilled and the conclusions of the Theorem follow. (b) Appolonius’ identity reads kc − ak2 + kc − bk2 = 1 1 ka − bk2 + 2kc − (a + b)k2 2 2 or, equivalently, 1 ka − bk2 = 2 kc − ak2 + 2 kc − bk2 − 4kc − (a + b)k2 . 2 (3.0.4) To prove uniqueness of the minimizing vector in Theorem 3.3-1, assume that both y ∈ M and y0 ∈ M satisfy ky − xk = δ and ky0 − xk = δ. By Appolonius’ identity (3.0.4) with a = y, b = y0 and c = x, we obtain 1 ky − y0 k2 = 2 kx − yk2 + 2 kx − y0 k2 − 4kx − (y + y0 )k2 2 1 = 2δ 2 + 2δ 2 − 4kx − (y + y0 )k2 . 2 Because M is convex, 21 (y + y0 ) ∈ M , so that kx − 12 (y + y0 )k ≥ δ, which implies ky − y0 k2 ≤ 2δ 2 + 2δ 2 − 4δ 2 = 0. 114 But since ky − y0 k2 ≥ 0, we must have ky − y0 k = 0 and thus y = y0 , which proves that the minimizing vector is unique. Ex. 3.3.10 We want to show that if Y ⊂ H is a closed subspace such that M ⊂ Y implies that M ⊥⊥ ⊂ Y. As a first step let y⊥ ∈ Y ⊥ . This implies hy⊥ , mi = 0 ∀m ∈ M since M ⊂ Y , which shows us that y⊥ ∈ M ⊥ . This means that Y ⊥ ⊂ M⊥ (3.0.5) since y⊥ was arbitrary. L ⊥ Let now x ∈ M ⊥⊥ . Since Y is a closed subspace we have that X = Y Y . So x can ⊥ be decomposed as x = y + y⊥ with y ∈ Y and y⊥ ∈ Y . But from (3.0.5) we have that hx, y⊥ i = 0 so we get 0 = x, y ⊥ = hy + y⊥ , y⊥ i = ky⊥ k2 , which means that y⊥ = 0. Thus x = y ∈ Y , and since x was arbitrary we get M ⊥⊥ ⊂ Y. Ex. 3.3.10 If M ⊂ H, show that M ⊥⊥ is contained in any closed subspace Y ⊂ H such that M ⊂ Y . There are two cases: Case I) if M is closed; then, M = M ⊥⊥ by Theorem (3.3-6). Since M ⊂ H, then M ⊥⊥ is the smallest subset of H that contains M . Case II) if M is not closed; we know that M ⊥⊥ contains M ; i.e., ∀x ∈ M, x⊥M ⊥ ⇒ x ∈ M ⊥⊥ ⇒ M ⊂ M ⊥⊥ If there exists a closed subspace Y ⊂ H, then, M ⊂ Y ⇒ Y ⊥ ⊂ M ⊥ ⇒ M ⊥⊥ ⊂ Y ⊥⊥ = Y Y was arbitrary, so M ⊥⊥ is the smallest closed subset of H which contains M . 115 Ex. 3.4.6 (Minimum property of Fourier coefficients) Let {e1 , . . . , en } be an orthonormal set in an inner product space X, where n is fixed. Let x ∈ X be any fixed element and y = β1 e1 + . . . + βn en . Then kx − yk depends on β1 , . . . , βn . Show by direct calculation that kx − yk is minimum if and only if βj = hx, ej i, where j = 1, . . . , n. Solution: kx − yk2 = hx − y, x − yi = hx, xi − hx, yi − hy, xi + hy, yi = kxk2 + kyk2 − hx, yi − hx, yi = kxk2 + kyk2 − 2Re {hx, yi} ≥ kxk2 + kyk2 − 2|hx, yi| n n X X = kxk2 + |βk |2 − 2|hx, βk ek i| k=1 ≥ kxk2 + 2 = kxk + By choosing βk = hx, ek i, n X k=1 n X k=1 k=1 |βk |2 − 2 n X k=1 |βk ||hx, ek i| 2 2 2 (|βk | − |hx, ek i|) − |hx, ek i| ≥ kxk − n X k=1 |hx, ek i|2 k = 1, . . . , n, we obtain kx − yk2 = kxk2 + kyk2 − 2Re {hx, yi} ) ( n n X X βk e k i = kxk2 + |βk |2 − 2Re hx, ( n k=1 ) X |βk |2 − 2Re βk hx, ek i k=1 = kxk2 + n X = kxk2 − n X k=1 k=1 ) ( n n X X |hx, ek i|2 = kxk2 + |hx, ek i|2 − 2Re k=1 k=1 k=1 |hx, ek i|2 which shows that the choice βk = hx, ek i, k = 1, . . . , n minimizes kx − yk 116 Ex. 3.4.6 Since X is an inner product space and x, y ∈ X, x − y is also in X, and we can use the Bessel inequality for x − y: 2 kx − yk ≥ = n X k=1 n X k=1 = = n X k=1 n X k=1 |hx − y, ek i|2 |hx, ek i − hy, ek i|2 |hx, ek i − n X j=1 βj hej , ek i|2 |hx, ek i − βk |2 . Thus, the minimun of kx − yk is obtained if and only if βk = hx, ek i. Ex. 3.4.6 Let {e1 , ..., en } be an orthonormal set in an inner product space X, where n is fixed. Let x ∈ X be any fixed element and y = β1 e1 + ... + βn en . We want to show that kx − yk is minimum if and only if βi = hx, ei i. The set {e1 , ..., en } is a basis of a closed finite vectorial subspace Y ⊂ X, i.e. Y = span(e1 , ..., en ), dim(Y ) = n. which is a convex and complete (see Theorem 2.4-2) subspace of X (see Exercise 3.3.4). Then we can say that y = β1 e1 + ... + βn en ∈ Y for any list of n scalars (β1 , ..., βn ). From Theorem 3.3-1, we know that for every given x ∈ X there exists a unique y ∈ Y such that kx − yk = inf kx − y˜k, y˜∈Y (3.0.6) which in our context we can rephrase as follows: there exists a unique list of n scalars (β1 , ..., βn ) such that y = β1 e1 + ... + βn en minizes kx − yk. From Lemma 3.3-2 we also know that a such y must be the orthogonal projection of x on Y , i.e. z = x−y ∈ Y ⊥ . Then we need to show that if we choose the coeffincients of y to Pbe βi = hx, ei i, then z = x − y is orthogonal to any other element of Y . We take v ∈ Y , nk=1 αk ek and we multiply it by 117 z with respect to the inner product of X: hz, vi = hx − y, vi = hx, vi − hy, vi = hx, = = = = Pn k=1 αk ek i − h Pn αk hx, ek i − Pn αk hx, ek i − k=1 Pn k=1 k=1 Pn k=1 αk hx, ek i − Pn k=1 Pn k=1 Pn k=1 Pn k=1 βk ek , Pn βk hek , βk Pn k=1 Pn j=1 αk ek i k=1 αk ek i αj hek , ej i βk α k αk [hx, ek i − βk ] which prove Pnthat hz, vi = 0 iff βi = hx, ei i . Note that in the second last equality we have used that j=1 αj hek , ej i = αk since hek , ej i = 1 only when k = j and zero otherwise. 3.4.8 From Bessel’s inequality, we have ∞ X |hx, ei i|2 j=1 ||x||2 ≤ 1. Assume we have nm inner products hx, ei i satisfying |hx, ei i| > nm X |hx, ek i|2 k=1 ||x||2 > nm X k=1 (3.0.7) 1 . m We have 1 nm = . m2 ||x||2 m2 ||x||2 We must then (at least) have nm < m2 ||x||2 , since (3.0.7) cannot hold otherwise. This proves the statement. Ex. 3.5.4 If (xj ) is a sequence in an inner product space X such that the series kx1 k + kx2 k + kx3 k + . . . converges, show that Sn is a Cauchy sequence, where Sn = x1 + . . . + xn 118 (3.0.8) Solution: We let S n = x1 + x2 + x3 + . . . + xn (3.0.9) Then kSn − Sm k = kxm + xm+1 + · · · + xn k ≤ kxm k+kxm+1 k+. . .+kxn k → 0 as n, m → ∞ (3.0.10) Since the series (3.0.8) converges. Ex. 3.5.6 Given: x = P∞ i=1 αi ei and y = hx, yi = h = ∞ X αi ei , i=1 ∞ ∞ XX P∞ j=1 ∞ X j=1 βj ej where, (ei ) is an orthonormal sequence. Then, ∞ X ∞ ∞ X ∞ X X βj ej i = hαi ei , βj ej i = αi βj hei , ej i αi βj δij = i=1 j=1 i=1 j=1 ∞ X i=1 j=1 αi βi i=1 It must be added that the series of x and y are absolutely convergent. For example, for x: ∞ ∞ ∞ ∞ X X X X kxk = hx, xi = h αi e i , αj e j i = αi αi = |αi |2 2 i=1 i=1 i=1 i=1 Ex. 3.6.4 We study hx + y, x + yi = hx, xi + hx, yi + hy, xi + hy, yi which can be re-written by using Parsevals identity X X X hx + y, ek ihx + y, ek i = hx, ek ihx, ek i + hy, ek ihy, ek i + hx, yi + hy, xi. (3.0.11) By using the linear properties of the inner product and rearranging, (3.0.11) can be written X X hx, yi + hy, xi = hy, ek ihx, ek i + hx, ek ihy, ek i. k k P P P Since hx, yi+hy, xi = 2Re(hx, yi) and k hy, ek ihx, ek i+ k hx, ek ihy, ek i = 2Re( k hy, ek ihx, ek i) we can conclude that X Re(hx, yi) = Re( hx, ek ihy, ek i). k 119 As for the imaginary part, we have hx + iy, x + iyi = hx, xi − ihx, yi + ihy, xi + hy, yi, where i is now the usual imaginary number. Repeating the same procedure on this, one arrives at X X i(−hx, yi + hy, xi) = i( hy, ek ihx, ek i − hx, ek ihy, ek i). k k P P By using i(−hx, yi + hy, xi) = 2Im(hx, yi) and i( k hy, ek ihx, ek i − k hx, ek ihy, ek i) = P 2Im( k hy, ek ihx, ek i), one can conclude X Im(hx, yi) = Im( hx, ek ihy, ek i). k Since both the real and imaginary parts are equal, we must have X hx, ek ihy, ek i, hx, yi = k which completes the proof. Ex. 3.6.4 Let M be a total orthonormal set in a Hilbert space H. Then ∀ x, y ∈ H we know that X X x= hx, em i em and y= hy, en i en (3.0.12) m n where en and em are the elements of M which generate a countable set of Fourier coeffincients (Lemma 3.5-3) for x and y respectively. In the next step we consider (ek ) = (el ) = (em ) ∪ (en ) which is a countable set since it is the countable union of countable sets. P P hx, yi = h k hx, ek i ek , l hy, el i el i = = = P k hx, ek i hek , P P k P k P l hy, el i el i l hx, ek i hy, el i hek , el i hx, ek i hy, ek i which is another version of the Parseval relation. 120 (3.0.13) Ex. 3.6.4 P P We have Parseval’s relation k |hx, ek i|2 = kxk2 and wish to show that hx, yi = k hx, ek ihy, ek i. Consider kx − yk. We have kx − yk2 = hx − y, x − yi = kxk2 + kyk2 − 2<[hx, yi] but also from Parseval’s relation that X X kx − yk2 = |hx − y, ek i|2 = |hx, ek i − hy, ek i|2 X X X = |hx, ek i|2 + |hy, ek i|2 − 2<[ hx, ek ihy, ek i] X = kxk2 + kyk2 − 2<[ hx, ek ihy, ek i], where we again have used Parseval’s relation in the final equality. Comparing the two expressions tells us that X <[hx, yi] = <[ hx, ek ihy, ek i]. k Repeating the exercise using kx + yk yields the analogous result for the imaginary parts, X =[hx, yi] = =[ hx, ek ihy, ek i]. k Consequently the desired result follows. Ex. 3.6.4 Let x ∈ X, M = {ek } ⊂ X be a total orthonormal set and X x˜ = hx, ej i ej . j Since hx − x˜, ej i = 0 ∀i we have x − x˜ ⊥ M. This implies x = x˜ by Theorem 3.6.2 a) since M is total in X. Thus we can write any element y ∈ X as X y= hy, ek i ek . k This now gives hx, yi = = * X j X k hx, ej i ej , X k hy, ek i ek + hx, ek i hy, ek i . 121 = XX j k hx, ej i hy, ek i hek , ej i = Ex. 3.6.10 Let M be a subset of a Hilbert space H, and let u, v ∈ H. Suppose that hv, xi = hw, xi for all x ∈ M implies v = w. If this holds for all v, w ∈ H, show that M is total in H. Solution: By the linearity of the inner product hv, xi = hw, xi ⇐⇒ hv − w, xi = 0. Thus, hv, xi = hw, xi for all x ∈ M implies v = w, for all v, w ∈ H ⇐⇒ hv − w, xi = 0 for all x ∈ M implies v = w, for all v, w ∈ H ⇐⇒ hu, xi = 0 for all x ∈ M implies u = 0 ⇐⇒ u⊥M implies u = 0 ⇐⇒ M is total in H. In the last step we used Theorem 3.6-2. Here it is important that H is complete. Ex. 3.6.10 Let us consider M subset of H Hilbert space and v, w ∈ H. Moreover we admit that hv, xi = hw, xi , ∀x ∈ M ⇒ v = w holds ∀v, w ∈ H. We want to prove that M is total: in order to do this we consider z ⊥ M , that satisfies for construction hz, xi = 0, ∀x ∈ M . But since also h0, xi = 0, ∀x ∈ M , then we have hz, xi = h0, xi and this implies that z = 0. Thus, by Theorem 3.6.2 the claim follows. Ex. 3.8.10 Sesquilinearity follows by h(x1 + x2 , y) = hx1 + x2 , yi = hx1 , yi + hx2 , yi = h(x1 , y) + h(x2 , y) h(x, y1 + y2 ) = hx, y1 + y2 i = hx, y1 i + hx, y2 i = h(x, y1 ) + h(x, y2 ) h(αx, y) = hαx, yi = α hx, yi = αh(x, y) ¯ h(x, βy) = hx, βyi = β¯ hx, yi = βh(x, y). 122 From the Cauchy-Schwartz inequality we get: khk = sup sup x6=0 y6=0 | hx, yi | kxk kyk ≤ sup sup = 1. kxk kyk x6=0 y6=0 kxk kyk But note that for x = y we have h(x, x) kxk2 = =1 kxk2 kxk2 so we get khk = 1. Ex. 3.8.14 Schwarz inequality. Let X be a vector space and h a Hermitian form on X × X. This form is said to be positive semidefinite if h(x, x) ≥ 0 for all x ∈ X. Show that then h satisfies the Schwarz inequality |h(x, y)|2 ≤ h(x, x)h(y, y). Since h is positive semidefinite, we have h(x − λy, x − λy) ≥ 0. By using the fact that h is a Hermitian sesquilinear form, we obtain ¯ ¯ h(x, x) − λh(x, y) − λh(y, x) + λλh(y, y) ≥ 0. Choose λ= (3.0.14) h(x, y) . h(y, y) Thus, ¯ = h(x, y) = h(y, x) . λ h(y, y) h(y, y) Substituting this into inequality (3.0.14), we obtain h(x, y) h(x, y)h(y, x) h(y, x) h(x, y) − h(y, x) + h(y, y) h(y, y) h(y, y) h(y, x) = h(x, x) − h(x, y) h(y, y) |h(x, y)|2 = h(x, x) − . h(y, y) 0 ≤ h(x, x) − Therefore, |h(x, y)|2 ≤ h(x, x)h(y, y). Remark: It may not be straightforward to get the correct λ value. The idea is that the equality in the Schwarz inequality is taken when x and y are linear dependent. Therefore, we use λ to project y so that λy and x are linear dependent. 123
© Copyright 2024