Computational Engineering and Technical Physics Technomathematics
Joël Right Dzokou Talla
ELLIPTIC BOUNDARY VALUES PROBLEMS WITH GAUSSIAN WHITE NOISE
Master’s Thesis
Examiners: Associate Professor Lassi Roininen Dr. Sari Lasanen
Supervisors: Associate Professor Lassi Roininen Dr. Sari Lasanen
Lappeenranta-Lahti University of Technology School of Engineering Science
Computational Engineering and Technical Physics Technomathematics
Joël Right Dzokou Talla
Elliptic Boundary Values Problems with Gaussian White Noise
Master’s Thesis 2019
52 pages.
Examiners: Associate Professor Lassi Roininen Dr. Sari Lasanen
Keywords: Boundary value problem ·Gaussian random field · Gaussian white noise · inverse problem·Bayesian estimation.
This work focuses on the study of elliptic boundary value problems with Gaussian white noise on the formLX =W, whereL=−∆ +ν2, withν > 0. First, the definitions and some properties of some important tools for this study such as, Gaussian random field and Gaussian white noise are given. Next, the boundary value problem is solved in dimensions d = 1,2, for the case of Dirichlet boundary condition. The spectral theorem is used to show that the operatorL admits eigenfunctions that form a complete orthonormal basis and the boundary value problem is solved in the eigenbasis ofL. Finally, the Bayesian estimation approach for inverse problem is derived, and as application, a one dimensional deconvolution problem is considered. The mathematical model is discretized on finite interval and the Bayesian estimation approach is used to estimate the truncated unknown.
First of all, I would like to thankAlmighty God who has kept me this far and who has helped me to provide this work.
We could not have achieved such a work without the collaboration of many people.
I want to thank my supervisor, Dr.Sari Lasanen. She is incredibly kind. Interviews with Sari have always comforted me and given energy to continue. I admire her approach in research. I thank her especially for the knowledge she has impacted in me and the freedom she has given me during this work. I appreciate Associate ProfessorLassi Roininenfor his advices, the trust and freedom he has given me all through this work. I appreciate your kind personality.
I thank my family, especially my parents for their unconditional love and support they have provided me throughout my studies. I thank my friends and classmates for being there with me and for the good time we spent together.
Finally, I thank the department of Technomathematics, School of Computational Engi- neering and Technical Physics of LUT University for given me the opportunity to pursue this Master’s program. Thank you for the scholarship provided to aid my study and for allowing me to work in better conditions.
Lappeenranta, June 24, 2019
Joël Right Dzokou Talla
CONTENTS
1 INTRODUCTION 5
2 PRELIMINARIES 7
2.1 Banach and Hilbert spaces . . . 7
2.2 Sobolev spaces . . . 10
2.3 Probability measures . . . 12
2.4 The Cameron-Martin space . . . 14
2.5 White noise . . . 16
3 EXISTENCE AND UNIQUENESS OF SOLUTION 20 3.1 Presentation of the equation . . . 20
3.2 Existence and uniqueness . . . 24
4 BAYESIAN ESTIMATION 32 4.1 Bayes theorem . . . 32
4.2 Estimators . . . 34
5 EXAMPLES 35 5.1 One-dimensional deconvolution problem with white noise prior . . . 35
5.2 One-dimensional deconvolution problem with the priorX ∼ N(0, γ2L−2) 38 5.3 Discretization of the problem by truncation . . . 40
5.3.1 Discretization with truncated prior . . . 40
5.3.2 Posterior covariance matrix . . . 43
5.3.3 Numerical reconstruction . . . 45
6 CONCLUSION AND DISCUSSION 50
REFERENCES 51
1 INTRODUCTION
In general, many phenomena in nature are modelled using partial differential equations.
These phenomena can sometimes contain a certain level of uncertainty thus making mod- elling more difficult. Modelling using stochastic partial differential equations turn out to be the most appropriate way to capture the behaviour of such phenomena. The stochastic differential equation is used to define the prior is the statistical inverse problem [1, 2].
In this work, we study the elliptic Dirichlet boundary values problem modelled by the following stochastic partial differential equations
LX =W, inD, X = 0, on∂D.
(1.1)
whereDis a bounded Lipschitz domain, the operatorLis defined as−∆ +ν2withν >0 a constant, the unknownX is a random field andW is a zero-mean Gaussian white noise onD. In the analytical resolution of the above Equation (1.1), we encounter many diffi- culties due to the inherent complexity of partial differential equations and the irregularity of the noise thermW which appears in this equation.
In many problems such as the stochastic Navier-Stokes equations, the stochastic heat equation [1] and the inverse seismic problem [2], we encounter some irregular stochastic effects which make them more difficult to solve. Due to the irregularity of these stochastic effects, it becomes then very difficult to determine the unknownXfrom the mathematical model. Sometimes the problems may be ill-posed in the sense of Hadamard, that is, may not admit a unique solution or the solution may not depend continuously on the data. The ill-posed problems may be view as inverse problems [3]. To overcome these difficulties, some methods such as Bayesian statistical method in inverse problems turn out to be a promising alternative that offers more useful results [3, 4].
There exist a vast literature on solving Equation (1.1) in which many authors such as Walsh in [5], Buckdahn and Pardoux in [6] and Lasanen et al [7] have brought partial answers under various constraints more or less restrictive on the irregularity of the white noiseW. In [7] it is shown in the case of Dirichlet, Neumann and Robin boundary con- ditions that (1.1) has a unique Gaussian zero mean Banach space-valued random field solution.
Our main objectives in this work is first to base ourselves on the work done by Lasa- nen et al [7] to show the existence and the uniqueness of a random field solution of (1.1)
in the eigenbasis of the operatorL, where the white noiseW has actually a representation in the eigenbasis ofL. Next, we will derive the Bayesian approach to inverse problem as in [4] and then we will use it to approximate the prior by truncation, that is, to approxi- mate it by a finite sum.
In order to achieve this goal, we structure our work as follows: Chapter 2 is devoted to some preliminary results known on Sobolev spaces, Gaussian random variables and some properties of Gaussian white noise. In Chapter 3 we study the existence and uniqueness of a Gaussian random field X in the eigenbasis of operatorL in dimensionsd = 1,2.
In Chapter 4 we derive the Bayesian estimation approach for linear inverse problems. In Chapter 5 we apply the Bayesian approach to a discretized problem.
2 PRELIMINARIES
In this section, we are going to present the concepts of Sobolev spaces, probability mea- sure and Gaussian white noise that will be used later on to study the existence and unique- ness of the solution of Equation (1.1).
2.1 Banach and Hilbert spaces
We recall the definition and some properties of Banach and Hilbert spaces. The proofs of some results given in these subsections can be found in the books of Douglas Farenick [8]
and Sopka et al [11] .
Definition 2.1.1 ( [8, p. 13]). Let M be a non-empty set. We call distance on M the function
d:M ×M →[0,+∞[
satisfying the following properties, for allx, y ∈M 1. d(x, y) = 0⇔x=y,
2. d(x, y) =d(y, x),
3. d(x, z)≤d(x, y) +d(y, z), z ∈M.
The spaceM endowed with the distancedis calledmetric spaceand we denote it(M, d).
We are going to define a complete metric space which will allow us to define the concept of Banach space.
Definition 2.1.2( [11, p. 28]). Let(M, d)be a metric space and (xn)n∈
N a sequence in M. The sequence(xn)n∈
Nis said to beCauchy sequenceif(xn)n∈
Nsatisfies the following condition:
∀ε >0, there existsN ∈N such that d(xm, xn)< ε, ∀m, n≥N.
Definition 2.1.3( [11, p. 28]). A metric space(M, d)iscompleteif all Cauchy sequences inM are convergent.
Definition 2.1.4( [11, p. 59]). A normon a vector spaceV is a function k.k : V → R such that, for allu, v ∈V andα∈C, the following properties hold:
1. kuk ≥0
2. kuk= 0⇔u= 0 3. kαuk=|α| kuk
4. ku+vk ≤ kuk+kvk.
If k·k is a norm on V, then the space V endowed with the norm k·k is called normed vector spaceand we denote it by(V,k·k).
Definition 2.1.5( [8, p. 170]). LetM be a vector space and letk·kbe a norm onM. A normed vector spaceM is said to be complete if all Cauchy sequences inM are conver- gent. A complete normed vector space is called Banach space. The spaceLp(Ω), with Ω⊂Cdand1≤p < ∞defined as
Lp(Ω) =
f : Ω→C, Z
Ω
|f(x)|pdx <∞
, (2.1)
is Banach space under the normkfk= R
Ω|f(x)|pdx1/p
.
Proposition 2.1.1 ( [8, p. 170]). In finite dimensions, every normed vector space is a Banach space.
Definition 2.1.6( [8, p. 23]). LetM be a Banach space. A subspaceAofM is said to be denseinM if A = M, whereA denotes the closure ofA inM. That is, for an element a∈M, there exists a sequence inAthat converges toa.
Proposition 2.1.2( [8, p. 57]). A Banach spaceM isseparableif there exist a countable subsetA ⊂M which is dense inM.
Definition 2.1.7 ( [8, p. 195]). An inner product on a complex vector space H is a complex-valued function h·,·ion the Cartesian product H×H satisfying the following properties: for allξ, η, ξ1, ξ2 ∈H, and α1, α2 ∈C
1. hξ, ξi ≥0,
2. hξ, ξi= 0⇔ξ = 0,
3. hα1ξ1+α2ξ2, ηi=α1hξ1, ηi+α2hξ2, ηi, 4. hξ, ηi=hη, ξi.
The following lemma gives the Cauchy-Schwartz inequality in a complex vector space.
Its proof can be found in [8, p. 196].
Lemma 2.1.1. LetHbe a complex vector space endowed with an inner producth·,·i. For allξ, η ∈H, we have
|hξ, ηi| ≤ hξ, ξi12hη, ηi12.
Proposition 2.1.3( [8, p. 196]). Ifh., .iis an inner product on a vector spaceH, then
kξk=hξ, ξi12 (2.2)
defines a norm onH.
Proof. The three first conditions of Definition 2.1.4 are immediate by using Definition 2.1.7. Now letξ, η ∈H, using the fact thathξ, ηi+hη, ξi= 2 Re(hξ, ηi), we have
kξ+ηk2 =hξ+η, ξ+ηi
=kξk2+ 2 Re(hξ, ηi) +kηk2
≤ kξk2+ 2|hξ, ηi|+kηk2
≤ kξk2+ 2kξk kηk+kηk2 by using Lemma2.1.1
= (kξk+kηk)2.
Hence (2.2) defines a norm onH.
Definition 2.1.8. A vector spaceHendowed with an inner product is calledinner product space.
Definition 2.1.9( [8, p. 197]). AHilbert spaceis an inner product space such thatH is a Banach space with respect to the normkξk=hξ, ξi12.
Forp= 2, the Banach spaceL2(Ω)defined in Equation (2.1) is an Hilbert space with the norm induce by the inner product
hf, gi= Z
Ω
f(x)g(x)dx.
Definition 2.1.10( [8, p. 221]). LetBbe a Banach space. Thedual spaceofX denoted byB∗, is the space of all continuous linear functionals onX. That is,
B∗ ={T :B→C, T is continuous and linear}.
Proposition 2.1.4 ( [8, p. 221]). The dual space of a Banach space is also a Banach space.
Let us denote byB∗∗ the dual space ofB∗, we say that the Banach spaceBisreflexiveif Bis the dual of its dual, that is, the applicationj :B→B∗∗define by
j(ξ)(Φ) = Φ(ξ), ∀ξ∈B, ∀Φ∈B∗
is surjective.
Definition 2.1.11( [8, p. 202]). LetH be a Hilbert space. A family(en)of elements ofH is anorthonormal basisofHif
1. kenk= 1, ∀n,
2. hen, emi= 0, ∀n, m, n6=m;
3. The linear span of the family(en)is dense inH.
If(en)is an orthonormal basis ofH, then everyu∈Hcan be written as u=
∞
X
n=1
hu, enien, with kuk2 =
∞
X
n=1
|hu, eni|2.
Theorem 2.1.1( [8, p. 203]). Every separable Hilbert space admits an orthonormal basis.
2.2 Sobolev spaces
We introduce some function spaces calledSobolev spaceswhich have a great importance in the development of partial differential equations. To know more about these spaces, the reader can refer to Golse [12] and Brezis [10].
Definition 2.2.1( [10, p. 157]). LetΩbe an open set ofRd. Let1≤p≤ ∞andm∈N. The linear spaceWm,p(Ω)defined as
Wm,p(Ω) =
u∈Lp(Ω) ; Dαu∈Lp(Ω), ∀α∈Nd:|α| ≤m ,
where
Dα =
∂|α|
∂xα11 · · ·∂xαNd
is calledSobolev space. Here|α|=α1+· · ·+αd.
The spaceWm,p(Ω)is a Banach space with respect to the norm
kukWm,p(Ω) =
P
|α|≤m
kDαukpLp(Ω)
!1p
, if1≤p <∞
|α|≤mmaxkDαukL∞(Ω), if p= +∞
(2.3)
We denote byW0m,p(Ω)the closure ofD(Ω)inWm,p(Ω), for allm∈Nand1≤p <∞, where D(Ω) is the space of test functions, that is, the space of infinitely differentiable functions with compact support on Ω. If p = 2, we denote Wm,2(Ω) = Hm(Ω) and analogously,H0m(Ω) is the closure ofD(Ω)inHm(Ω). For m ∈ Nand1 ≤ p < ∞the
space
W−m,p0(Ω) =
u∈ D0(Ω), u= X
|α|≤m
Dαuα, uα ∈Lp(Ω)
defines the dual space of W0m,p(Ω), with 1p + p10 = 1, where D0(Ω) is the space of dis- tribution on D(Ω), that is, the space of continuous linear forms on D(Ω). The space W−m,p0(Ω)is a Banach space with the norm
kukW−m,p0
(Ω)(f) = sup
kvkW m,p(Ω)=1
hu, vi
We can also define a Sobolev space of non-integer order using Fourier transform. Let us first recall the definition of Fourier transform.
Let us denote byC∞(Rd)the space of infinitely differentiable functions.
Definition 2.2.2( [12, p. 150]). TheSchwartzspaceS(Rd)is the space ofC∞(Rd)func- tions decreasing rapidly with all its derivatives, that is,
S Rd
=
u∈C∞ :α, β ∈Nd, sup
x∈Rd
xβ|Dαu(x)|
<+∞
,
for all multi-indicesαandβ.
The functionu(x) = exp(−a|x|2), witha >0belongs to theSchwartzspaceS(Rd).
Definition 2.2.3 ( [12, p. 162]). A tempered distribution on Rd is a continuous linear form onS(Rd).
We denote the dual space of S(Rd)by S0(Rd) the space of continuous linear forms on S(Rd).
Definition 2.2.4( [12, p. 154]). The Fourier Transform ofu ∈S0(Rd)is the distribution Fudefined by
hFu, vi=hu,Fvi, for allv ∈S(Rd).
Definition 2.2.5( [12, p. 160]). Lets∈ R. We denote byHs Rd
the space of functions u∈S0 Rd
such that
Z
Rd
1 +|ξ|2s
|Fu(ξ)|2dξ < ∞.
For allu, v ∈Hs Rd
, the inner product is defined by (u, v)Hs(Rd) =
Z
Rd
1 +|ξ|2s
Fu(ξ)Fv(ξ)dξ.
The associated norm is
kukHs(Rd) =
Z
Rd
1 +|ξ|2s
|Fu(ξ)|2dξ
1 2
, u∈Hs Rd .
Proposition 2.2.1 ( [10, p. 125]). Lets ∈ RandΩ ⊂ Cd. The Sobolev spaceHs(Ω) is an reflexive and separable Hilbert space.
2.3 Probability measures
We recall here some measure theoretic concepts that will help to define a Gaussian mea- sure.
Definition 2.3.1 ( [8, p. 77]). LetΩbe an arbitrary non-empty set. Let Σ be a class of subsets ofΩ, that is,Σ⊆2Ω. The classΣis calledσ-algebraif the following conditions are satisfied
1. Ω∈Σ
2. ifA ∈Σ, thenΩ\A∈Σ
3. ifAi ∈Σ, fori= 1,2, . . . , then ∞∪
i=1Ai ∈Σ.
Let a functionP : Σ→[0,1]and{Ai}∞i=1 ⊆Σbe a countable collection of disjoint sets.
ThenP is calledprobability measureif the following conditions are satisfied 1. P ∞
i=1∪ Ai
=
∞
P
i=1
P (Ai), and
2. P (Ω) = 1.
The triple(Ω,Σ, P)is calledprobability space, whereΣis aσ-algebra andP is a prob- ability measure.
Remark 2.3.1. • The couple(Ω,Σ)is calledmeasurable space.
• The triple(Ω,Σ, P)is said to becompleteprobability space if for allB ∈Σ, with P (B) = 0and for allA⊂B one hasA∈Σ.
Definition 2.3.2 ( [8, p. 78]). Let C0 be the collection of all open sets inΩ. The Borel σ-algebrais theσ-algebra generated byC0and it is denoted byB(Ω).
An element ofB(Ω)is called aBorel measurable set, or simplyBorel set.
Definition 2.3.3 ( [13, p. 42]). Let (Ω,Σ, P) be a probability space and (E,Π) be a measurable space. Arandom variableis a functionX : Ω→E such thatX−1(A)∈Σ, for allA∈Π.
In general, if we takeE to be a Banach space andΠ = B(E), then in this caseX is an E-valued random variable.
A familyX = {X(t), t∈I}of random variables is calledrandom field, whereI is the index set.
Definition 2.3.4( [13, p. 1]). LetE(Ω)be the Borelσ-algebra. A measureµ:E(Ω)→R is calledBorel measure.
IfP is a Borel probability measure on(R,B(R)), thenP is Gaussian if and only if P(B) =
Z
B
√1
2πσexp
−(x−m)2 2σ2
dx, for all Borel setsB.
The parametersmandσ2 aremeanandvarianceofP, respectively.
Remark 2.3.2. • Ifm = 0, themP is said to be centered or symmetric.
• Gaussian distribution are also called normal.
Definition 2.3.5( [13, p. 203]). Let(Ω,Σ, P)be a probability space. A random variable on(Ω,Σ, P)is Gaussian if its distribution is Gaussian, that is,
P(X ∈B) = Z
B
√1
2πσexp
−(x−m)2 2σ2
dx, for all Borel setsB.
A Gaussian random variable with a centered or symmetric distribution is called centered or symmetric. We can also define a Gaussian random variable on an infinite dimensional Banach spaceB.
Let us denote byB∗ the topological dual ofBandh·,·i
B,B∗ the duality betweenBandB∗. Definition 2.3.6 ( [7, p. 5]). Let B be a separable Banach space and let P be a Borel probability measure on B. ThenP is Gaussian if and only if for each continuous linear functionalλ ∈B∗, thepush-forwardmeasure
µλ =P ◦λ−1 (2.4)
is Gaussian onR.
Definition 2.3.7( [7, p. 5]). LetXbe aB-valued random variable. Xis called Gaussian if and only ifhX, b∗i
B,B∗is Gaussian for allb∗ ∈B∗.
For sake of simplicity, let us take B to be a reflexive separable Banach space. We will adopt the following notations for the mean and the covariance operator of a B-valued
random variableX. Let us denote bymthe mean ofX, that is, hm, b∗i
B,B∗ =EhX, b∗i
B,B∗ ∀b∗ ∈B∗
and we denote byCX the covariance operator ofX defined fromB∗ toBsuch that hCXb∗, b∗i
B,B∗ =EhX−m, b∗i
B,B∗hX−m, b∗i
B,B∗ ∀b∗ ∈B∗. Here
E[f] = Z
Ω
f(w)dP(w).
2.4 The Cameron-Martin space
Definition 2.4.1( [7, p. 5]). LetBbe a separable reflexive Banach space and letX be a GaussianB-valued random variable with zero mean and covariance operatorCX which is non trivial. Set
kb∗kµ
X =q
hCXb∗, b∗i
B,B∗, ∀b∗ ∈B∗ and we denote byB∗µX the closure ofB∗ in the normk·kµ
X. Let us recall the definition ofµX-measurable linear functional.
Definition 2.4.2( [13, p. 80]). LetµX be a Gaussian measure on a Banach space B. A functionf onBis calledµ-measurable linear functional on(B, µX)if there exist a linear subspaceB1of full measure and linear in the usual sense functionf0onB1that coincides withf µX-almost surely.
If aµX-measurable linear functionalf : B → Ris defined in such a way thatf0 will be linear on all ofB, then such a version off is called proper linear version.
Proposition 2.4.1 ( [13, p. 80]). Let µX be a centered Gaussian measure on B. If f is µX-measurable linear functional onX, thenf is a centered Gaussian random variable.
Through this work, we denote the elements of B∗µX by fˆwhich can be identified with µX-measurable linear functionals onB.
Lemma 2.4.1. EveryfˆinB∗µX can be identified with theµX-measurable linear functional b 7→fb(b)that are Gaussian random variables with zero mean on the space(B,B, µX).
Moreover, the covariance
f ,bbg
µX = Z
fb(b)bg(b)µX(db),
where f ,b bg ∈ B∗µX, defines an inner product on B∗µX and the corresponding norm is
f ,bfb
µX
= fb
2 µX
.
Theorem 2.4.1( [1, p. 10]). (Fernique)
LetBbe a reflexive separable Banach space, and letX be aB-valued Gaussian random variable. There exists a constantα >0such thatE
exp αkXk2
<∞.
The covariance operator CX : B∗ 7→ B can be extended to a continuous mapping from B∗µX to B. In order to show this, we use the fact thatEkXkp < ∞ for all1 ≤ p < ∞ which is a consequence of Fernique theorem. Let us define a mapping onB∗µX ×B∗by
f ,b˜b∗
7→Ef(X)b D X,˜b∗E
B,B∗
.
It is clear that this mapping is bilinear. Let us show that it is bounded. Using Cauchy- Schwartz inequality, we have:
|Efb(X)D X,˜b∗E
B,B∗
| ≤
Ef(X)b 212
EkXk2
B
˜b∗
2 B∗
12
≤C fb
µ
X
˜b∗ B∗
. (2.5) Since B is reflexive, then the covariance operator CX can be extended to a continuous mapping also denoted byCX fromB∗µX toB. That is,
D
CXf , bb ∗E
B,B∗
=Efb(X)hX, b∗i
B,B∗, ∀fb∈B∗µX, ∀b∗ ∈B∗.
Definition 2.4.3( [13, p. 60]). LetBbe a separable reflexive Banach space andX be a Gaussian B-valued zero mean random field with covariance operator CX : B∗µX → B. The Cameron-Martin space ofX is defined as the set
HµX =CX(B∗µX).
The inner product onHµX is defined as (f, g)µ
X = Z
f(b)b bg(b)µX(db),
and the corresponding norm is
(f, f)µ
X =kfk2µ
X. For allf ∈HµX orfb∈B∗µX, we denoteCXfb=f.
We have the following theorem that gives a fundamental property of the Cameron-Martin space.
Theorem 2.4.2 ( [13, p. 102]). Let X be a Gaussian random variable. The Cameron- Martin space and the spaceB∗µX of all measurable linear functional onBare separable Hilbert spaces.
2.5 White noise
Now, we are going to give the definition and some properties of Gaussian white noise.
Definition 2.5.1( [7, p. 102]). LetD⊂RdandB(D)be Borel sets ofD. We definewhite noiseW as a collection{W(A) :A∈B}of random variables with zero mean and finite covarianceE[W(A)W(B)] =A∩B, for allA , B ∈B(D).
The white noise W is Gaussian if each random variable has a normal distribution. The way to construct functional is through stochastic integral
W(A) = Z
1A(x)dWx, (2.6)
with respect tod-dimensional field which is actually a Gaussian field with zero mean and covariance
E[WxWy] = min(x1, y1)· · ·min(xd, yd), ∀x y∈Rd.
We can define white noise functional by replacing characteristic functions1Ain Equation (2.6) by functionsφ∈L2(D)such that
W(φ) = Z
φ(x)dWx. (2.7)
White noise defined by Equation (2.7) is a Gaussian random variable with zero mean and variance
EW(φ)W(φ) = E
|W(φ)|2
=E
Z
D
φ(x)dWx
2
=E
Z
D
|φ(x)|2dx
=kφk2L2(D).
In this work, we consider a version of white noise which actually has a representation W =
∞
X
k=1
Wkek, (2.8)
whereWkare Gaussian random variable such thatWk∼ N(0,1), and(ek)is an orthonor- mal basis ofL2(D).
We are going to prove some properties of white noise which actually has representation
(2.8). Let(ek)be an orthonormal basis of L2(D)andW = P∞
k=1Wkek. By Proposition 2.2.1, L2(D) is a separable reflexive Banach space and the covariance operator of W denoted byCW maps fromL2(D)toL2(D). Moreover, we have
hCWφ, φiL2 =EhW, φihW, φi=kφk2L2
and by Definition 2.4.3, it follows that the Cameron-Martin space of the white noiseW is L2(D).
Proposition 2.5.1. LetW be as in Equation(2.8). The following conditions are satisfied:
1. W has zero mean and covariance
EhW, φihW, ψi=hφ, ψiL2, for allφ, ψ∈L2. 2. Letφ ∈L2, we have
Eexp(ihW, φi) = exp
−1 2kφk2
.
Proof. The random variableW has zero mean, sinceWk∼ N(0,1)and we have
EhW, φi=E
∞
X
k=1
Wkφk =
∞
X
k=1
EWkφk = 0.
Forφ, ψ∈L2, we have
EhW, φi hW, ψi=E
"* ∞ X
k=1
Wkek, φ
+ * ∞ X
k=1
Wkek, ψ +#
=E
" ∞ X
k=1
∞
X
l=1
hek, φi hel, ψiWkWl
!#
=
∞
X
k=1
hek, φi hek, ψiEWk2 =hφ, ψi.
SinceWkhas Gaussian distribution, we have:
Eexp(ihW, φi) =
∞
Y
k=1
Eexp (ihWkek, φi)
=
∞
Y
k=1
∞
Z
−∞
√1
2πexp
−x2k 2
exp (ihxkek, φi)dxk
=
∞
Y
k=1
∞
Z
−∞
√1
2πexp
−x2
2 +ihxkek, φi
dxk
=
∞
Y
k=1
∞
Z
−∞
√1
2πexp
−x2k
2 +ihek, φixk
dxk
=
∞
Y
k=1
∞
Z
−∞
√1
2πexp
−x2k
2 +i(xkφk) + 1
2(hek, φi)2− 1
2(hek, φi)2
dxk
=
∞
Y
k=1
∞
Z
−∞
√1
2πexp (xk− hek, φi)2 exp
−1 2φ2k
dxk
=
∞
Y
k=1
exp
−1 2φ2k
Z∞
−∞
√1
2π exp (xk− hek, φi)2 dxk
= exp
−1 2kφk2
,
we have thus the proof.
Proposition 2.5.2. The white noiseW is aH−d2−(D)-valued Gaussian random variable.
Proof. We show the claim forD= [0,1]d. Let us show thatW belongs toH−d2−, for all >0according to Definition 2.2.5. LetWebe an extension ofW onRdsuch that
We=
∞
X
k=1
Wkek1[0,1]d.
Let us denote by FWe the Fourier transform ofW. We have thatFWe ∈ S0. Now we just need to show that
Z
Rd
1 +|ξ|2s
|FW(ξ)|2dξ <∞.
By the definition of Fourier transform, we have
FW(ξ) = hWe,exp(−iξ·x)i,
and Z
Rd
1 +|ξ|2s
|FW(ξ)|2dξ = Z
Rd
1 +|ξ|2s
|hWe,exp(−iξ·x)i|2dξ.
So, taking the expectation of this, we have
E
Z
Rd
1 +|ξ|2s
|FW(ξ)|2dξ
=E
Z
Rd
1 +|ξ|2s
|hWe,exp(−iξ·x)i|2dξ
= Z
Rd
1 +|ξ|2s E
|hWe,exp(−iξ·x)i|2 dξ
= Z
Rd
1 +|ξ|2s Z
[0,1]d
exp(−iξ·x) exp(iξ·x)dξdx
= Z
Rd
1 +|ξ|2s
dξ < C
∞
Z
0
1 +|r|2s
rd−1dr
< C0
∞
Z
1
r2s+d−1dr <+∞.
So whens <−d2, we can deduce Z
Rd
1 +|ξ|2s
|FW(ξ)|2dξ < +∞
henceWe belongs to Sobolev spaceHs, and in particular toH−d2−, for all > 0. Us- ing Pettis measurability theorem (see [13]), we deduce that W is a H−d2−(D)-valued Gaussian random variable.
3 EXISTENCE AND UNIQUENESS OF SOLUTION
Our main objective in this chapter is to show the existence and uniqueness of a Gaussian random fieldXsatisfying the following partial differential equation with Gaussian white noise
LX =W. (3.1)
3.1 Presentation of the equation
LetD ⊂Rdbe a open bounded domain. We define the differential operatorLonDby
L:=−∆ +ν2, (3.2)
where ∆ is the Laplacian operator defined for any function φ : D → R as ∆φ = Pd
i=1
∂2φ
∂x2i, andν > 0is a constant.
Definition 3.1.1( [8, p. 216]). LetHbe an Hilbert space. A linear operatorL:H →H isboundedif there existsM >0such that
kLφkH ≤MkφkH.
Definition 3.1.2( [8, p. 331]). LetH be an Hilbert space and h·,·ibe the inner product inH. A bounded linear operatorL:H →H is said to beself-adjointif
hLφ, ψi=hφ, Lψi, for allφ, ψ ∈H.
Definition 3.1.3( [8, p. 284]). LetHbe an Hilbert space. A linear operatorL:H →H iscompactif for every bounded sequence(uk)k≥1inH, the image sequence(Luk)k≥1has a convergent subsequence.
Theorem 3.1.1( [8, p. 348]). (Spectral Theorem)
LetV be an infinite dimensional, real Hilbert space. LetA:V →V be a continuous lin- ear, self-adjoint, compact operator. Then the eigenvalues ofAform a decreasing sequence (µk)k≥1of strictly positive reals which tends to zero and there exist an orthonormal basis (ψk)k≥1of eigenfunctions ofA.
Theorem 3.1.2( [14, p. 7]). (Lax-Milgram)
LetH be a Hilbert space and letabe a bilinear form onH×Hsuch that 1. ais bounded onH, that is,
|a(φ, ψ)| ≤CkφkHkψkH, for all φ, ψ∈H,
2. aiscoercive, that is, there existc >0such that
|a(φ, φ)| ≥ckφkH, for all φ∈H.
Then for anyf ∈H∗, there exist a unique solutionφ ∈Hsuch that a(φ, ψ) = chf, φiH, for all ψ ∈H.
Theorem 3.1.3( [10, p. 162]). (Poincaré Inequality)
LetD⊂Rdbe an open bounded domain. Then there exists a constantCD, depending on Dandpsuch that, for allφ∈H01(D)
kφkL2(D) ≤Ck∇φkL2(D).
Theorem 3.1.4( [10, p. 165]). (Rellich)
Let D ⊂ Rd be a bounded domain. The inclusion map form H01(D) into L2(D) is a compact operator.
Now, let D ∈ Rd be an open bounded domain. We consider the following variational eigenvalue problem that consist of findingλ∈Randφ∈H01(D)− {0}such that
a(φ, ψ) =λhφ, ψiL2, for all ψ ∈H01(D), (3.3) whereais a bilinear symmetric form given by
a(φ, ψ) = Z
D
∇φ(x)∇ψ(x)dx+ν2 Z
D
φ(x)ψ(x))dx.
Using Green’s formula, the above problem can be reformulated into the following
Lφ=λφ, inD, with φ= 0, on∂D (3.4)
that is λ and φ are eigenvalue and eigenfunction of L respectively. The solutions of Equation (3.3) can be given by the following theorem.
Theorem 3.1.5. The eigenvalues of Equation(3.3)form an increasing sequence(λk)k≥1
of positive reals which tends to infinity, and there exists an orthonormal basis(φk)k≥1 of L2(D)of associated eigenvectors, that is,
φk∈H01(D) and a(φk, ψ) =λkhφk, ψiL2, for all ψ ∈H01(D). (3.5)
Proof. Letf ∈L2(D). We define the following problem
Findφ ∈H01(D)such that a(φ, ψ) = hf, ψiL2, for all ψ ∈H01(D). (3.6) Using Theorem 3.1.2 we show that Equation (3.6) admits a unique solutionφ ∈ H01(D).
If we takeψ =φin Equation (3.6), we have
φ∈H01(D) such that a(φ, φ) = hf, φiL2, for all φ ∈H01(D).
Set
T φ =hf, φiL2.
The bilinear formais continuous and bounded inH1. Moreoverais coercive since
|a(φ, φ)|=|T φ|=|hf, φiL2|= Z
D
f φdx
≤ kfkL2kφkL2 ≤CkfkL2kφkH1
0,
since the embedding ofH01(D)intoL2(D)is continuous by Rellich Theorem. It follows thatT andaare continuous. Furthermore, using Poincaré Inequality we have
|a(φ, φ)|= Z
D
|∇φ|2dx+ν2 Z
D
|Af|2dx=k∇φk2L2 +ν2kφk2L2
≥ k∇φk2L2 ≥ckφk2H1 with c >0.
Henceais coercive. Let us denote byφ =Af the unique solution of (3.6), whereAis a linear operator defined fromL2(D)intoH01(D)as follows:
Af ∈H01(D), such that a(Af, ψ) = hf, ψiL2, for all ψ ∈H01(D). (3.7) Since the operator A : L2(D) → H01(D)and by Rellich Theorem H01(D) is compactly embedded into L2(D), then it follows that the operator A : L2(D) → L2(D) and is compact. To show thatAis self-adjoint, let us takeψ =Af in Equation (3.7). Then we have
hf, AgiL2 =a(Af, Ag) = a(Ag, Af) =hg, AfiL2
sinceais symmetric. Since all the hypothesis of Theorem 3.1.1 are satisfied, we can now apply it and we have that the eigenvalues of A form a decreasing sequence (µk)k≥1 of strictly positive reals which tends to zero and there exist an orthonormal basis(ψk)k≥1 of
eigenfunctions ofA, with
Aψk=µkψk.
Coming back to Equation (3.3), we have
a(φ, ψ) =λhφ, ψiL2 =λa(Aφ, ψ), for all ψ ∈H01(D), so,
a(φ−λAφ, ψ) = 0 ⇒φ=λAφ⇒Aφ= 1 λφ.
Hence the eigenvalues(λk)k≥1 from Equation (3.3) form an increasing sequence consist- ing of the inverses of eigenvalues(µk)k≥1 ofA. The eigenfunctions are exactly the same as(φk)k≥1 and form an orthonormal basis ofL2(D). Thus, the theorem is proved.
We have then shown that Ladmits an orthonormal basis of eigenfunctions, which form a complete orthonormal set ofL2(D). This complete set of orthonormal eigenfunctions form a basis for the infinite dimensional Hilbert space. Let us denote the eigenfunctions of Lbyekunder Dirichlet conditions, and let us denote byλkthe corresponding eigenvalues.
ThenLmust satisfy the following problem
Lek =λkek.
Next, we consider the Gaussian white noise W that we have presented in the previous chapter which has the following representation
W =
∞
X
k=1
Wkek, (3.8)
where Wk ∼ N(0,1) are independent, and (ek) is a complete orthonormal basis in a Hilbert space.
Based on all what we have defined so far, we are now able to reformulate our main objec- tive and thus we have the following problem:
Problem 3.1.1. Show the existence and uniqueness of a Gaussian random fieldXwhich is solution of the equation
LX =W,
in the eigenbasis ofL.
3.2 Existence and uniqueness
In this subsection, we are interested solving Problem 3.1.1. We are going to solve the problem in dimensions one and two.
Lemma 3.2.1. LetD ⊂Rdbe a Lipschitz domain. LetHµX be the Cameron-Martin space of a zero meanH−r-valued Gaussian random variableX, such thatHµX is continuously embedded into H1(D). Then X has realizations in H1−
d
2−(D) almost surely, for all >0.
The following result shows the existence of a unique random field solution of Problem 3.1.1.
Theorem 3.2.1. LetD ⊂ R be a bounded open interval. LetW be the Gaussian white noise define in Equation (3.8). Then there exists a zero mean H−r(D)-valued Gaussian random fieldX, withr > 12 −1 = 12 such that
1. The Cameron-Martin space of X denoted by HµX(D) is continuously embedded intoH1(D)and for allh∈HµX(D)we have∆h∈L2(D).
2. The operator L defined in Equation (3.2) has eigenspace representation and the solution of Problem(3.1.1)and can be represented as eigenspace expansion.
Moreover,H−r(D)-valued Gaussian random fieldX is unique.
Proof. We prove the theorem for dimensionsd = 1,2. The prove can be done similarly for dimension for d ≥ 3. Let D = (0,1) ⊂ R. We have shown that L admits an orthonormal basis of eigenfunctions(ek)andLmust satisfy the following problem:
Lek =λkek.
Let us consider the case of Dirichlet boundary condition, the eigenvalues and eigenfunc- tions ofLcan be computed as follow. We know that
Lek(x) = (−∆ +ν2)ek(x) =λkek(x) x∈D= (0,1).
So
∆ek(x) +αek(x) = 0, with α=λ−ν2 >0, (3.9) since under Dirichlet boundary condition, the Equation (3.9) has solution only whenα >
0. Letα =b2, the characteristic equation is
r2+b2 = 0⇒r=±ib.
Figure 1.Dirichlet eigenfunctions of operatorLin dimension one.
The general solution is given by
ek(x) =c1exp(ax) cos(bx) +c2exp(ax) sin(bx),
whereais the real part ofr. Applying boundary conditionek(0) =ek(1) = 0we obtain ek(x) = sin(kπx) and bk=kπ, kinteger.
Hence we deduce that eigenfunctions and eigenvalues ofLare defined by:
ek(x) =√
2 sin(kπx) and λk= (kπ)2+ν2, kinteger andν > 0. (3.10) Eigenfunctions of operator L in dimension one for fixed value of parameter ν are pre- sented in Figure 1.
We consider the expansion ofXin term of normed eigenfunction as X(x) =
+∞
X
k=1
Xkek(x), (3.11)
and we writeLin its eigenspace as LX(x) =
+∞
X
k=1
XkLek(x) =
+∞
X
k=1
Xkλkek(x) =
+∞
X
k=1
√2Xk[(kπ)2 +κ2] sin(kπx).
Furthermore, the Gaussian white noise defined by Equation (3.8) can also be represent in term of eigenfunctions as
W =
+∞
X
k=1
√
2·Wksin(kπx).
We have
LX(x) =W(x)⇒
+∞
X
k=1
√
2Xk[(kπ)2+ν2] sin(kπx) =
∞
X
k=1
√
2Wksin(kπx).
Since, eigenfunctionsek(x) = sin(kπx),k integer are orthonormal inL2(D), taking the inner product with a fixed eigenfunctioneko(x), we have
Xk[(kπ)2+ν2] =Wk.
It follows that
Xk= Wk
(kπ)2+ν2, (3.12)
which is well defined since ν > 0andWk ∼ N(0,1). Substituting Equation (3.18) into Equation (3.11), we obtain
X(x) =
+∞
X
k=1
√2·Wk
(kπ)2+ν2 sin(kπx) (3.13) In order to show the uniqueness of a Gaussian random fieldX, we assume that there exists two Gaussian random fieldsX1andX2 which are solutions of Problem 3.1.1. That is,
LX1 =W andLX2 =W.
Then we have
L(X1−X2) = 0, almost surely. (3.14) It is known from deterministic problem
LX =f, withf ∈L2(D)
Figure 2.Gaussian random fieldXin dimension one.
that there exists a unique solution, for any function. Since0is a function ofL2(D), we then have the that the solutionX =X1−X2 of Equation (3.14) is unique. Then
X1−X2 = 0, thus X1 =X2 almost surely.
The following Figure 2 shows the zero mean Gaussian random field for a one dimensional Problem 3.1.1 and Figure 3 shows its covariance for a fixed value of parameterν.
We are now going to solve Problem 3.1.1 in two dimensional case. We mainly prove Theorem 3.2.1 by consideringD = (0,1)2 ⊂ R2. We recall that in dimension two, that Laplace operator is defined as
∆ek(x, y) = ∂ek(x, y)
∂x2 + ∂ek(x, y)
∂y2 .
From the same idea as in dimension one, we have
∆ek(x, y) +αek(x, y) = 0, with α =λ−ν2 >0, (x, y)∈D= (0,1)2. (3.15)
Figure 3. Covariance of a Gaussian random fieldXin dimension one.
Letting
ek(x, y) =uk(x)vk(y),
plugging into Equation (3.15) and separating gives u00k
uk
+ vk00 vk
=γ,
which can be true only if each term on the left hand side is a constant. So we have u00k+γxuk = 0 and vk00+γyvk = 0,
withγx andγy constant. Solving these two equations as in dimension one with Dirichlet boundary conditions
uk(0) = 0 =uk(1) and vk(0) = 0 = vk(1), and using the fact that
ek(x, y) =uk(x)vk(y),
we get eigenfunctions
ek(x, y) = 2·emk,lk(x, y) = sin(mkx) sin(lky),
and eigenvalues
γk =γmk,lk = (mkπ)2+ (lkπ)2+ν2.
Figure 4. Dirichlet eigenvalues of operatorLin dimension two.
Figure 4 and Figure 5 show respectively the Dirichlet eigenvalues and eigenfunctions of operatorLin dimension two for fixed value of parameterν.
We consider the expansion ofXin term of eigenfunctions as X(x, y) =
+∞
X
k=1
Xkek(x, y), (3.16)
and we writeLin its eigenspace as LX(x, y) =
+∞
X
k=1
XkLek(x, y) =
+∞
X
k=1
Xkλkek(x, y)
=
+∞
X
k=1
2·Xk[(mkπ)2+ (lkπ)2+ν2] sin(mkπx) sin(lkπ).
Furthermore, the Gaussian white noise representation defined by (3.8) in dimension two is
W =
+∞
X
k=1
2·Wksin(mkπx) sin(lkπy).
Figure 5. Dirichlet eigenfunctions of operatorLin dimension two.
We have
LX(x, y) = W(x, y),
which implies
+∞
X
k=−∞
Xk[(mkπ)2+ (lkπ)2+ν2] sin(mkπx) sin(lkπy) =
∞
X
k=1
Wksin(mkπx) sin(lkπy).
(3.17) Since, eigenfunctionsek(x, y) = sin(mkπx) sin(lkπx), mk and lk integer, are orthonor- mal inL2(D), taking the inner product with a fixed eigenfunctioneko(x, y), we have
Xk[(mkπ)2+ (lkπ)2+ν2] =Wk,
It follows that
Xk = Wk
(mkπ)2+ (lkπ)2+ν2, (3.18) which is well defined since ν > 0andWk ∼ N(0,1). Substituting Equation (3.18) into Equation (3.16), we obtain the solution
X(x, y) =
+∞
X
k=−∞
2·Wk
(mkπ)2+ (lkπ)2+ν2 sin(mkπx) sin(lkπy). (3.19)
Figure 6. Gaussian random fieldXin dimension two.
Figure 7.Covariance of a Gaussian random fieldXin dimension two.
By the same way as in dimension one, we prove that the Gaussian random field X is unique. The zero mean Gaussian random field solution of Problem 3.1.1 and its covari- ance are illustrated in the following Figure 7 for fixed value of parameterν.
4 BAYESIAN ESTIMATION
In this section, we develop the Bayesian estimation approach that arises in Bayesian ap- proximation for inverse problems. We follow the same approach as in [4, 15, 16]. The Bayesian approach to inverse problem consist of determined the unknown from the math- ematical model and finite-dimensional noisy measurement.
In the Bayesian approach, the solution of the inverse problem isposterior distribution and it is determined with the knowledge of the measurement, the mathematical model and the prior distribution. The posterior distribution gives us the degree of confidence that any candidate solution might be the true function. Theprior contains our beliefs about the unknown that we wish to update with the help of new information.
We start our discussion by considering the linear inverse problem that can be modelled in finite dimensions as
Y =GX+E, (4.1)
where, Y ∈ RM is a finite dimensional vector of measurement, the functionG : RN → RM is an operator mapping the unknown Gaussian random variable X ∈ RM into the measurement in the absence of measurement noise. We suppose that the measurement noiseE is independent of the unknownXand such thatE ∼ N(0, C), whereCis its the covariance matrix.
4.1 Bayes theorem
We recall that the Bayes formula from probability theory, that gives the conditional prob- ability of an assumptionAknowing assumptionB. LetP be a probability measure given in Definition 2.3.1. LetAbe the assumption andB be that observed measurements under assumptionA. Then the Bayes formula is given by
P(A|B) = P(B|A)P(A)
P(B) , withP(B)6= 0. (4.2) We are going to adapt this version of Bayes formula to give an expression of the posterior distribution of the unknownXin Equation (4.1). We will now denote withπa probability density function (pdf) of a distribution of a random vector.
Theorem 4.1.1. LetX be a Gaussian random field which has a known prior distribution π(X), let Y be a vector of measurements such thatπ(Y) >0. Then the posterior pdf of
X given the measurementY is given by
π(X|Y) = π(X)π(Y|X) π(Y) .
Note thatπ(Y)does not depend onX and therefore can just be considered as a normal- ization constant. The posterior probability distribution becomes
π(X|Y)∝π(Y|X)π(X), (4.3)
where the symbol ” ∝ ” means "is proportional to". The probability of observing the measurement Y, given the random field X defines the likelihood. Based on Equation (4.1), the likelihood can be written as
π(Y|X)∝exp
−1
2(Y −GX)TC−1(Y −GX)
.
SinceX is a Gaussian random variable, we choose the prior distribution to be Gaussian with a given zero mean and covariance matrixΣ. The prior density is then written as:
π(X)∝exp
−1
2XTΣ−1X
.
It follows from Equation (4.3) that the unnormalized posterior distribution is:
π(X|Y)∝exp
−1
2(Y −GX)TC−1(Y −GX)−1
2XTΣ−1X
. (4.4)
Let’s simplify Equation (4.4) and write it in a form of a Gaussian probability distribution function. Let
Q:=−1
2(Y −GX)TC−1(Y −GX)− 1
2XTΣ−1X.
We have Q=−1
2(Y −GX)TC−1(Y −GX)−1
2XTΣ−1X
=−1
2YTC−1Y +1
2XTGTC−1Y + 1
2XTGTC−1Y − 1
2XTGTC−1GX −1
2XTΣ−1X
=−1
2YTC−1Y +1
2XTGTC−1Y + 1
2YTC−1GX − 1
2XT(GTC−1G+ Σ−1)X If we denote by
Cpost = (GTC−1G+ Σ−1)−1, (4.5)