Computational Engineering and Technical Physics Techno-Mathematics
Arnold Kaynet Muchatibaya
BACKWARD STOCHASTIC DIFFERENTIAL EQUATIONS WITH APPLICATIONS
Master’s Thesis
Examiners: Professor Simo S¨arkk¨a Professor Heikki Haario Supervisors: Professor Simo S¨arkk¨a
Professor Heikki Haario
Lappeenranta University of Technology School of Engineering Science
Computational Engineering and Technical Physics Techno-Mathematics
Arnold Kaynet Muchatibaya
Backward Stochastic Differential Equations with Applications
Master’s Thesis 2018
60pages.
Examiners: Professor Simo S¨arkk¨a Professor Heikki Haario
Keywords: Stochastic differential equations, Backward stochastic differential equations, Doob’s h-transform, Feynman-Kac, L´evy process
In this thesis we study backward stochastic differential equations driven by a Brownian motion and by a L´evy process and their applications, focusing on their applications to fi- nancial markets. We give results on the existence and uniqueness of solution of backward stochastic differential equations when the drift is Lipschitz continuous and the termi- nal condition is square integrable and measurable with respect to the terminal filtration.
Backward stochastic differential equations associated with a forward stochastic differen- tial equation are investigated. We use the generalisation of the Feynman-Kac formula to show the relationship between a backward stochastic differential equation associated with a forward stochastic differential equation and a partial differential equation in the Brown- ian motion case and a partial differential integral equation for the L´evy process case. The Doob’s h-transform is studied for the Brownian motion and applied to stochastic differ- ential equations. Finally, we conclude with an application to option pricing and hedging of a European calls for both Brownian and L´evy processes.
All glory and praise be given to the most High for giving me the life and sustaining me during my studies until the completion of this thesis. I wish to thank my supervisors Professor Simo S¨arkk¨a and Professor Heikki Haario for their help, resources and super- vision throughout the preparation of this thesis. I would also like to thank my family for the strength and support during my studies. Finally, thank you to African Institute for Mathematical Sciences (AIMS-Tanzania) and Lappeenranta University of Technology for arranging and awarding of the scholarship to undertake studies at the later.
Lappeenranta, October 19, 2018
Arnold Kaynet Muchatibaya
CONTENTS
1 INTRODUCTION 6
1.1 Objectives . . . 8
1.2 Outline . . . 9
2 BACKWARD STOCHASTIC DIFFERENTIAL EQUATIONS 10 2.1 Preliminaries . . . 10
2.2 Background . . . 21
2.3 Existence and Uniqueness . . . 23
2.3.1 Brownian Motion . . . 23
2.3.3 Levy Process . . . 33
2.4 Feynman-Kac Theorem . . . 37
2.5 Doob’s h Transform . . . 47
3 APPLICATION 51 3.1 Model 1 . . . 51
3.2 Model 2 . . . 53
4 DISCUSSION 56 4.1 Future Work . . . 57
5 CONCLUSION 58
REFERENCES 59
ABBREVIATIONS AND SYMBOLS
These are the abbreviations we will use throughout this essay:
a.s. Almost surely C Complex numbers A∈B Element A in set B
∀ For all
⇒ Implies
h·,·i Inner product 7→ Maps to
|| · || Norm of (·) R Real numbers
∃ There exists
∪ Union of sets w.r.t with respect to
BSDE Backward stochastic differential equation EMM Equivalent martingale measure
FPK Fokker-Planck-Kolmogorov
FBSDE Forward-backward stochastic differential equation FSDE Forward stochastic differential equation
ODE Ordinary differential equation PDE Partial differential equation
PDIE Partial differential integral equation SDE Stochastic differential equation
1 INTRODUCTION
Allowing randomness in the coefficients of an ordinary differential equation results in realistic mathematical models of physical phenomena. Stochastic differential equations arise when we allow randomness in the coefficients of ordinary differential equations or when the forcing is an irregular stochastic process like Gaussian white noise [1]. There are two types of stochastic differential equations; if the initial condition is specified, we have a forward stochastic differential equation and if the terminal condition is specified, we have a backward stochastic differential equation.
The theory of backward stochastic differential equations has found wide applications in areas such as stochastic optimal control, theoretical economies, and mathematical finance problems such as the theory of hedging and non-linear pricing in incomplete markets [2].
Backward stochastic differential equations can be driven by a L´evy process, Brownian motion, Poisson process, or a combination of these.
Bismut [3] first introduced backward stochastic differential equations in a linear form as the equation for the conjugate variable in the stochastic version of the Pontryagin maxi- mum principle. Pardoux et al. [4] were the first to consider general backward stochastic differential equations. Their main result was the existence and uniqueness of an adapted pair of processes as a solution of a backward stochastic differential equation.
Several authors have extended their results. Peng [5] used backward stochastic differen- tial equations to obtain a probabilistic interpretation for systems of second order quasi- linear parabolic partial differential equations. Pardoux et al. [6] introduced a new class of backward stochastic differential equations, which allowed them to produce a probabilistic representation of a certain quasi-linear stochastic partial differential equation thus ex- tending the Feynman-Kac formula for stochastic partial differential equations. Antonelli [7] showed the existence and uniqueness of a solution of a backward stochastic differen- tial equation inspired from the stochastic differential utility in finance theory. Ma et al.
[8] investigated adapted solutions to a class of forward-backward stochastic differential equations in which the forward stochastic differential equation is non-degenerate. They showed that the adapted solution can be sought over an arbitrarily prescribed time du- ration via a direct four step scheme. Using this scheme, they proved that the backward components of the adapted solution are determined explicitly by the forward component via the solution of a certain quasi-linear parabolic partial differential equation system.
El-Karoui et al. [9] summarized the existence and uniqueness of solutions of backward
stochastic differential equations by Pardoux et al. [4] and gave new shorter proofs. They stated the a priori estimates of the difference of two backward stochastic differential equa- tions, and the uniqueness and existence was proved using a fixed point theorem. They also looked at the solution of a backward stochastic differential equation associated with a for- ward stochastic differential equations. The main property was that the solution is Marko- vian in the sense that it can be written as a function of time and state process. The general- isation of the Feynman-Kac formula is given, and they also showed that under smoothness assumptions, the solution of the backward stochastic differential equation corresponds to a solution of a system of quasi-linear parabolic partial differential equations. These re- sults could be applied to option pricing of a European call in the constrained Markovian cases.
Buckdahn and Pardoux [10] proved the existence and uniqueness of a solution to a back- ward stochastic differential equation with respect to both the Brownian motion and Pois- son random measure and the associated integro-partial differential equation of parabolic type. They proved that under certain conditions, the solution of a backward stochastic dif- ferential equation provides the unique viscosity solution of the associated integro-partial differential equation. Situ [11] studied backward stochastic differential equations driven by Brownian motion and Poisson point process. A new existence and uniqueness result for the solution of the partial differential integral equation with non Lipschitz force is obtained. Oukine [12] considered a backward stochastic differential equation driven by a Poisson random measure. The integral representation of the square integral random variable in terms of a Poisson random measure is the main result.
Nualart and Schoutens [13] proved a martingale representation theorem for a L´evy pro- cess satisfying some exponential moment condition. Nuarlart et al. [14] used results from [13] to establish existence and uniqueness of solutions for backward stochastic differential equations driven by a L´evy process. Our work is primarily based on [9] and [14].
1.1 Objectives
This thesis is based on the articles by Nualart et al. [13] and El-Karoui et al.[9]. We expand the proof studied in the articles, especially the existence and uniqueness of so- lutions of a BSDE under Lipschitz conditions on the drift driven by a L´evy process and Brownian motion. We are also concerned with the application of these BSDEs in finance.
To achieve our purpose, we consider some specific objectives:
(i) We study the existence and uniqueness of a general BSDE under the Lipschitz con- dition on the drift driven by a L´evy process and Brownian motion expanding the proofs in these sections.
(ii) Apply the Feynman-Kac formula to the BSDEs to get the relationship between BSDEs, PDEs and PDIEs.
(iii) Consider the application of the theory above to European call options.
(iv) We study the Doob’s h-transform applied to an SDE in the Brownian Motion case.
1.2 Outline
In Chapter2, we study the existence and uniqueness of a solution of BSDEs with a Lips- chitz driver and driven by a L´evy process and Brownian motion in Section2.3. We also consider the Feynman-Kac formula for the BSDEs to establish the relationship with the partial differential equation in the case of BSDE driven by Brownian motion and the par- tial differential integro equation in the case of the BSDE driven by a L´evy process in Section2.4. In Section2.5we study the Doob’s h-transform and application to stochastic differential equations to come up with BSDE. In Chapter 3, we look at the application of these studied BSDEs to option pricing. In Chapter 4, we discuss the results we have obtained in our work and mention possible future work to be done. Finally we conclude our work in Chapter5.
2 BACKWARD STOCHASTIC DIFFERENTIAL EQUA- TIONS
2.1 Preliminaries
In this section we will give the background of SDEs, theorems and inequalities which will be necessary to refer in the forthcoming chapters. We start by defining the space we will be working on,
Definition 1(Probability space). IfΩis a given set, then aσ-algebraF onΩis a family F of subsets ofΩwith the following properties:
(i) ∅ ∈ F,
(ii) F ∈ F ⇒Fc ∈ F whereFcis the complement ofF inΩandFc = Ω\F,and (iii) A1, A2,· · · ∈ F ⇒ A:=S∞
i=1∈ F.
Then the pair(Ω,F)is called a measurable space. A probability measurePon(Ω,F)is a functionP:F 7−→[0,1]such that
(i) P(∅) = 0, P(Ω) = 1,and
(ii) ifA1, A2,· · · ∈ F and{Ai}∞i=1is disjoint, then
P
∞
[
i=1
!
=
∞
X
i=1
P(Ai).
Then the triplet(Ω,F,P)is a probability space [15].
A martingale is a stochastic process for which at a time in the realised sequence the expectation of the next value is the current observed value given prior observation. Now we give the mathematical definition is as follows.
Definition 2(Martingale). A filtration on(Ω,F)is a familyF = {Ft}t≥0 ofσ-algebras Ft⊂ F such that
0≤s≤T ⇒ Fs ⊂ Ft.
An n-dimensional stochastic process {Xt}t≥0 on (Ω,F,P) is called a martingale [15]
with respect to a filtration{Ft}t≥0(and with respect toP) if
(i) XtisFt−measurable for allt, (ii) E[|Xt|]<∞for allt,and (iii) E[Xs|Ft] =Xta.s. for alls ≥t.
Now let us define a L´evy process.
Definition 3 (L´evy process). Given the probability space in Definition 1, a L´evy pro- cessX = {Xt, t ≥ 0}taking values inRd is a stochastic process having stationary and independent increments and we always assumeX0 = 0with probability1.So
• Xt: Ω−→Rd.
• Given any selection of distinct time points 0 ≤ t1 < t2 < · · · < tn the random vectorsXt1, Xt2 −Xt1, Xt3 −Xt2,· · · , Xtn−Xtn−1 are all independent.
• Given any two distinct times0≤s < t <∞,the probability distribution ofXt−Xs
coincides with that ofXt−s.[16]
Brownian motion and Poison process are examples of L´evy processes. Thus we have,
Definition 4 (Brownian motion). A standard Brownian motion in Rd is a L´evy process W = (Wt, t ≥0)for which
• Wt∼N(0, tI)for eacht≥0
• W has continuous sample paths
Every L´evy process is characterised by its characteristic function which is defined as follows [7].
Definition 5(Characteristic function). LetXbe a random variable defined on the proba- bility space in Definition1taking values inRdwith the probability lawPx.It’s character- istic functionφx :Rd−→Cis
φx(u) =E ei(u,x)
= Z
Ω
ei(u,X(ω))P(dω)
= Z
Rd
ei(u,y)Px(dy)
for eachu∈Rd
We define the indicator functionχA as a function defined to be1onAand0elsewhere.
Now the characteristic function is given by the L´evy -Khintchine formula defined below as.
Definition 6(L´evy-Khintchine formula). IfX = {Xt, t ≥ 0}is a L´evy process, then it has a specific form for its characteristic function [16]. More precisely∀t≥0, u∈Rd
E(ei(u,Xt)) =etη(u) where
η(u) =i(b, u)− 1
2(u, au) + Z
Rd−{0}
ei(u,y)−1−i(u, y)χ0<|y|<1(y)
ν(dy).
where
• b∈Rd
• ais a positive definite symmetricd×dmatrix
• νis a L´evy measure onRd− {0}so that Z
Rd−{0}
min{1,|y|2ν(dy)}<∞.
A L´evy process can be decomposed into a linear drift, Brownian motion, and a pure jump process [16]. This result is called the L´evy-Itˆo decomposition and defined as follows.
Theorem 2.1.1(The L´evy-Itˆo decomposition). IfX is a L´evy process, then there exists b ∈ Rd,a Brownian motionB with diffusion matrixQand an independent Poisson ran- dom measureN onR+×(Rd− {0})such that for eacht ≥0,
Xt =bt+Bt+ Z
|x|<1
xN˜(t, dx) + Z
|x|>1
xN(t, dx).
Proof. See [16]
Definition 7. Let us consider the SDE,
dXt=b(t, x)dt+σ(t, x)dWt
X(0) =x, t≥0.
A strong solution of this SDE on the given probability space with respect to the fixed Brownian motion W and initial condition x is a process X = {Xs; 0 ≤ s ≤ T} with continuous sample paths and with the following properties:
(i) Xis adapted to the filtrationFs, (ii) P[X0 =x] = 1,
(iii)
P
T
Z
0
{|b(s, Xs)|+σ2ij(s, Xs)}ds <∞
= 1
holds for every1≤i≤d,and (iv) The integral version is
Xs =X0+
T
Z
0
b(s, Xs)ds+
T
Z
0
σ(s, Xs)dWs.
[17].
For any martingale adapted with respect to a Brownian motion can be expressed as an Itˆo integral with respect to the same Brownian motion as follows.
Theorem 2.1.2(Martingale representation). Let(Wt,0≤ t≤ T)be a Brownian motion on(Ω,F,P). Let{Ft; 0 ≤ t ≤ T}be the filtration generated by this Brownian motion.
Let{Xt; 0≤t ≤T}be a martingale (underP) relative to this filtration (i.e., for everyt, Xtis Ft−measurable, and for 0 ≤ s ≤ t ≤ T,E[Xt|Fs] = Xs a.s.). Then there is an adapted process{At; 0≤t≤T},Atsquare integrable such that
Xt =X0+
t
Z
0
AudWu, 0≤t ≤T.
Proof. See [15]
The Burkholder-Davis-Gundy inequalities relate the maximum of a local martingale with it’s quadratic variation. This result is important in the proofs in the next chapters.
Theorem 2.1.3 (Burkholder-Davis-Gundy inequalities). Let T > 0 and (Mt)0≤t≤T be a continuous local martingale such that M0 = 0. For every 0 < p < ∞, there exists universal constantscp, Cp independent ofT and(Mt)0≤t≤T such that,
cpE hMti
p 2
T
≤E
sup
0≤t≤T
|Mt| p
≤CpE hMti
p 2
T
.
Proof. See [17].
The Banach fixed point theorem is important in the proof of existence and uniqueness of solution of a BSDE, hence we first define a metric space then give theorem.
Definition 8. Let (X, d) be a metric space. A mapping T : X 7−→ X is Lipschitz continuous if there exists a constant α > 0 such that d(T x, T y) ≤ α d(x, y) for all x, y ∈X.If0≤α <1, thenT is called a contraction mapping, andαis called the factor ofT [18].
Theorem 2.1.4(Banach fixed point theorem). Suppose that(X, d)is a generalised com- plete metric space, and that the functionT :X 7−→X is a contraction.
Letx0 ∈X,and consider the sequence of successive approximations with initial element x0
x0, T x0, T2x0,· · · , Tix0,· · ·. (1) Then either
1. For every integeri= 0,1,2,· · ·,one has
d(Tix0, Ti+1x0) =∞, or
2. The sequence of approximations, Equation(1)isd-convergent to a fixed point ofT.
Proof. See [18]
We need to define anLp space before the H¨older inequality as follows.
Definition 9. Consider the measurable space in Definition 1 and 1 ≤ p, q ≤ ∞. The spaceLp(Ω)consists of equivalence classes of measurable functionsf : Ω7→Rsuch that
Z
|f|pP(dω ∈ F)<∞,
where ω ∈ Ω and two measurable functions are equivalent if they are equalP a.e [19].
TheLp norm off ∈L(Ω)is defined by
||f||Lp = Z
|f|pP(dω∈ F) 1p
. Whenp= 1the spaceL1consists of all integrable functions onΩ.
Theorem 2.1.5(H¨older inequality). Assume that a functionf ∈ Lp andg ∈ Lq,where p, q∈(1,∞)are conjugate numbers, that is,
1 p+ 1
q = 1.
Thenf g ∈L1,and the following inequality holds
Z
f g dx
≤ Z
|f g|dx≤ ||f||p||g||q. (2)
Proof. See [19]
Theorem 2.1.6(Markov inequality). Suppose(Ω,F,P)is a measure space,f is a mea- surable extended real valued function, andε >0.Then
P({ω∈Ω :|f(ω)| ≥ ε})≤ 1 ε
Z
Ω
|f|P(dω ∈ F).
Proof. See [19]
Definition 10(Itˆo process). LetWtbe a one-dimensional Brownian motion on(Ω,F,P).
An Itˆo process (or stochastic integral) is a stochastic processXton(Ω,F,P)of the form
Xt=X0+
t
Z
0
b(s, ω)ds+
t
Z
0
σ(s, ω)dWs,
where
P
t
Z
0
σ(s, ω)2ds <∞for allt≥0
= 1,
and
P
t
Z
0
|b(s, ω)|ds <∞for allt≥0
= 1.
[15]
We define the quadratic variation and cross variance as follows.
Definition 11. IfXt(·) : Ω7−→ Ris a continuous stochastic process, then forp > 0the p’th variation process ofXt;hX, Xi(p)t is defined by
hX, Xi(p)t (ω) = lim
△tk→0
X
tk≤t
|Xtk+1(ω)−Xtk(ω)|p(limit in probability)
where0 = t1 < t2 < · · ·< tn = tand△tk =tk+1−tk. Ifp = 1then it is called total variation and ifp= 2is called quadratic variation. We also have
hX, Xi(p)t =hX, Xit=hXit.
More generally for the cross variation between two processesXtandYtwe have hX, Yit(ω) = lim
△tk→0
X
tk≤t
(Xtk+1(ω)−Xtk(ω))(Ytk+1(ω)−Ytk(ω)).
[15].
Let us consider the Itˆo formula for Brownian motion and L´evy process driven Itˆo process as follows.
Theorem 2.1.7 (Itˆo formula). Let Xt be an Itˆo process, and let f(t, x) be a function for which the partial derivativesft, fx, fxx are defined and continuous. Then for every T ≥0,
f(T, Xt) =f(0, X0) +
T
Z
0
ftdt+
T
Z
0
fxdXt+ 1 2
T
Z
0
fxxdhXit. (3) Letf(t, x, y)be a function whose partial derivativesft, fx, fy, fxx, fxy, fyy are defined and are continuous. LetXtandYtbe Itˆo processes. The two dimensional Itˆo formula in differential form is
df(t, x, y) =ftdt+fxdXt+fydYt+1
2fxxdhXit+fxydhX, Yit+ 1
2fyydhYit. (4) Proof. See [20]
Theorem 2.1.8 (Itˆo formula for L´evy process driven SDEs). Let X = (X1,· · · , Xn) be an n-tuple of semi martingales and let f : Rn 7−→ R have continuous second order partial derivatives. Thenf(x)is a semi-martingale and the following formula holds
f(Xt)−f(X0) =
n
X
i=1 t
Z
0+
∂f
∂xi
(Xs−)dXs(i)+1 2
X
1≤i,j≤n t
Z
0+
∂2f
∂xi∂xj
(Xs−)d[X(i), X(j)]cs
+ X
0<s≤t
(
f(Xs)−f(Xs−)−
n
X
i=1
∂f
∂xi
(Xs−)△Xs(i) )
Proof. See [21]
Theorem 2.1.9 (Doob’s maximal inequality). Let (Xt; Ft, 0 ≤ t < ∞) be a semi- martingale whose every path is right-continuous. Letα < β be real numbers, and[0, T] is a sub interval of[0,∞).Then
E
sup
0≤t≤T
Xt
p
≤ p
p−1 p
E(XTp), p >1, providedXt≥0a.s. Pfor everyt≥0andE(XTp)<∞.
Proof. See [17]
Theorem 2.1.10(Gronwall’s inequality). Letg(t)andh(t)be regular non-negative func- tions on[0, T].Then for any regularf(t)≥0satisfying the inequality for all0≤t≤T,
f(t)≤g(t) +
t
Z
0
h(s)f(s)ds,
we have
f(t)≤g(t) +
t
Z
0
h(s)g(s) exp
t
Z
s
h(u)du
ds. (5) In particular, if g is non-decreasing, Equation(5)simplifies to give
f(t)≤g(t)eR0th(s)ds. In its simplest form wheng =Aandh=B are constants,
f(t)≤AeBt.
Proof. See [22]
Theorem 2.1.11 (Fubini’s theorem). Let f(x, t) be continuous on [t, T]×[0, b]. Then iterated integrals:
T
Z
t b
Z
0
f(x, t)dx dt=
b
Z
0 T
Z
t
f(x, t)dt dx.
Proof. See [23]
We consider the Fubini theorem for stochastic processes. We first of all define an FV process and some notation we will use for the theorem.
Definition 12. An FV process is a cadlag adapted stochastic process such that all its paths are of finite variation on each compact interval onR+[21].
The following Notation is by [21]. Notation: Let A be an FV process and letF be jointly measurable process such that
t
Z
0
F(s, ω)dAs(ω)
exists and is finite∀t >0,a.s. we let
(F ·A)t(ω) =
t
Z
0
F(s, ω)dAs(ω).
We also writeF ·Ato denote the processF ·A= (F ·At)t≥0. Then the Fubini theorem is as follows.
Theorem 2.1.12(Fubini’s theorem (stochastic processes)). LetX be a semi-martingale, Hta =H(a, t, ω)be a boundedA ⊗Pmeasurable function, and letµbe a finite measure on A. Let Zta = Rt
0 HsadXs be A ⊗ B(R+)⊗ F measurable such that for each a, Za is c´adl´ag version of Ha·X.Then Yt = R
AZtaµ(da)is c´adl´ag version of H ·X, where Ht=R
Aµ(da)[21].
The Itˆo isometry is essential for computation of variances of random variables given as an Itˆo integral. The following theorem is by [20].
Theorem 2.1.13 (Itˆo Isometry). Let W : [0, T]×Ω −→ R be the standard Brownian motion defined to T > 0. Let X : [0, T]× Ω −→ R be a stochastic process that is adapted to the natural filtration of the Brownian motion then,
E
T
Z
0
XsdWs
2
=E
t
Z
0
Xs2ds
Theorem 2.1.14 (Comparison theorem). Let (f1, ξ1) and (f2, ξ2) be two standard pa- rameters of BSDEs, and let(X1, Y1) and(X2, Y2)be the associated square integrable solutions. We suppose that
1. ξ1 ≥ξ2 Pa.s.
2. δ2ft =f1(t, Xt2, Yt2)−f2(t, Xt2, Yt2)≥0, dPN dt a.s.
Then we have almost surely for any timet,
Xt1 ≥Xt2.
Moreover, the comparison is strict. That is, if, in addition, X01 = X02, then ξ1 = ξ2, f1(t, Xt2, Yt2) = f2(t, Xt2, Yt2), dPN
dt a.s., and X1 = X2 a.s. More generally,
ifXt1 =Xt2on a setA∈ Ft, thenXs1 =Xs2almost surely on[0, T]×A, ξ1 =ξ2a.s. on A, andf1(t, Xt2, Yt2) = f2(t, Xt2, Yt2)onA×[t, T]dPN
dta.s.
Proof. See [9]
Definition 13 (Generalised generator). The generalised (infinitesimal) generator for a time dependent functionφ(x, t)can be defined as [24]
Atφ(x, t) = lim
s↓0
E[φ(xt+s, t+s)]−φ(xt, t)
s (6)
for a time dependent SDE and At(·) = ∂(·)
∂t +X
i
∂(·)
∂xi
bi(t, Xt) + 1 2
X
i,j
∂2(·)
∂xi∂xj
[σ(t, Xt)Qσ∗(t, Xt)]ij
whereσ∗ is the transpose ofσand the SDE is
dXt=b(t, Xt)dt+σ(t, Xt)dWt, whereQis the diffusion matrix of Brownian motion.
2.2 Background
In this section we will discuss the theory of BSDEs. First, we give some notations used in this chapter and the rest of the work, by [9]. Forx∈Rd,|y|denotes the Euclidean norm, andhy, zidenotes the inner product. Ann×dmatrix will be considered as an element y∈Rn×d,and the Euclidean norm is given by
|z| =p
trace(zz∗), wherez∗ is the transpose ofz, and
hy, zi=trace(yz∗), fory, z∈Rn×d.
Given a probability space, andRn-valued Brownian motionW, we consider the following definitions:
• {Ft; 0 ≤ t ≤ T},the filtration generated by the Brownian motionW and, P the σ-field of predictable sets ofΩ×[0, T].
• L2T(Rd),the space of allFT measurable random variablesX : Ω7−→Rdsatisfying
||Y||2 =E(|Y|2)<+∞.
Usually denoted asL2,d
T .
• H2
T(Rd),the space of all predictable processesϕ : Ω×[0, T]7−→Rdsuch that
||ϕ||2 =E
T
Z
0
|ϕt|2dt <+∞.
Usually denoted asH2,d
T .
• H1T(Rd),the space of all predictable processesϕ : Ω×[0, T]7−→Rdsuch that
E v u u u t
T
Z
0
|ϕt|2dt < +∞.
Usually denotedH1,d
T .
• Forβ > 0andφ ∈H2
T(Rd),
||φ||2β =E
T
Z
0
eβt|φt|2dt,
andH2
T,β(Rd)denotes the spaceH2
T(Rd)equipped with the norm || • ||β. Usually denotedH2,d
T,β.
2.3 Existence and Uniqueness
In this section we are going to prove the existence and uniqueness of the solution of a BSDE driven by L´evy process and another driven by Brownian motion. The drift for both BSDEs is considered to be Lipschitz.
2.3.1 Brownian Motion
Consider a BSDE of the form [9],
−dYt=f(t, Yt, Zt)dt−Zt∗dWt, YT =ξ. (7) or equivalently,
Yt=ξ+
T
Z
t
f(s, Ys, Zs)ds−
T
Z
t
Zs∗dWs (8)
where
• The terminal value is anFT-measurable random variable,ξ: Ω7−→Rd.
• The generatorf : Ω×R+×Rd×Rn×d7−→Rdand isPN BdN
Bn×d-measurable.
HereBddenotes Borel-measurable sets inRd,andBn×ddenotes Borel-measurable sets in Rn×d.
Definition 14. A solution of Equation (7) is a pair (Y, Z) such that {Yt ; t ∈ [0, T]}
is a continuous Rd-valued adapted process, and {Zt ; t ∈ [0, T]} is an Rn×d-valued predictable process satisfying
T
Z
0
|Zs|2ds <+∞, Pa.s.
Definition 15. Suppose thatξ ∈L2,d
T ,f(·,0,0)∈H2,d
T , andf is uniformly Lipschitz that is, there existsC > 0such thatdPN
dta.s.
|f(ω, t, x1, y1)−f(ω, t, x2, y2)| ≤C(|x1−x2|+|y1−y2|), ∀(x1, y1),∀(x2, y2)∈R2. Then(f, ξ)are said to be standard parameters for the BSDE.[9]
Proposition 1. Let ((fi, ξi);i = 1,2) be two standard parameters of the BSDE and ((Yi, Zi);i= 1,2)be two square integrable solutions. LetCbe a Lipschitz constant for f1,and putδYt = Yt1 −Yt2 andδ2ft =f1(t, Yt2, Zt2)−f2(t, Yt2, Zt2).For any(λ, µ, β) such thatµ >0, λ2 > C, β ≥C(2 +λ2) +µ2,it follows that
||δY||2β ≤T
eβTE(|δYT|2) + 1
µ2||δ2f||2β
, (9)
||δZ||2β = λ2 λ2−C
eβTE(|δYT|2) + 1
µ2||δ2f||2β
. (10)
[9].
The Proof of Proposition1is done by [9] and is a follows.
Proof. Let(Y, Z) ∈H2,d
T ×H2T(Rn×d)be a solution of Equation (7). From Equation (8) using triangle inequality and H¨older inequality Theorem2.1.5, we have
|Yt| ≤ |ξ|+
T
Z
t
f(s, Ys, Zs)ds
+
T
Z
t
Zs∗dWs
≤ |ξ|+
T
Z
t
|f(s, Y s, Zs)|ds+
T
Z
t
Zs∗dWs
.
Now taking the supremum, we have
sup
0≤t≤T
|Yt| ≤ |ξ|+
T
Z
0
|f(s, Ys, Zs)|ds+ sup
0≤t≤T
T
Z
t
Zs∗dWs
. (11)
We claim that Yt ∈ L2,d
T .In fact, it is enough to show that each component of the right hand side of Equation (11) is in L2,d
T . Thus using the triangular inequality, Itˆo Isometry
(Theorem2.1.13)and the Burkholder-Davis Gundy inequality (Theorem2.1.3), we get
E
sup
0≤t≤T
T
Z
t
Zs∗dWs
2
=E
sup
0≤t≤T
T
Z
0
Zs∗dWs−
t
Z
0
Zs∗dWs
2
≤E
sup
0≤t≤T
2
T
Z
0
Zs∗dWs
2
+
t
Z
0
Zs∗dWs
2
= 2E
T
Z
0
Zs∗dWs
2
+ 2E
sup
0≤t≤T
t
Z
0
Zs∗dWs
2
= 2E
T
Z
0
|Zs|2ds
+ 2E
sup
0≤t≤T
t
Z
0
Zs∗dWs
2
≤2E
T
Z
0
|Zs|2ds
+ 2C2E
T
Z
0
|Zs|2ds
.
Thussup0≤t≤T |RT
t Zs∗dWs| ∈L2,1
T .Sinceξ ∈L2,1
T , fis uniformly Lipschitz,f(·,0,0)∈ H2,1
T , X ∈H2,d
T , Y ∈H2,n×d
T ,then|ξ|+RT
0 |f(s, Xs, Ys)|ds ∈L2,1
T ,thusYt∈L2,1
T .
Now consider two solutions(Y1, Z1)and(Y2, Z2)associated with(f1, ξ1)and(f2, ξ2), respectively. LetδYs=Ys1−Ys2 such that
δYs=ξ1−ξ2+
T
Z
t
[f1(s, Ys1, Zs1)−f2(s, Ys2, Zs2)]ds−
T
Z
t
δZs∗dWs.
From the Itˆo’s formula Equation (3) applied froms =ttos =T to the semi-martingale eβt|δYt|2, we letf(t, x) =eβt|x|2 then substitutingxwithδYtandywithδZt,we have
eβT|δYT|2−eβt|δYt|2 =
T
Z
t
βeβs|δYs|2ds+
T
Z
t
2eβshδYs, dδYsi+ 1 2
T
Z
t
2eβsdhδYs, δYsi
=
T
Z
t
βeβs|δYs|2ds+
T
Z
t
2eβshδYs, dδYsi+
T
Z
t
eβs|δZs|2ds.
Then
eβT|δYT|2−2
T
Z
t
eβshδYs, dδYsi=eβt|δYt|2+
T
Z
t
βeβs|δYs|2ds
+
T
Z
t
eβs|δZs|2ds.
But we have
2
T
Z
t
eβshδYs, dδYsi=−2
T
Z
t
eβshδYs,(f1(s, Ys1, Zs1)−f2(s, Ys2, Zs2))dsi
+ 2
T
Z
t
eβshδYs, δZs∗dWsi.
Thus we have
eβt|δYt|2+
T
Z
t
βeβs|δYs|2ds+
T
Z
t
eβs|δZs|2ds
=eβT|δYT|2+ 2
T
Z
t
eβshδYs,(f1(s, Ys1, Zs1)−f2(s, Ys2, Zs2))dsi
−2
T
Z
t
eβshδYs, δZs∗dWsi.
Using the Burkholder-Davis Gundy inequality Theorem2.1.3, we have
E
sup
0≤t≤T
T
Z
t
eβsδZsδYsdW s
≤CE
T
Z
0
|δYs|2|δYs|2ds
1 2
≤C
E
sup
0≤t≤T
|δYs|2
E
T
Z
0
|δZs|2ds
1 2
<∞.
Sincesups≤T|δYs|belongs toL2,1
T ,andeβsδZsδYsbelongs toH2,n
T ,and the stochastic in- tegral
RT
t eβshδYs, δZs∗dWsiisP−integrable, with zero expectation. Using the triangle inequal-
ity and Lipschitz condition onf,we have
|f1(s, Ys1, Zs1)−f2(s, Ys2, Zs2)|=|f1(s, Ys1, Zs1)−f1(s, Ys2, Zs2) +f1(s, Ys2, Zs2)−f2(s, Ys2, Zs2)|,
≤ |f1(s, Ys1, Zs1)−f1(s, Ys2, Zs2)|
+|f1(s, Ys2, Zs2)−f2(s, Ys2, Zs2)|,
≤C(|δYs|+|δZs|) +|δ2fs|.
Using2ab≤a2ǫ+ bǫ2 applied to2yCz and2yt,we have the inequality, 2yCz ≤y2ǫ+ (Cz)2
ǫ . Takingǫ=Cλ2 >0, we have
2yCz ≤y2Cλ2+ Cz2 λ2 . For2yt, we have
2yt≤y2ǫ+ t2 ǫ. Takingǫ=µ2 >0, we have
2yt≤y2µ2+ t2 µ2.
Now since2y(Cz+t) = 2Cyz+ 2yt, we have the inequality 2y(Cz+t)≤ Cz2
λ2 + t2
µ2 +y2(µ2+Cλ2). (12)
E
eβt|δYt|2 +βE
T
Z
t
eβs|δYs|2ds
+E
T
Z
t
eβs|δZs|2ds
=E
eβT|δYT|2 + 2E
T
Z
t
eβshδYs,(f1(s, Ys1, Zs1)−f2(s, Ys2, Zs2))dsi
≤E
eβT|δYT|2 +E
T
Z
t
eβs2hδYs,[C(|δYs|+|δZs|) +|δ2fs|]ids
=E
eβT|δYT|2 +E
T
Z
t
eβs(2ChδYs,|δYs|i+ 2ChδYs,|δZs|i+ 2hδYs,|δ2fs|i) ds
.
Using|hy, zi|=|y||z|, we have
E
eβt|δYt|2 +βE
T
Z
t
eβs|δYs|2ds
+E
T
Z
t
eβs|δZs|2ds
≤E
eβT|δYT|2 +E
T
Z
t
eβs 2C|δYs|2+ 2|δYs|(C|δZs|+|δ2fs|) ds
. Applying Equation (12) withy =|δYs|,Z =|δZs|,t =|δ2fs|andC =C,we get
E
eβt|δYt|2 +βE
T
Z
t
eβs|δYs|2ds
+E
T
Z
t
eβs|δZs|2ds
≤E
eβT|δYT|2 +E
T
Z
t
eβs
2C|δYs|2+ C|δZs|2
λ2 + |δ2fs|2
µ2 +|δYs|2 µ2+Cλ2
ds
=E
eβT|δYT|2 +E
T
Z
t
eβs
C|δYs|2(2 +λ2) + C|δZs|2
λ2 +|δ2fs|2
µ2 +µ2|δYs|2
ds
=E
eβT|δYT|2
+ [C(2 +λ2) +µ2]E
T
Z
t
eβs|δYs|2ds
+ C λ2E
T
Z
t
eβs|δZs|2ds+ 1 µ2E
T
Z
t
eβs|δ2fs|2ds,
which gives
E[eβt|δYt|2] + β−[C(2 +λ2) +µ2] E
T
Z
t
eβs|δYs|2ds+
1− C λ2
E
T
Z
t
eβs|δZs|2ds
≤E
eβT|δYT|2 + 1
µ2E
T
Z
t
eβs|δ2fs|2ds. (13)
If we takeβ ≥[C(2 +λ2) +µ2]andC ≤λ2,we have
E[eβt|δYt|2]≤E
eβT|δYT|2 +E
T
Z
t
eβs|δ2fs|2 1 µ2ds.
Taking the integral from0 −→ T, using Fubini’s theorem (Theorem2.1.11) and Defini-
tion2.2, we have the control of the norm for the process|δY|as
||δY||2β ≤T eβTE[|δYT|2] +
T
Z
t
1
µ2||δ2fs||2βds
≤T
eβTE[|δYT|2] + 1
µ2||δ2fs||2β
. The control of the process for|δZ|from Equations (13) is
λ2−C λ2 E
T
Z
t
eβs|δZs|2ds≤eβTE[|δYT|2] + 1
µ2||δ2fs||2β. That is,
λ2−C
λ2 ||δZ||2β ≤eβTE[|δYT|2] + 1
µ2||δ2fs||2β. Hence,
||δZ||2β ≤ λ2 λ2−C
eβTE[|δYT|2] + 1
µ2||δ2fs||2β
.
We have developed the necessary tools we need to prove uniqueness and existence of a solution. A detailed proof of the following theorem by [4], is given in their article. As in [9] we will prove it using the Banach fixed point theorem and a priori estimates.
Theorem 2.3.2. Given standard parameters (f, ξ), there exists a unique pair(Y, Z) ∈ H2,d
T ×H2
T(Rn×d)which solves Equation(7)[9].
The solution is often referred to as a square-integrable solution.
Proof. This proof is by [9], we expand on the proof. We use the Banach fixed point theorem (Theorem2.1.4) for the mapping fromH2,d
T,β×H2
T,βRn×donto itself, which maps (y, z)onto the solution(Y, Z)of the BSDE with generatorf(t, xyt, zt),that is
Yt=ξ+
T
Z
t
f(s, ys, zs)ds−
T
Z
t
Zs∗dWs.
The assumption that (f, ξ) are standard parameters implies f is uniformly Lipschitz, f(·,0,0) ∈ H2,d
T , andξ ∈ L2,d.Thus (f(t, yt, zt); t ∈ [0, T])belongs to H2,d
T . Now we show why the solution to the BSDE is defined as a pair of adaptable processes. Consider the continuous versionM of a square integrable martingaleEh
ξ+RT
0 f(s, Ys, Zs)ds|Ft
i,
Mt =E
ξ+
T
Z
0
f(s, ys, zs)ds
Ft
.
[9]. By the Martingale representation theorem (Theorem 2.1.2) there exists a unique integrable processZ ∈H2,n×d
T,β such that
Mt =M0+
t
Z
0
Zs∗dWs.
Define the adapted and continuous process [9]
Yt =Mt−
t
Z
0
f(s, ys, zs)ds.
Substitute forMt,we have
Yt=M0+
t
Z
0
Zs∗dWs−
t
Z
0
f(s, ys, zs)ds
=ξ+
T
Z
0
f(s, ys, zs)ds−
T
Z
0
Zs∗dWs+
t
Z
0
Zs∗dWs−
t
Z
0
f(s, ys, zs)ds
=ξ+
T
Z
t
f(s, ys, zs)ds−
T
Z
t
Zs∗dWs.
SinceYtis adapted, we have Yt=E[Yt|Ft]
=E
ξ+
T
Z
t
f(s, ys, zs)ds−
T
Z
t
Zs∗dWs|Ft
=E
ξ+
T
Z
t
f(s, ys, zs)ds|Ft
−E
T
Z
t
Zs∗dWs|Ft
=E
ξ+
T
Z
t
f(s, ys, zs)ds|Ft
.
Y is square integrable sincef, ξ are square integrable. Let (y1, z1), and(y2, z2)be two elements ofH2,d
T,β ×H2
T,βRn×d, and let(Y1, Z1)and(Y2, Z2)be the associated solutions.
By Proposition 1applied with C = 0and β = µ2,we have(f, ξ)standard parameters, δ2fs = f(s, ys1, zs1)−f(s, ys2, zs2)and δYT = 0.Then from Equations (9) and (10), we have
||δY||2β ≤ T βE
T
Z
0
eβs|f(s, ys1, z1s)−f(s, y2s, zs2)|2ds
, and
||δZ||2β ≤ 1 βE
T
Z
0
eβs|f(s, ys1, zs1)−f(s, ys2, zs2)|2ds
.
Sincef is Lipschitz with constantCand using(a+b)2 ≤2(a2+b2), we have
||δY||2β +||δZ||2β ≤ T
β + 1 β
CE
T
Z
0
eβs(|δy|+|δz|)2ds
≤ T
β + 1 β
CE
T
Z
0
eβs2(|δy|2+|δz|2)ds
= 2(1 +T)C β
||δy||2β+||δz||2β
. (14)
Choosingβ > 2(1+T)C, we see that this mapping is a contraction fromH2,d
T,β×H2T,βRn×d onto itself and that there exists a fixed point, which is a unique continuous solution of the BSDE.
From Equation (14), we show that the Picard iterative sequence converges almost surely