1,347
Views
1
CrossRef citations to date
0
Altmetric
Reviews

Forward-backward stochastic differential equation games with delay and noisy memory

Pages 708-729 | Received 13 Sep 2018, Accepted 20 Dec 2019, Published online: 21 Jan 2020

Abstract

The goal of this paper is to study a stochastic game connected to a system of forward-backward stochastic differential equations (FBSDEs) involving delay and noisy memory. We derive sufficient and necessary maximum principles for a set of controls for the players to be a Nash equilibrium in the game. Furthermore, we study a corresponding FBSDE involving Malliavin derivatives. This kind of equation has not been studied before. The maximum principles give conditions for determining the Nash equilibrium of the game. We use this to derive a closed form Nash equilibrium for an economic model where the players maximize their consumption with respect to recursive utility.

AMS SUBJECT CLASSIFICATION:

1. Introduction

The aim of this paper is to study a stochastic game between two players. The game is based on a forward stochastic differential equation (SDE) for the process X. In applications to economy, this process can be thought of as the market situation, e.g. the financial market, the housing market or the oil market. This SDE includes two kinds of memory of the past; regular memory and noisy memory. Regular memory (also called delay, see f. ex. the survey paper by Ivanov et al. [Citation1]) means that the SDE can depend on previous values of the process X. That is, for some given δ>0, X(t) depends on X(tδ). For more on stochastic delay differential equations and optimal control with delay, see Øksendal et al. [Citation2] and Agram and Øksendal [Citation3]. In contrast, noisy memory means that the SDE may involve an Itô integral over previous values of the process, so for δ>0, X(t) depends on tδtX(s)dB(s) where {B(s)}s[0,T] is a Brownian motion. For more on noisy memory, see Dahl et al. [Citation4].

Connected to this SDE are two backward stochastic differential equations (BSDEs). These BSDEs are connected to the SDE in the sense that they depend on {X(t)}t[0,T], as well as the delay and noisy memory of this process. Hence, this forms an FBSDE system. Each of these BSDEs corresponds to one of the players in the stochastic game; corresponding to player i = 1, 2 is a BSDE in the process {Wi(t)}t[0,T]. The length of memory can be different for the two players, so for i = 1, 2, player i has memory span δi. The players may also have different levels of information, which is included in the model by having (potentially) different filtrations {Et(i)}t[0,T], i = 1, 2.

Each of the players aim to find an optimal control ui which maximizes their personal performance (objective) function, Ji. Seminal work in stochastic optimal control has been done by Krylov and his students, see e.g. Krylov [Citation5, Citation6]. The performance function of each of the agents will be defined in such a way that it depends on the player’s profit rate, the market process X and the process Wi coming from the player’s BSDE (more on this in Section 2, EquationEquation (11)). This kind of problem, where both players maximize their performance which depends on an FBSDE, is called an FBSDE stochastic game, and has been studied by e.g. Øksendal and Sulem [Citation7]. However, they do not include memory in their model. We study conditions for a pair of controls (u1, u2) to be a Nash equilibrium for such a stochastic game. That is, we would like to determine controls such that the players cannot benefit by changing their actions. In order to do so, we derive sufficient and necessary maximum principles giving conditions for a control to be Nash optimal. This is done in Sections 3 and 4. Maximum principles for forward-backward stochastic differential equations (FBSDEs) have been studied by Øksendal and Sulem [Citation7], Wang and Wu [Citation8], Wu [Citation9] and Wang et al. [Citation10], but these papers do not consider delay and noisy memory.

In connection with these maximum principles, there are adjoint equations (see e.g. Øksendal [Citation11] for an introduction to stochastic maximum principles and adjoint equations, or Øksendal and Sulem [Citation12] for maximum principles and adjoint equations where delay is involved). In our case, these adjoint equations are a system of coupled forward-backward stochastic differential equations involving Malliavin derivatives (see Di Nunno et al. [Citation13] for more on Malliavin derivatives). To the best of our knowledge, such equations have not been studied before. In Section 5 we study a slightly simplified version of these adjoint FBSDEs, and establish a connection between these equations and a system of FBSDEs without Malliavin derivatives. Finally, in Section 6, we apply our results to a specific example in order to determine the optimal consumption with respect to recursive utility.

2. The problem

Let (Ω,F,P) be a probability space, and let B(t), t[0,T] be a Brownian motion in this space. Let (N([0,t],B),0tT,BR{0}) be an independent Poisson random measure. Denote by ν(B) the associated Lèvy measure such that E[N([0,t],B)]=ν(B)t. Also, let N˜(t,·) be the corresponding compensated Poisson random measure, i.e., N˜(dt,dμ):=N(dt,dμ)ν(dμ)dt.

Let (Ft)t[0,T] be the P-augmented filtration generated by B(t) and N˜(t,·).

We will consider a game between two players: player 1 and player 2. Let ui(t) be the control process chosen by player i = 1, 2, and denote u(t)=(u1(t),u2(t)). Let Ai, i = 1, 2, denote the set of admissible controls for player i. It is contained in a given set of càdlàg processes in L2(Ω×[0,T]), with values in a subset Vi of R. Let A=A1×A2 be the combined controls for both players, and denote by V:=V1×V2. Let δi0, i = 1, 2 be the memory span of player 1 and 2, respectively. We define δ:=maxδ1,δ2 to be the longest memory span of the two agents.

We consider a controlled forward stochastic differential equation for a process X(t)=Xu(t,ω),ωΩ,t[0,T] determining the market situation (in the following, we omit the ω for notational ease unless it is important to highlight its dependence): (1) dX(t)=b(t,X(t),Y(t),Λ(t),u(t),ω)dt+σ(t,X(t),Y(t),Λ(t),u(t),ω)dB(t)+Rγ(t,X(t),Y(t),Λ(t),u(t),ζ,ω)N˜(dt,dζ), t[0,T],X(t)=ξ(t), t[δ,0),(1) where ξ(t) is some (given) initial process, Y(t)=(Y1(t),Y2(t)),Λ(t)=(Λ1(t),Λ2(t)), and Yi(t):=X(tδi),Λi(t):=tδitX(s)dB(s), and δi0 for i = 1, 2. The superscript t means that we are taking the left limit of the process is question (that is, the value before a potential jump at time t), see Øksendal and Sulem [Citation14] for more on this.

Remark 2.1. N

ote that ξ(t) is a given initial process which can not be controlled (i.e., there is no dependency on u in ξ). Hence, we do not need to define the filtration (Ft) for t[δ,0).

Here, the delay processes Yi, and the noisy memory processes Λi correspond to player i = 1, 2 respectively. Hence, the two players may have memories for different time intervals, depending on the values of δi. Also, on the coefficient functions (2) b:[0,T]×R×R2×R2×VR×Ω,(2) (3) σ:[0,T]×R×R2×R2×VR×Ω,(3) (4) γ:[0,T]×R×R2×R2×V×R×ΩR,(4) we impose the following set of assumptions.

Assumption 2.2.

  1. The functions b(ω,t,·),σ(ω,t·) and γ(ω,t,ζ,·) are assumed to be C1 for each fixed ω,t,ζ.

  2. The functions b(·,x,y,z,u) and σ(·,x,y,z,u), and γ(·,x,y,z,u,ζ) are predictable for each x,y,z,u.

  3. Lipschitz condition: The functions b,σ are Lipschitz continuous in the variables x,y,z, with the Lipschitz constant independent of the variables t,u,ω. Also, there exists a function LL2(ν), independent of t,u,ω, such that (5) |γ(ω,t,x1,y1,z1,u,ζ)γ(ω,t,x2,y2,z2,u,ζ)|(5) (6) L(ζ){|x1x2|+||y1y2||+||z1z2||},νa.e.ζ.(6)

  4. Linear growth: The functions b,σ,γ satisfy the linear growth condition in the variables x,y,z, with the linear growth constant independent of the variables t,u,ω Also, there exists a non-negative function KL2(ν), independent of t,u,ω, such that (7) |γ(ω,t,x,y,z,u,ζ)|(7) (8) K(ζ){1+|x|+||y||+||z||},νa.e.ζ.(8)

Assumptions 1 and 2 are sufficient to ensure the integrands in EquationEquation (1) have predictable versions, whenever X is càdlàg and adapted. It is always assumed that the N˜-integral is taken with respect to the predictable version of γ(t,X(t),Y(t),Z(t),u(t),ζ). Together with the Lipschitz and linear growth conditions, this ensures that for every uA, there exists a unique càdlàg adapted solution X=Xu to the EquationEquation (1) satisfying (9) E[supt[δ,T]|X(t)|2]<.(9)

This can be seen, for example, by regarding EquationEquation (1) as a stochastic functional differential equation. See Dahl et al. [Citation4] for more on this.

In addition to this, the players (potentially) have different levels of information, represented by different subfiltrations (Et(i))0tT where Et(i)Ft for all t[0,T], i = 1, 2.

For i = 1, 2, let gi(·,x,y,Λ,wi,zi,ki(·),u,ω) be a given predictable process w.r.t. (Et(i))0tT, i = 1, 2, and let hi(x,ω) be an FT-measurable function. Associated to the FSDE (1), we have a pair of backward stochastic differential equations (BSDEs) in the unknown stochastic processes (Wi,Zi,Ki), i = 1, 2: (10) dWi(t)=gi(t,X(t),Y(t),Λ(t),Wi(t),Zi(t),Ki(t,·),u(t),ω)dt+Zi(t)dB(t)+RKi(t,ζ)N˜(dt,dζ), t[0,T],Wi(T)=hi(X(T),ω).(10)

Note that these BSDEs are coupled to the SDE (1) due to the dependency on X. Also, the BSDEs depend on the memory of the market process X, due to the dependency on the processes Y and Λ. However, EquationEquation (10) is a standard BSDE with jumps, hence the conditions for existence and uniqueness of solution are well known, see e.g. Theorem 1.5 in Øksendal and Sulem [Citation15]. Essentially, we require that g is square integrable w.r.t. t when all other inputs are 0 and that g is Lipschitz in W,Z and K.

For i = 1, 2, let fi:[0,T]×R×R×R×A×ΩR,φi:RR,ψi:RR be functions representing a profit rate, bequest function and risk evaluation, respectively. Then, the performance function of each player i = 1, 2 is defined by: (11) Ji(u)=E[0Tfi(t,Xu(t),Yiu(t),Λiu(t),ui(t))dt+φi(Xu(T))+ψi(Wiu(0))],(11) where we must assume all conditions necessary for the integrals and the expectation to exist.

Also, note that the performance Ji of player i is a function of the control u(t)=(u1(t),u2(t)), which is determined by both players. Therefore, this problem setting specifies a stochastic game.

A pair of controls (û1,û2) is called a Nash equilibrium for this stochastic game if the following holds: (12) J1(u1,û2)J1(û1,û2) for all u1A1,J2(û1,u2)J2(û1,û2) for all u2A2.(12)

In words, this means that in the Nash equilibrium, neither player would like to change their control.

Assume there exists a Nash equilibrium for this forward-backward stochastic differential (FBSDE) game with delay and noisy memory. We would like to find this Nash equilibrium, and we will do so by proving sufficient and necessary maximum principles for this problem. Therefore, we define a Hamiltonian function for each player i = 1, 2 as follows: (13) Hi(t,x,y,Λ,wi,zi,ki,u1,u2,λi,pi,qi,ri)=fi(t,x,yi,Λi,ui)+λigi(t,x,y,Λ,wi,zi,ki,u1,u2)+pib(t,x,y,Λ,u1,u2)+qiσ(t,x,y,Λ,u1,u2)+Rri(ζ)γ(t,x,y,Λ,u1,u2,ζ)ν(dζ).(13)

Assume Hi is C1 in x,y1,y2,Λ1,Λ2,wi,zi,ki,u1,u2 for i = 1, 2. In the following, for ease of notation, we will use the abbreviation Hi(t)=Hi(t,x(t),y(t),Λ(t),wi(t),zi(t),ki(t),u1(t),u2(t),λi(t),pi(t),qi(t),ri(t)).

For i = 1, 2, we define a system of FBSDEs associated to these Hamiltonians in the unknown adjoint processes (λi,pi,qi,ri):

FSDE in λi (which depends on pi,qi,ri): (14) dλi(t)=Hiwi(t)dt+Hizi(t)dB(t)+Rki(Hi(t,ζ))N˜(dt,dζ),λi(0)=ψi(Wi(0)),(14) where ki(Hi(t,ζ)) is the Fréchet derivative of Hi at ki, see the appendix in Øksendal and Sulem [Citation7] for a closer explanation of this gradient.

We also define a BSDE in pi,qi,ri, which depends on λi: (15) dpi(t)=E[μi(t)|Ft]dt+qi(t)dB(t)+Rri(t,ζ)N˜(dt,dζ),pi(T)=φi(X(T))+hi(X(T))λi(T),(15) where μi(t)=Hix(t)Hiyi(t+δi)1[0,Tδi](t)tt+δiDt[HiΛi(s)1[0,T](s)ds] and Dt[·] denotes the Malliavin derivative (see Remark 2.3). Note that the conditional expectation in (15) is well defined by the extension of the Malliavin derivative introduced by Aase et al. [Citation16], see Remark 2.3. EquationEquations (14) and Equation(15) form an FBSDE-system involving Malliavin derivatives. To the best of our knowledge, such systems have not been studied before.

Remark 2.3.

We refer to Nualart [Citation17], Sanz-Solè [Citation18] and Di Nunno et al. [Citation13] for information about the Malliavin derivative Dt for Brownian motion B(t) and, more generally, Lévy processes. In Aase et al. [Citation16], Dt was extended from the space D1,2 to L2(P), where D1,2 denotes the classical space of Malliavin differentiable FT-measurable random variables. The extension is such that for all FL2(FT,P), the following holds:

  1. DtF(S)*, where (S)*L2(P) denotes the Hida space of stochastic distributions,

  2. the map (t,ω)E[DtF|Ft] belongs to L2(FT,λ×P), where λ denotes the Lebesgue measure on [0,T].

  3. Moreover, the following generalized Clark–Ocone theorem holds: (16) F=E[F]+0TE[DtF|Ft]dB(t).(16)

    See [Citation16], Theorem 3.11, and also [Citation13], Theorem 6.35. Notice that by combining Itô’s isometry with the Clark–Ocone theorem, we obtain (17) E[0TE[DtF|Ft]2dt]=E[(0TE[DtF|Ft]dB(t))2]=E[(F2E[F]2)].(17)

  4. As observed in Agram et al. [Citation19], we can also apply the Clark–Ocone theorem to show the following generalized duality formula:

    Let FL2(FT,P) and let φ(t)L2(λ×P) be adapted. Then, (18) E[F0Tφ(t)dB(t)]=E[0TE[DtF|Ft]φ(t)dt].(18)

Remark 2.4. N

ote that Equation Equation(14) is linear in λi, and hence, if pi,qi,ri were given, it could be solved by using the Itô formula. However, this solution will depend on the processes X,Yi,Λi and Wi, so in order to find an explicit solution for λi, we must also solve the coupled FBSDE system (1)–(10).

The BSDE (15) is linear in pi, and hence, if λi was given, it would be possible to find a unique solution to this equation by using e.g. Proposition 6.2.1 in Pham [Citation20] or Theorem 1.7 in Øksendal and Sulem [Citation15]. However, as for the adjoint SDE (14), this solution will depend on the coupled FBSDE system (1)–(10).

In the remaining part of the paper, we will prove a sufficient (Section 3) and a necessary maximum principle (Section 4) for this kind of FBSDE game with delay and noisy memory. Then, we will study existence and uniqueness of solutions of the FBSDE system (14) and (15) (Section 5). Finally, we will present an example which illustrates our results: optimal consumption rate with respect to recursive utility (see Section 6).

3. Sufficient maximum principle for FBSDE games with delay and noisy memory

We prove a sufficient maximum principle which roughly states that under concavity conditions, a control (û1,û2) satisfying a conditional maximum principle and an 2-condition is a Nash equilibrium for the stochastic game.

Theorem 3.1.

Let û1A1 and û2A2 with corresponding solutions X̂(t),Ŷi(t),Λ̂i(t),Ŵi(t),Ẑi(t), K̂i(t),λ̂i(t), p̂i(t),q̂i(t),r̂i(t,ζ) of the FSDE (1), the BSDE (10), and the FBSDE system (14) and (15) for i = 1, 2. Also, assume that:

  • (Concavity I) The functions xhi(x),xφi(x),xψi(x) are concave for i = 1, 2.

  • (The conditional maximum principle) ess sup vA1E[H1(t,X̂(t),Ŷ(t),Λ̂(t),Ŵ1(t),Ẑ1(t),K̂1(t,·),v,û2(t),λ1̂(t),p̂1(t),q̂1(t),r̂1(t,·))|Et(1)]=E[H1(t,X̂(t),Ŷ(t),Λ̂(t),Ŵ1(t),Ẑ1(t),K̂1(t,·),û1(t),û2(t),λ̂1(t),p̂1(t),q̂1(t),r̂1(t,·))|Et(1)],

and similarly ess sup vA2E[H2(t,X̂(t),Ŷ(t),Λ̂(t),Ŵ2(t),Ẑ2(t),K̂2(t,·),û1,v,λ2̂(t),p̂2(t),q̂2(t),r̂2(t,·))|Et(2)]=E[H2(t,X̂(t),Ŷ(t),Λ̂(t),Ŵ2(t),Ẑ2(t),K̂2(t,·),û1(t),û2(t),λ̂2(t),p̂2(t),q̂2(t),r̂2(t,·))|Et(2)].
  • (Concavity II) The functions Ĥ1(t,x,y1,Λ1,w1,z1,k1):=ess sup vA1E[H1(t,x,y1,ŷ2,Λ1,Λ̂2,w1,z1,k1,v,û2,λ̂1,p̂1,q̂1,r̂1)|Et(1)]

and Ĥ2(t,x,y2,Λ2,w2,z2,k2):=ess sup vA2E[H2(t,x,ŷ1,y2,Λ̂1,Λ2,w2,z2,k2,û1,v,λ̂2,p̂2,q̂2,r̂2)|Et(2)]

are concave for all t a.s.

  • Finally, assume that the following 2 conditions hold: E[0T{p̂i2(t)[(σ(t)σ̂(t))2+R(ri(t,ζ)r̂i(t,ζ))2ν(dζ)]+(X(t)X̂(t))2[q̂i2(t)+Rr̂i2(t,ζ)ν(dζ)]+(Yi(t)Ŷi(t))2[(Ĥiz)2(t)+R||kĤi(t,ζ)||2ν(dζ)]+λ̂i2(t)[(Λi(t)Λ̂i(t))2+R(Ki(t,ζ)K̂i(t,ζ))2ν(dζ)]}]<

for i = 1, 2.

Then, (û1,û2) is a Nash equilibrium.

Proof.

We would like to show that J1(u1,û2)J1(û1,û2) for all u1A1. Choose u1A1. By the definition of the performance function J1, Δ:=J1(u1,û2)J1(û1,û2)=I1+I2+I3, where I1=E[0T{f1(t,x,y,Λ,u)f1(t,x̂,ŷ,Λ̂,û)}dt],I2=E[φ1(X(T))φ1(X̂(T))],I3=E[ψ1(W1(0))ψ1(Ŵ1(0))].

Note that from the definition of the Hamiltonian, (19) I1=E[0T{H1(t)Ĥ1(t)λ̂1(t)(g1(t)ĝ1(t))p̂1(t)(b(t)b̂(t))q̂1(t)(σ(t)σ̂(t))Rr̂1(t,ζ)(γ(t,ζ)γ̂(t,ζ)ν(dζ))}dt],(19) where we have used the abbreviation Ĥ1(t):=H1(t,X̂(t),Ŷ(t),Λ̂(t),Ŵ1(t),Ẑ1(t),K̂1(t,·),û(t),λ̂1(t),p̂1(t),q̂1(t),r̂1(t),ω) and corresponding abbreviations for H1(t),b(t),b̂(t),σ,σ̂(t),γ(t) and γ̂(t).

Also, (20) I2=E[φ1(X(T))φ1(X̂(T))]E[φ1(X̂(T))(X(T)X̂(T))]=E[(p̂1(T)h1(X̂(T))λ̂1(T))(X(T)X̂(T))]=E[p̂1(T)(X(T)X̂(T))]E[λ̂1(T)h1(X̂(T))(X(T)X̂(T))]=E[0Tp̂1(t)(dX(t)dX̂(t))+0T(X(t)X̂(t))dp̂1(t)+0Tq̂1(t)(σ(t)σ̂(t))dt+0TRr̂1(t,ζ)(γ(t,ζ)γ̂(t,ζ))ν(dζ)dt]E[λ̂1(T)h1(X̂(T))(X(T)X̂(T))]=E[0Tp̂1(t)(b(t)b̂(t))dt+0T(X(t)X̂(t))(Ĥ1x(t)Ĥ1y1(t+δ1)1[0,Tδ1](t)+tt+δ1Dt[Ĥ1Λ1(s)]1[0,T](s)ds)dt+0Tq̂1(t)(σ(t)σ̂(t))dt+0TRr̂1(t,ζ)(γ(t,ζ)γ̂(t,ζ))ν(dζ)dt]]E[λ̂1(T)h1(X̂(T))(X(T)X̂(T))],(20) where the first inequality follows from the concavity of φ1, the second equality follows from EquationEquation (15), the fourth equality from Itô’s product rule applied to p̂1X and p̂1X̂, the fifth equality follows from EquationEquation (15), the double expectation rule and EquationEquation (1).

Also, note that (21) I3=E[ψ1(W1(0))ψ1(Ŵ1(0))]E[ψ1(Ŵ1(0))(W1(0)Ŵ1(0))]=E[λ̂1(T)(W1(T)Ŵ1(T))]{E[0T(W1(t)Ŵ1(t))dλ̂1(t)+0Tλ̂1(t)(dW1(t)dŴ1(t))+0TĤ1z1(t)(Z1(t)Ẑ1(t))dt+0TRk1Ĥ1(t)(K1(t)K̂1(t))ν(dζ)dt]}=E[λ̂1(T)(h1(X(T))h1(X̂(T)))]{E[0TĤ1w1(t)(W1(t)Ŵ1(t))dt+0Tλ̂1(t)(g1(t)+ĝ1(t))dt+0TĤ1z1(t)(Z1(t)Ẑ1(t))dt+0TRkĤ1(t)(K1(t)K̂1(t))ν(dζ)dt]}E[λ̂1(T)h1(X̂(T))(X(T)X̂(T))]{E[0TĤ1w1(t)(W1(t)Ŵ1(t))dt+0Tλ̂1(t)(g1(t)+ĝ1(t))dt+0TĤ1z1(t)(Z1(t)Ẑ1(t))dt+0TRk1Ĥ1(t)(K1(t)K̂1(t))ν(dζ)dt]},(21) where the first inequality follows from the concavity of ψ1, the second equality follows from EquationEquation (14), the third equality follows from Itô’s product rule applied to λ̂1Y1 and λ̂1Ŷ1, the fourth equality follows from EquationEquation (10) as well as EquationEquation (14). The final inequality follows from the concavity of h1 and that λ̂1(T)0.

Hence, (22) Δ=I1+I2+I3E[0T{H1(t)Ĥ1(t)(Ĥ1x(t)+Ĥ1y1(t+δ1)1[0,Tδ1](t)+tt+δ1Dt[Ĥ1Λ1(s)]1[0,T](s)ds)(X(t)X̂(t))dt}0T{Ĥ1w1(t)(W1(t)Ŵ1(t))+Ĥ1z1(t)(Z1(t)Ẑ1(t))+Rk1Ĥ1(t)(K1(t,ζ)K̂1(t,ζ))ν(dζ)}dt].(22)

Note that by changing the order of integration and using the duality formula for Malliavin derivatives (see Di Nunno et al. [Citation13]), we get: (23) E[0TĤ1Λ1(s)(Λ1(s)Λ̂1(s))ds]=E[0TĤ1Λ1(s)sδ1s(X(t)X̂(t))dB(t)ds]=0TE[Ĥ1Λ1(s)sδ1s(X(t)X̂(t))dB(t)]ds=0TE[sδ1sE[Dt(Ĥ1Λ1(s))|Ft](X(t)X̂(t))dt]ds=E[0Ttt+δ1E[Dt(Ĥ1Λ1(s))|Ft]1[0,T](s)ds(X(t)X̂(t))dt]=E[0Ttt+δ1Dt(Ĥ1Λ1(s))1[0,T](s)ds(X(t)X̂(t))dt].(23)

Also, note that (24) E[0TĤy1(t)(Y1(t)Y1̂(t))dt]=E[0TĤy1(t)(X(tδ)X̂(tδ1))dt]=E[0TĤy1(t+δ1)1[0,Tδ1](t)(X(t)X̂(t))dt].(24)

Hence, by the inequality (22) combined with EquationEquations (23) and Equation(24), (25) ΔE[0T{H1(t)Ĥ1(t)Ĥ1x(t)(X(t)X̂(t))Ĥ1y1(t)(Y1(t)Ŷ1(t))Ĥ1Λ1(t)(Λ1(t)Λ̂1(t))dtĤ1w1(t)(W1(t)Ŵ1(t))Ĥ1z1(t)(Z1(t)Ẑ1(t))+Rk1Ĥ1(t)(K1(t,ζ)K̂1(t,ζ))ν(dζ)}dt].(25)

Fix some t[0,T]. By assumption, Ĥ1(x):=Ĥ1(t,x) is concave, so it is superdifferentiable1 (see Rockafellar [Citation21]) at the point x :=(X̂,Ŷ1,Λ̂1,Ŵ1,Ẑ1,K̂1). Thus, there exists a supergradient a:=(a0,a1,a2,a3,a4,a5(·)) such that for all y:=(x,y,Λ,w,z,k), the following holds: (26) Ĥ1(x)+a·(yx)Ĥ1(y).(26)

Define (27) ϕ1(t,x,y,Λ,w,z,k):=Ĥ1(t,x,y,Λ,w,z,k)Ĥ1(t,X̂,Ŷ1,Λ̂1,Ŵ1,Ẑ1,K̂1){a0(xX̂)+a1(yŶ1)+a2(ΛΛ1)+a3(wŴ1)+a4(zẐ1)+Ra5(ζ)(kK̂1)ν(dζ))}.(27)

Then, by EquationEquation (26), (28) ϕ1(t,x,y,Λ,w,z,k)0 for all x,y,Λ,w,z,k,ϕ1(t,X̂,Ŷ1,Λ̂1,Ŵ1,Ẑ1,K̂1)=0 (by definition).(28)

Therefore, by differentiating EquationEquation (27) and using EquationEquation (28), we find that a0=Ĥ1x(t,X̂,Ŷ1,Λ̂1,Ŵ1,Ẑ1,K̂1)=Ĥ1x,a1=Ĥ1y1(t,X̂,Ŷ1,Λ̂1,Ŵ1,Ẑ1,K̂1)=Ĥ1y1,a2=Ĥ1Λ1(t,X̂,Ŷ1,Λ̂1,Ŵ1,Ẑ1,K̂1)=Ĥ1Λ1,a3=Ĥ1w1(t,X̂,Ŷ1,Λ̂1,Ŵ1,Ẑ1,K̂1)=Ĥ1w1,a4=Ĥ1z1(t,X̂,Ŷ1,Λ̂1,Ŵ1,Ẑ1,K̂1)=Ĥ1z1,a5=k1Ĥ1(t,X̂,Ŷ1,Λ̂1,Ŵ1,Ẑ1,K̂1)=k1Ĥ1.

Therefore, it follows from this, EquationEquations (25) and Equation(28) that Δ=ϕ1(t,X(t),Y1(t),Λ1(t),W1(t),Z1(t),K1(t,·))0 t[0,T] (where the final inequality follows since Ĥ1 is concave). This means that J1(u1,û2)J1(û1,û2) for all u1A1. In a similar way, one can prove that J2(û1,u2)J2(û1,û2) for all u2A2. This completes the proof that (û1,û2) is a Nash-equilibrium. □

4. Necessary maximum principle for FBSDE games with delay and noisy memory

In the following, we need some additional assumptions and notation:

  • For all t0[0,T] and all bounded Ei(t)-measurable random variables αi(ω), the control (29) βi(t):=1(t0,T)(t)αi(ω) is in Ai for i=1,2.(29)

  • For all ui,βiAi with βi bounded, there exists κi>0 such that the control (30) ui(t)+sβi(t) for t[0,T],(30)

belongs to Ai for all s(κi,κi), i = 1, 2.

  • Also, assume that the following derivative processes exist and belong to L2([0,T]×Ω): (31) x1(t)=ddsX(u1+sβ1,u2)(t)|s=0,y1(t)=ddsY1(u1+sβ1,u2)(t)|s=0,(31) Λ˜1(t)=ddsΛ1(u1+sβ1,u2)(t)|s=0,w1(t)=ddsW1(u1+sβ1,u2)(t)|s=0,z1(t)=ddsZ1(u1+sβ1,u2)(t)|s=0,k1(t)=ddsK1(u1+sβ1,u2)(t)|s=0,

and similarly for x2(t)=ddsX(u1,u2+sβ2)(t)|s=0 etc. Here, the derivative processes are directional derivatives, defined in the following way: (32) x1(t)=ddsX(u1+sβ1,u2)(t):=limΔs0X(u1+(s+Δs)β1,u2)(t)X(u1+sβ1,u2)(t)Δs.(32)

For more on this, see (4.11) in Di Nunno et al. [Citation22] and Appendix A in Øksendal and Sulem [Citation7]. Note also that xi(0)=0 for i = 1, 2 since X(0)=x.

If these assumptions hold, we can prove a necessary maximum principle for our noisy memory FBSDE game. The proof of the following theorem is based on the same idea as the proof of Theorem 2.2 in Øksendal and Sulem [Citation7], however the presence of noisy memory in our problem requires some extra care.

Theorem 4.1.

Suppose that uA with corresponding solutions X(t),Yi(t), Λi(t),Wi(t), Zi(t),Ki(t,ζ), λi(t),pi(t), qi(t),ri(t,ζ), i = 1, 2, of EquationEquations (1), Equation(10), Equation(14) and Equation(15). Also, assume that conditions (29)–(31) hold. Then, the following are equivalent:

  1. sJ1(u1+sβ1,u2)|s=0=sJ2(u1,u2+sβ2)|s=0=0 for all bounded β1A1,β2A2.

  2. E[H1(t,X(t),Y(t),Λ(t),W1(t),Z1(t),K1(t,·),v1,u2(t),λ1(t),p1(t),q1(t),r1(t,·))v1]|v1=u1(t) =E[H2(t,X(t),Y(t),Λ(t),W2(t),Z2(t),K2(t,·),u1(t),v2,λ2(t),p2(t),q2(t),r2(t,·))v2]|v2=u2(t)=0.

Proof.

We only prove that sJ1(u1+sβ1,u2)|s=0=0 for all bounded β1A1 is equivalent to E[H1(t,X(t),Y(t),Λ(t),W1(t),Z1(t),K1(t,·),v1,u2(t),λ1(t),p1(t),q1(t),r1(t,·))v1]|v1=u1(t)=0.

The remaining part of the theorem (i.e., the same statement for J2andH2) is proved in a similar way.

Note that, by the definition of J1 and by interchanging differentiation and integration, (33) D1:=sJ1(u1+sβ1,u2)|s=0=E[0T{f1x(t)x1(t)+f1y(t)y1(t)+f1Λ(t)Λ˜1(t)f1u1(t)β1(t)}dt+φ1(X(T))x1(T)+ϕ1(W1(0))w1(0)].(33)

Note that the interchange of differentiation and integration is justified since everything in EquationEquation (33) is well defined and square integrable by assumption and P×[0,T] is a finite measure space. Hence, we can apply Theorem 11.5 in Shilling [Citation23] to change the order of the expectation/integral and the differentiation. Also, note that D1 is a directional derivative of J1, defined similarly as in EquationEquation (32). For more details on directional (also called Gâteaux derivative), see Appendix A in Øksendal and Sulem [Citation7]. Furthermore, note that fx is the partial derivative of the function f wrt. x inserted the corresponding processes at time t. For proofs of the differentiability of the performance functional in a similar context, see Dahl et al. [Citation4].

We study the different parts of D1 separately. First, by the Itô product rule, the adjoint BSDE (15) and the definition of x1(t), (34) I1:=E[φ1(X(T))x1(T)]=E[p1(T)x1(T)]E[h1(X(T))λ1(T)x1(T)]=E[p1(0)x1(0)]+E[0Tp1(t)dx1(t)+0Tx1(t)dp1(t)+0Td[p1,x1](t)]E[h1(X(T))λ1(T)x1(T)]=E[0Tp1(t)(bx(t)x1(t)+by1(t)y1(t)+bΛ1(t)Λ˜1(t)+bu1(t)β1(t))dt]+E[0Tx1(t)E[μ1(t)|Ft]dt]+E[0Tq1(t)(σx(t)x1(t)+σy1(t)y1(t)+σΛ1Λ˜1(t)+σu1(t)β1(t))dt]+E[0TRr1(t,ζ)(γx(t)x1(t)+γy1(t)y1(t)+γΛ1Λ˜1(t)+γu1(t)β1(t))dν(ζ)dt]E[h1(X(T))λ1(T)x1(T)].(34)

Also, by the FSDE (14), the BSDE (10), the definition of x1(t) and the Itô product rule, (35) I2:=E[ϕ1(W1(0))w1(0)]=E[λ1(0)w1(0)]=E[λ1(T)w1(T)]E[0Tλ1(t)dw1(t)+0Tw1(t)dλ1(t)(35) +0Tz1(t)H1z1(t)dt+0TRk1H1(t,ζ)k1(t,ζ)ν(dζ)dt]=E[λ1(T)h1(X(T))x1(T)]+E[0Tλ1(t)(g1x(t)x1(t)+g1y1(t)y1(t)+g1Λ1(t)Λ˜(t)+g1w1(t)w1(t)+g1z1(t)z1(t)+k1g1(t)k1(t)+g1u1(t)β1(t))dt]E[0TH1w1(t)w1(t)dt]E[0Tz1(t)H1z1(t)dt+0TRkH1(t,ζ)k1(t,ζ)ν(dζ)dt].

By the definition of D1 as well as EquationEquations (34) and Equation(35), (36) D1=A+E[0Tβ1(t)(f1u1(t)+bu1(t)p1(t)+σu1(t)q1(t)+γu1(t)r1(t)+g1u1(t)λ1(t))dt]+E[0Tw1(t){H1w1(t)+g1w1(t)λ1(t)}dt+0Tz1(t){H1z1(t)+g1x(t)λ1(t)}dt+0Tk1(t){k1H1(t)+kg1(t)λ1(t)}dt],(36) where (37) A:=E[0Tx1(t){f1x(t)+bx(t)p1(t)+E[μ1(t)|Ft]+σx(t)q1(t)+γx(t)r1(t)+g1x(t)λ1(t)}dt+0Ty1(t){f1y1(t)+by1(t)p1(t)+σy1(t)q1(t)+γy1(t)r1(t)+g1y1(t)λ1(t)}dt+0TΛ˜1(t){f1Λ1(t)+bΛ1(t)p1(t)+σΛ1(t)q1(t)+γΛ1(t)r1(t)+g1Λ1(t)λ1(t)}dt]=E[0Tx1(t){H1x(t)+E[μ1(t)|Ft]}dt]+E[0Ty1(t)H1y1(t)]+E[0TΛ˜1(t)H1Λ1(t)].(37)

Then, by using the definition of the Hamiltonian H1, see EquationEquation (13), we see that everything inside the curly brackets in EquationEquation (36) is equal to zero. Hence, D1=A+E[0Tβ1(t)H1u1(t)dt].

Recall that from the definitions of y1andΛ˜1, y1(t)=x1(tδ1) and Λ˜1(t)=tδ1tx1(u)dB(u).

This implies, by change of variables E[0Ty1(t)H1y1(t)]=E[0Tx1(tδ1)H1y1(t)dt]=δ1Tδ1x1(u)H1y1(u+δ1)du]=E[0Tx1(u)1[0,Tδ1](u)H1y1(u+δ1)du].

Also, by the duality formula for Malliavin derivatives (see Di Nunno et al. [Citation13]) and changing the order of integration E[0TΛ˜1(t)H1Λ1(t)]=E[0Ttδ1tx1(u)dB(u)H1Λ1(t)dt]=E[0Ttδ1tE[Du(H1Λ1(t))|Fu]x1(u)du dt]=E[0Tuu+δ1E[Du(H1Λ1(t))|Fu]1[0,T](t)dt x1(u) du].

But, from the definition of μ1, E[0Tx1(t)E[μ1(t)|Ft]dt]=E[0TE[x1(t)μ1(t)|Ft] dt]=E[0TE[x1(t){H1x(t)H1y1(t+δ1)1[0,Tδ1]tt+δ1Dt[H1Λ1(s)]1[0,T](s)ds}|Ft]dt].

So, by the rule of double expectation and the calculations above, A=0. This implies that D1=E[0Tβ1(t)H1u1(t)dt], so sJ1(u1+sβ1,u2)|s=0=E[0Tβ1(t)H1u1(t)dt], which was what we wanted to prove. □

5. Solution of the noisy memory FBSDE

In this section, we consider a slightly simplified version of the system of noisy memory FBSDEs in EquationEquations (14) and Equation(15). Instead, consider the following noisy memory FBSDE:

FSDE in λ, (38) dλ(t)=Hw(t)dt+Hz(t)dB(t)+RkH(t,ζ)N˜(dt,dζ),λ(0)=ϕ(W(0)).(38)

BSDE in p,qandr, (39) dp(t)=E[μ(t)|Ft]dt+q(t)dB(t)+Rr(t,ζ)N˜(dt,dζ),p(T)=φ(X(T))+h(X(T))λ(T)(39) where H(t,x,y1,y2,Λ1,Λ2,w,z,k,u1,u2,λ,p,q,r)=f(t,x,y,Λ,u1,u2)+λg(t,x,y1,y2,Λ1,Λ2,w,z,k,u1,u2)+pb(t,x,y1,y2,Λ1,Λ2,u1,u2)+qσ(t,x,y1,y2,Λ1,Λ2,u1,u2)+Rr(ζ)γ(t,x,y1,y2,Λ1,Λ2,u1,u2,ζ)ν(dζ) and μ(t)=Hx(t)+Hy(t+δ)1[0,Tδ](t)+tt+δE[Dt[HΛ(s)]|Ft]1[0,T](s)ds.

Note that the set of EquationEquations (14) and Equation(15) are two such systems such as (38) and (39) involving the same X process as well as the same controls u1,u2.

Also, consider the following system consisting of an FSDE and two BSDEs:

FSDE in λ, (40) dλ˜(t)=Hw(t)dt+Hz(t)dB(t)+RkH(t,ζ)N˜(dt,dζ),λ˜(0)=ϕ(W(0)).(40)

BSDE in p1,q1andr1, (41) dp1(t)=E[μ1(t)|Ft]dt+q1(t)dB(t)+Rr1(t,ζ)N˜(dt,dζ),p1(T)=φ(X(T))+h(X(T))λ˜(T).(41)

BSDE in p2,q2andr2, (42) dp2(t)=E[μ2(t)|Ft]dt+q2(t)dB(t)+Rr2(t,ζ)N˜(dt,dζ),p2(T)=0(42) where (43) H(t,x,y1,y2,Λ1,Λ2,w,z,k,u1,u2,λ˜,p1,p2,q1,q2,r1,r2)=q2(t)x+H(t,x,y1,y2,Λ1,Λ2,w,z,k,u1,u2,λ˜,p1,q1,r1),μ1(t)=q2(t)+Hx(t)+Hy(t+δ)1[0,Tδ](t)(43) and μ2(t)=HΛ(t)HΛ(t+δ)1[0,Tδ](t).

Note that HΛ(t)=HΛ(t),HΛ(t)=q2(t)+HΛ(t)andHy(t)=Hy(t). Hence, EquationEquations (38) and Equation(40) are structurally equal.

Then, by similar techniques as in Dahl et al. [Citation4], we can show the following theorem:

Theorem 5.1.

Assume that (pi,qi,ri)fori=1,2andλ˜ solve the FBSDE system (40)–(42). Define λ=λ˜,p(t)=p1(t),q(t)=q1(t)andr(t,·)=r1(t,·) and assume that E[0T(H(t)z)2]dt<. Then, (p,q,r,λ) solves the noisy memory FBSDE (38) and (39) and q2(t)=tt+δE[Dt[HΛ(s)]|Ft]ds.

Proof.

The jump terms do not make a difference here, so assume for simplicity that r=r1=r2=0 everywhere.

In general, we know that if dp2(t)=θ(t,p2,q2)dt+q2(t)dB(t),p2(T)=F, then (44) q2(t)=Dtp2(t).(44)

Now, note that the solution p2 of the BSDE (42) can be written p2(t)=E[tTE[μ2(s)|Fs]ds|Ft]=tTE[μ2(s)|Ft]ds=tTE[HΛ(t)HΛ(t+δ)1[0,Tδ](t)|Ft]ds=tt+δE[H(s)Λ|Ft]1[0,T](s)ds, where the equalities follow from Fubini’s theorem, the rule of double expectation, the definition of μ2 and a change of variables. Hence, by EquationEquation (44): q2(t)=Dtp2(t)=Dt[tt+δE[H(s)Λ|Ft]1[0,T](s)]ds=tt+δE[Dt(H(s)Λ)|Ft]1[0,T](s)ds, which is part of what we wanted to prove.

By inserting this expression for q2 into the definition of μ1, we see that μ1(t)=tt+δE[Dt[H(s)Λ]|Ft]1[0,T](s)ds+H(t)x+H(t+δ)y1[0,T](t+δ).

Hence, we see that the BSDE (41) is the same as (39), so they have the same solution. This completes the proof of the theorem. □

We can also prove the following converse result.

Theorem 5.2.

If p,q,r,λ solve the FBSDE (38) and (39) and we define λ˜=λ,p1=p,q1=q,r1=r and p2(t)=tt+δE[HΛ(s)|Ft]1[0,Tδ](s)ds,q2(t)=tt+δE[Dt[HΛ(s)]|Ft]1[0,Tδ](s)ds,r2(t,·)=0.

Then, (pi,qi,ri)fori=1,2andλ˜ solve the system of EquationEquations (40)–(42).

Proof. A

gain, the jump parts make no crucial difference, so we consider the no-jump situation for simplicity.

It is clear that EquationEquation (40) holds from the assumptions above (from the definition of H, see (43)). Also, the BSDE (41) holds: Clearly, the terminal condition holds, and by the computations in the proof of Theorem 5.1, the remaining part of EquationEquation (41) also holds. Therefore, it only remains to prove that the BSDE (42) holds.

By the Itô isometry and the Clark–Ocone formula, E[0TE[Ds(H(r)Λ)|Fs]2ds]=E[(0TE[DsH(r)Λ|Fs]dBs)2]=E[(HΛ(r))2E[HΛ(r)]2].

Hence, 0TE[0TE[Ds(H(r)Λ)|Fs]2ds]12dr=0T(E[HΛ(r)2]E[HΛ(r)]2)12dt<.

Note that from the Clark–Ocone theorem, H(r)Λ=E[H(r)Λ|Ft]+trE[Ds(H(r)Λ)|Fs]dB(s).

Therefore, by the definition of q2 in the theorem and the Fubini theorem tTq2(s)dB(s)=tTtTE[Ds(H(r)Λ)|Fs]1[s,s+δ](r)drdB(s)=tTtTE[Ds(H(r)Λ)|Fs]1[rδ,r](s)dB(s)dr.

By some algebra and the Clark–Ocone theorem Equation(16), tTtTE[Ds(H(r)Λ)|Fs]1[rδ,r](s)dB(s)dr=tTrδrE[Ds(H(r)Λ)|Fs]dB(s)dr=tT(H(r)ΛE[H(r)Λ|Frδ])dr.

By splitting the integrals and using change of variables (twice) as well as some algebra, =tTH(s)ΛdstδTδE[H(s+δ)Λ|Fs]ds=tTH(s)ΛdstTE[H(s+δ)Λ|Fs]1[0,Tδ](s)dstt+δE[H(s)Λ|Ft]1[0,Tδ](s)ds=tTE[H(s)ΛH(s+δ)Λ1[0,Tδ](s)|Fs]dsp2(t).

This proves that the BSDE (42) holds as well. □

Now, we have expressed the solution of the Malliavin FBSDE via the solution of the “double” FBSDE system (40)–(42). What kind of system of equations is this? The system consists of two connected BSDEs in (p1,q1,r1)and(p2,q2,r2) respectively, and these are again connected to a FBSDE in λ. However, from EquationEquation (42) and the definition of μ2, we see that the right hand side of (42) does not depend on p2. Hence, the BSDE (42) can be rewritten dp2(t)=h(t,λ,p1,q1,r1(·))dt+q2(t)dB(t)+Rr2(t,ζ)N˜(dt,dζ)p2(T)=0.

This can be solved to express p2usingλ,p1,q1andr1(·) by letting q2(t)=r2(t,·)=0 for all t and p2(t)=E[tTh(t,λ,p1,q1,r1(·))dt|Ft].

Now, we can substitute this solution for p2(t) into the FBSDE system (40) and (41). The resulting set of equations is a regular system of time advanced FBSDEs with jumps. There are to the best of our knowledge, no general results on existence and uniqueness of such systems of FBSDEs. However, if we simplify by removing the jumps and there was no time-advanced part (i.e., no delay process Yi in the original FSDE (1)), there are some results by Ma et al. [Citation24].

6. Optimal consumption rate with respect to recursive utility

In this section, we apply the previous results to the problem of determining an optimal consumption rate with respect to recursive utility (see also Øksendal and Sulem [Citation25] and Dahl and Øksendal [Citation26]). Let X(t)=Xc(t), where the consumption rate c(t) is our control, and assume that (45) dX(t)=X(t)[μ(t)dt+σ(t)dB(t)+Rγ(t,ζ)N˜(dt,dζ)][c1(t)+c2(t)]X(t)dt,X(0)=x>0,(45) and Wi(t) is given by dWi(t)=[αi(t)Wi(t)+ηi(t)ln(Yi(t))+κi(t)ln(Λi(t))+ln(ci(t)X(t))]+Zi(t)dB(t)+RKi(t,ζ)N˜(dt,dζ),Wi(T)=0.

Let the performance functional be defined by Ji(c1,c2):=Wi(0), i.e., Ji is the recursive utility for player i. Also, assume that both players have full information, so (Et(i))t=(Ft)tfori=1,2.

We would like to find a Nash equilibrium for this FBSDE game with delay. To do so we will use the maximum principle Theorem 3.1. Note that fi=φi=hi=0 and that ψi(w)=wfori=1,2. The Hamiltonians are: Hi(t,x,y1,y2,Λ1,Λ2,wi,zi,ki,c1,c2,λi,pi,qi,ri(ζ))=λi(αi(t)wi+ηi(t)ln(yi)+ln(cix))+pi(xμ(t)(c1+c2)x)+qiσ(t)x+Rxri(ζ)γ(t,ζ)ν(dζ) for i=1,2.

The adjoint BSDEs are dpi(t)=E[μi(t)|Ft]dt+qi(t)dB(t)+Rri(t,ζ)N˜(dt,dζ),pi(T)=0, where μi(t)=λi(t)X(t)λi(t+δi)ηi(t+δi)Yi(t+δi)1[0,Tδi](t)pi(t)(μ(t)(c1(t)+c2(t)))+qi(t)σ(t)+Rri(t,ζ)γ(t,ζ)ν(dζ) for i=1,2. Note that by the definition of Yi,Yi(t+δi)=X({t+δi}δi)=X(t).

The adjoint BSDEs are linear, and the solutions are given by (see Øksendal and Sulem [Citation15]) (46) Γi(t)pi(t)=E[tT(λi(s)X(s)+λi(s+δi)ηi(s+δi)Yi(s+δi)1[0,Tδi](s))Γi(s)ds|Ft]=E[tT(λi(s)X(s)+λi(s+δi)ηi(s+δi)X(s)1[0,Tδi](s))Γi(s)ds|Ft],(46) where dΓi(t)=Γi(t)[(μ(t)(c1(t)+c2(t)))dt+σ(t)dB(t)+Rγ(t,ζ)N˜(dt,dζ)],Γi(0)=1 for i=1,2.

Note that by the SDE (45), (47) xΓi(t)=X(t).(47)

Hence, by combining EquationEquations (46) and Equation(47), we see that (48) X(t)pi(t)=E[tT(λi(s)+λi(s+δi)ηi(s+δi)1[0,Tδi](s))ds|Ft].(48)

The adjoint FSDEs are dλi(t)=λi(t)αi(t)dt,λi(0)=1, for i=1,2.

These are (non-stochastic) differential equation with solution λi(t)=exp(0tαi(s)ds)fori=1,2.

We maximize Hi with respect to ci. For i=1,2, the first order condition is: ĉi(t)=λi(t)pi(t)X(t).

By substituting EquationEquation (48) into this, we find (by the sufficient maximum principle, Theorem 3.1) that the consumption rates leading to a Nash equilibrium for the recursive utility problem are given by: ci*(t)=λi(t)E[tT(λi(s)+λi(s+δi)ηi(s+δi)1[0,Tδi](t))ds|Ft]. where λi(t)=exp(0tαi(s)ds)fori=1,2.

7. Conclusion

In this paper, we have analyzed a two-player stochastic game connected to a set of FBSDEs involving delay and noisy memory of the market process. We have derived sufficient and necessary maximum principles for a set of controls for the two players to be a Nash equilibrium in this game. We have also studied the associated FBSDE involving Malliavin derivatives, and connected this to a system of FBSDEs not involving Malliavin derivatives. Finally, we were able to derive a closed form Nash equilibrium solution to a game where the aim is to find the optimal consumption with respect to recursive utility.

Notes

1 Defined similarly as subdifferentiability for convex functions.

References

  • Ivanov, A. F., Kazmerchuk, Y. I., Swishchuk, A. V. Theory, stochastic stability and applications of stochastic delay differential equations: a survey of recent results. Research report, http://www.math.yorku.ca/aswishch/sddesurvey.pdf.
  • Øksendal, B., Sulem, A., Zhang, T. (2011). Optimal control of stochastic delay equations and time-advanced backward stochastic differential equations. Adv. Appl. Probab. 43(2):572–596. DOI: 10.1239/aap/1308662493.
  • Agram, N., Øksendal, B. (2014). Infinite horizon optimal control of forward-backward stochastic differential equations with delay. J. Comput. Appl. Math. 259 (B):336–349. DOI: 10.1016/j.cam.2013.04.048.
  • Dahl, K., Mohammed, S., Øksendal, B., Røse, E. (2016). Optimal control of systems with noisy memory and BSDEs with malliavin derivatives. J. Funct. Anal. 271(2):289–329. DOI: 10.1016/j.jfa.2016.04.031.
  • Krylov, N. V. (2009). Controlled Diffusion Processes. Berlin, Heidelberg: Springer.
  • Krylov, N. V. (1972). Control of a solution of a stochastic integral equation. Theory Probab. Appl. 17(1):114–130. DOI: 10.1137/1117009.
  • Øksendal, B., Sulem, A. (2014). Forward-backward stochastic differential games and stochastic control under model uncertainty. J. Optim. Theory Appl. 22 (161):22–55. DOI: 10.1007/s10957-012-0166-7.
  • Wang, S., Wu, Z. (2017). Stochastic maximum principle for optimal control problems of forward-backward delay systems involving impulse controls. J. Syst. Sci. Complex. 30(2):280–306. DOI: 10.1007/s11424-016-5039-y.
  • Wu, Z. (2013). A general maximum principle for optimal control of forward-backward stochastic systems. Automatica 49 (5):1473–1480. DOI: 10.1016/j.automatica.2013.02.005.
  • Wang, G., Wu, Z., Xiong, J. (2013). Maximum principles for forward-backward stochastic control systems with correlated state and observation noises. SIAM J. Control Optim. 51 (1):491–524. DOI: 10.1137/110846920.
  • Øksendal, B. (2013). Stochastic Differential Equations: An Introduction with Applications. Berlin, Heidelberg: Springer.
  • Øksendal, B., Sulem, A. (2000). A maximum principle for optimal control of stochastic systems with delay, with applications to finance. In: J.M. Menaldi, E. Rofman, A. Sulem, eds. Optimal Control and Partial Differential Equations—Innovations and Applications. Amsterdam: IOS Press.
  • Di Nunno, G., Øksendal, B., Proske, F. (2009). Malliavin Calculus for Lévy Processes with Applications to Finance. Berlin, Heidelberg: Springer.
  • Øksendal, B., Sulem, A. (2007). Applied Stochastic Control of Jump Diffusions. Berlin, Heidelberg: Springer.
  • Øksendal, B., Sulem, A. (2014). Risk minimization in financial markets modeled by Itô Lévy processes. Research report, Department of Mathematics, University of Oslo.
  • Aase, K., Øksendal, B., Privault, N., Ubøe, J. (2000). White noise generalizations of the Clark–Haussmann–Ocone theorem with application to mathematical finance. Finance Stochast. Stochastics4 (4):465–496. DOI: 10.1007/PL00013528.
  • Nualart, D. (2006). The Malliavin Calculus and Related Topics. Springer, Berlin Heidelberg.
  • Sanz-Solè, M. (2005). Malliavin Calculus. Lausanne: EPFL Press.
  • Agram, N., Øksendal, B. (2015). Malliavin calculus and optimal control of stochastic Volterra equations. J. Optim. Theory Appl. 167(3):1070–1094. DOI: 10.1007/s10957-015-0753-5.
  • Pham, H. (2009). Continuous-time stochastic control and optimization with financial applications. Stochastic Modell. Appl. Probab. 61:139–169.
  • Rockafellar, R. T. (1970). Convex Analysis. Princeton: Princeton University Press.
  • Di Nunno, G., Øksendal, B., Proske, F. (2004). White noise analysis for Lévy processes. J. Funct. Anal. 206(1):109–148. DOI: 10.1016/S0022-1236(03)00184-8.
  • Shilling, R. L. (2005). Measures, Integrals and Martingales. Cambridge: Cambridge University Press.
  • Ma, J., Yin, H., Zhang, J. (2012). On non-Markovian forward-backward SDEs and backward stochastic PDEs. Stochastic Processes Their Appl. 122(12):3980–4004. DOI: 10.1016/j.spa.2012.08.002.
  • Øksendal, B., Sulem, A. (2016). Optimal control of predictive mean-field equations and applications to finance, Eds. Benth F., Di Nunno G.: Stochastics of Environmental and Financial Economics, Springer Proceedings in Mathematics & Statistics, Vol. 138. Cham: Springer.
  • Dahl, K., Øksendal, B. (2017). Singular recursive utility. Stochastics 89(6–7):994–1014. DOI: 10.1080/17442508.2017.1303067.