This paper is concerned with the switching game of a one-dimensional backward stochastic differential equation (BSDE). The associated Bellman-Isaacs equation is a system of matrix-valued BSDEs living in a special unbounded convex domain with reflection on the boundary along an oblique direction. In this paper, we show the existence of an adapted solution to this system of BSDEs with oblique reflection by the penalization method, the monotone convergence, and the a priori estimates.
In [5] the authors obtained Mean-Field backward stochastic differential equations (BSDE) associated with a Mean-field stochastic differential equation (SDE) in a natural way as limit of some highly dimensional system of forward and backward SDEs, corresponding to a large number of ``particles (or ``agents). The objective of the present paper is to deepen the investigation of such Mean-Field BSDEs by studying them in a more general framework, with general driver, and to discuss comparison results for them. In a second step we are interested in partial differential equations (PDE) whose solutions can be stochastically interpreted in terms of Mean-Field BSDEs. For this we study a Mean-Field BSDE in a Markovian framework, associated with a Mean-Field forward equation. By combining classical BSDE methods, in particular that of ``backward semigroups introduced by Peng [14], with specific arguments for Mean-Field BSDEs we prove that this Mean-Field BSDE describes the viscosity solution of a nonlocal PDE. The uniqueness of this viscosity solution is obtained for the space of continuous functions with polynomial growth. With the help of an example it is shown that for the nonlocal PDEs associated to Mean-Field BSDEs one cannot expect to have uniqueness in a larger space of continuous functions.
In this paper, we study the following nonlinear backward stochastic integral partial differential equation with jumps begin{equation*} left{ begin{split} -d V(t,x) =&displaystyleinf_{uin U}bigg{H(t,x,u, DV(t,x),D Phi(t,x), D^2 V(t,x),int_E left(mathcal I V(t,e,x,u)+Psi(t,x+g(t,e,x,u))right)l(t,e) u(de)) &+displaystyleint_{E}big[mathcal I V(t,e,x,u)-displaystyle (g(t, e,x,u), D V(t,x))big] u(d e)+int_{E}big[mathcal I Psi(t,e,x,u)big] u(d e)bigg}dt &-Phi(t,x)dW(t)-displaystyleint_{E} Psi (t, e,x)tildemu(d e,dt), V(T,x)=& h(x), end{split} right. end{equation*} where $tilde mu$ is a Poisson random martingale measure, $W$ is a Brownian motion, and $mathcal I$ is a non-local operator to be specified later. The function $H$ is a given random mapping, which arises from a corresponding non-Markovian optimal control problem. This equation appears as the stochastic Hamilton-Jacobi-Bellman equation, which characterizes the value function of the optimal control problem with a recursive utility cost functional. The solution to the equation is a predictable triplet of random fields $(V,Phi,Psi)$. We show that the value function, under some regularity assumptions, is the solution to the stochastic HJB equation; and a classical solution to this equation is the value function and gives the optimal control. With some additional assumptions on the coefficients, an existence and uniqueness result in the sense of Sobolev space is shown by recasting the backward stochastic partial integral differential equation with jumps as a backward stochastic evolution equation in Hilbert spaces with Poisson jumps.
In this paper we are concerned with a new type of backward equations with anticipation which we call neutral backward stochastic functional differential equations. We obtain the existence and uniqueness and prove a comparison theorem. As an application, we discuss the optimal control of neutral stochastic functional differential equations, establish a Pontryagin maximum principle, and give an explicit optimal value for the linear optimal control.
We study the problem of optimal inside control of an SPDE (a stochastic evolution equation) driven by a Brownian motion and a Poisson random measure. Our optimal control problem is new in two ways: (i) The controller has access to inside information, i.e. access to information about a future state of the system, (ii) The integro-differential operator of the SPDE might depend on the control. In the first part of the paper, we formulate a sufficient and a necessary maximum principle for this type of control problem, in two cases: (1) When the control is allowed to depend both on time t and on the space variable x. (2) When the control is not allowed to depend on x. In the second part of the paper, we apply the results above to the problem of optimal control of an SDE system when the inside controller has only noisy observations of the state of the system. Using results from nonlinear filtering, we transform this noisy observation SDE inside control problem into a full observation SPDE insider control problem. The results are illustrated by explicit examples.