Abstract

A classical problem in ergodic control theory consists in the study of the limit behaviour of λVλ(⋅) as λ↘0, when Vλ is the value function of a deterministic or stochastic control problem with discounted cost functional with infinite time horizon and discount factor λ. We study this problem for the lower value function Vλ of a stochastic differential game with recursive cost, i.e., the cost functional is defined through a backward stochastic differential equation with infinite time horizon. But unlike the ergodic control approach, we are interested in the case where the limit can be a function depending on the initial condition. For this we extend the so-called non-expansivity assumption from the case of control problems to that of stochastic differential games and we derive that λVλ(⋅) is bounded and Lipschitz uniformly with respect to λ>0. Using PDE methods and assuming radial monotonicity of the Hamiltonian of the associated Hamilton-Jacobi-Bellman-Isaacs equation we obtain the monotone convergence of λVλ(.) and we characterize its limit W0 as maximal viscosity subsolution of a limit PDE. Using BSDE methods we prove that W0 satisfies a uniform dynamic programming principle involving the supremum and the infimum with respect to the time, and this is the key for an explicit representation formula for W0.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call