Abstract
We develop a two-step estimator for a class of Markov decision processes with continuous control that is intuitive and simple to implement. Making use of the monotonicity assumption we estimate the expected continuation value functions nonparametrically in the first stage. In the second stage our estimator minimizes a minimum distance criterion that measures the divergence between the nonparametric conditional distribution function and a model implied simulated semiparametric counterpart. We show that our minimum distance estimator is asymptotically normal and converges at the parametric rate under some regularity conditions. We estimate the expected value function by kernel smoothing and derive its pointwise distribution theory. We illustrate how our estimation methodology forms a basis for the estimation of dynamic models with different class of control variable(s) as well as a class of Markovian games.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.