Abstract

The Merton problem determines the optimal intertemporal portfolio choice by maximizing the expected utility and is the basis of modern portfolio theory in continuous-time finance. However, its empirical performance is disappointing. The estimation errors of the expected rates of returns make the optimal policy degenerate, resulting in an extremely low (or unbounded) expected utility value for a high-dimensional portfolio. We further prove that the estimation error of the variance-covariance matrix leads to the degenerated policy of solely investing in the risk-free asset. This study proposes a constrained $\ell_1$-minimization approach to resolve the degeneracy in the high-dimensional setting and stabilize the performance in the low-dimensional setting. The proposed scheme can be implemented with simple linear programming and involves negligible additional computational time, compared to standard estimation. We prove the consistency of our framework that our estimate of the optimal control tends to be the true one. We also derive the rate of convergence. Simulation studies are provided to verify the finite-sample properties. An empirical study using S&P 500 component stock data demonstrates the superiority of the proposed approach.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call