arXiv Analytics

Sign in

arXiv:1709.06466 [math.OC]AbstractReferencesReviewsResources

Evaluation of the Rate of Convergence in the PIA

Jun Maeda, Saul D. Jacka

Published 2017-09-19Version 1

Folklore says that Howard's Policy Improvement Algorithm converges extraordinarily fast, even for controlled diffusion settings. In a previous paper, we proved that approximations of the solution of a particular parabolic partial differential equation obtained via the policy improvement algorithm show a quadratic local convergence. In this paper, we show that we obtain the same rate of convergence of the algorithm in a more general setup. This provides some explanation as to why the algorithm converges fast. We provide an example by solving a semilinear elliptic partial differential equation numerically by applying the algorithm and check how the approximations converge to the analytic solution.

Related articles: Most relevant | Search more
arXiv:1310.7063 [math.OC] (Published 2013-10-26, updated 2015-07-01)
On the Convergence of Decentralized Gradient Descent
arXiv:0803.2211 [math.OC] (Published 2008-03-14, updated 2010-05-09)
On Conditions for Convergence to Consensus
arXiv:1801.08691 [math.OC] (Published 2018-01-26)
On Quasi-Newton Forward--Backward Splitting: Proximal Calculus and Convergence