arXiv Analytics

Sign in

arXiv:1805.10129 [cs.LG]AbstractReferencesReviewsResources

Dyna Planning using a Feature Based Generative Model

Ryan Faulkner, Doina Precup

Published 2018-05-23Version 1

Dyna-style reinforcement learning is a powerful approach for problems where not much real data is available. The main idea is to supplement real trajectories, or sequences of sampled states over time, with simulated ones sampled from a learned model of the environment. However, in large state spaces, the problem of learning a good generative model of the environment has been open so far. We propose to use deep belief networks to learn an environment model for use in Dyna. We present our approach and validate it empirically on problems where the state observations consist of images. Our results demonstrate that using deep belief networks, which are full generative models, significantly outperforms the use of linear expectation models, proposed in Sutton et al. (2008)

Comments: 8 pages, 7 figures
Journal: 24th Annual Proceedings of the Advances in Neural Information Processing Systems (2010) pp. 1-9
Categories: cs.LG, cs.AI, cs.CV, stat.ML
Related articles: Most relevant | Search more
arXiv:2203.11132 [cs.LG] (Published 2022-03-21)
Review of Disentanglement Approaches for Medical Applications -- Towards Solving the Gordian Knot of Generative Models in Healthcare
arXiv:1812.09111 [cs.LG] (Published 2018-12-21)
Generative Models from the perspective of Continual Learning
arXiv:1804.09858 [cs.LG] (Published 2018-04-26)
Generative Model for Heterogeneous Inference