arXiv Analytics

Sign in

arXiv:1811.00128 [cs.LG]AbstractReferencesReviewsResources

Towards a Simple Approach to Multi-step Model-based Reinforcement Learning

Kavosh Asadi, Evan Cater, Dipendra Misra, Michael L. Littman

Published 2018-10-31Version 1

When environmental interaction is expensive, model-based reinforcement learning offers a solution by planning ahead and avoiding costly mistakes. Model-based agents typically learn a single-step transition model. In this paper, we propose a multi-step model that predicts the outcome of an action sequence with variable length. We show that this model is easy to learn, and that the model can make policy-conditional predictions. We report preliminary results that show a clear advantage for the multi-step model compared to its one-step counterpart.

Related articles: Most relevant | Search more
arXiv:2205.00403 [cs.LG] (Published 2022-05-01)
A Simple Approach to Improve Single-Model Deep Uncertainty via Distance-Awareness
arXiv:1903.10145 [cs.LG] (Published 2019-03-25)
Cyclical Annealing Schedule: A Simple Approach to Mitigating KL Vanishing
arXiv:2306.05401 [cs.LG] (Published 2023-06-08)
RDumb: A simple approach that questions our progress in continual test-time adaptation