arXiv Analytics

Sign in

arXiv:2212.08765 [cs.LG]AbstractReferencesReviewsResources

Latent Variable Representation for Reinforcement Learning

Tongzheng Ren, Chenjun Xiao, Tianjun Zhang, Na Li, Zhaoran Wang, Sujay Sanghavi, Dale Schuurmans, Bo Dai

Published 2022-12-17Version 1

Deep latent variable models have achieved significant empirical successes in model-based reinforcement learning (RL) due to their expressiveness in modeling complex transition dynamics. On the other hand, it remains unclear theoretically and empirically how latent variable models may facilitate learning, planning, and exploration to improve the sample efficiency of RL. In this paper, we provide a representation view of the latent variable models for state-action value functions, which allows both tractable variational learning algorithm and effective implementation of the optimism/pessimism principle in the face of uncertainty for exploration. In particular, we propose a computationally efficient planning algorithm with UCB exploration by incorporating kernel embeddings of latent variable models. Theoretically, we establish the sample complexity of the proposed approach in the online and offline settings. Empirically, we demonstrate superior performance over current state-of-the-art algorithms across various benchmarks.

Related articles: Most relevant | Search more
arXiv:1706.04711 [cs.LG] (Published 2017-06-15)
Reinforcement Learning under Model Mismatch
arXiv:1301.0601 [cs.LG] (Published 2012-12-12)
Reinforcement Learning with Partially Known World Dynamics
arXiv:1306.6189 [cs.LG] (Published 2013-06-26)
Scaling Up Robust MDPs by Reinforcement Learning