arXiv Analytics

Sign in

arXiv:1906.03574 [cs.LG]AbstractReferencesReviewsResources

Transfer Learning by Modeling a Distribution over Policies

Disha Shrivastava, Eeshan Gunesh Dhekane, Riashat Islam

Published 2019-06-09Version 1

Exploration and adaptation to new tasks in a transfer learning setup is a central challenge in reinforcement learning. In this work, we build on the idea of modeling a distribution over policies in a Bayesian deep reinforcement learning setup to propose a transfer strategy. Recent works have shown to induce diversity in the learned policies by maximizing the entropy of a distribution of policies (Bachman et al., 2018; Garnelo et al., 2018) and thus, we postulate that our proposed approach leads to faster exploration resulting in improved transfer learning. We support our hypothesis by demonstrating favorable experimental results on a variety of settings on fully-observable GridWorld and partially observable MiniGrid (Chevalier-Boisvert et al., 2018) environments.

Comments: Accepted at the ICML 2019 workshop on Multi-Task and Lifelong Reinforcement Learning
Categories: cs.LG, cs.AI, stat.ML
Related articles: Most relevant | Search more
arXiv:2107.10199 [cs.LG] (Published 2021-07-21)
Distribution of Classification Margins: Are All Data Equal?
arXiv:2006.10096 [cs.LG] (Published 2020-06-17)
Towards Recurrent Autoregressive Flow Models
arXiv:1603.02501 [cs.LG] (Published 2016-03-08)
Mixture Proportion Estimation via Kernel Embedding of Distributions