arXiv Analytics

Sign in

arXiv:1811.11441 [cs.LG]AbstractReferencesReviewsResources

Trajectory-based Learning for Ball-in-Maze Games

Sujoy Paul, Jeroen van Baar

Published 2018-11-28Version 1

Deep Reinforcement Learning has shown tremendous success in solving several games and tasks in robotics. However, unlike humans, it generally requires a lot of training instances. Trajectories imitating to solve the task at hand can help to increase sample-efficiency of deep RL methods. In this paper, we present a simple approach to use such trajectories, applied to the challenging Ball-in-Maze Games, recently introduced in the literature. We show that in spite of not using human-generated trajectories and just using the simulator as a model to generate a limited number of trajectories, we can get a speed-up of about 2-3x in the learning process. We also discuss some challenges we observed while using trajectory-based learning for very sparse reward functions.

Related articles: Most relevant | Search more
arXiv:2203.17193 [cs.LG] (Published 2022-03-31)
Learning from many trajectories
arXiv:2205.00403 [cs.LG] (Published 2022-05-01)
A Simple Approach to Improve Single-Model Deep Uncertainty via Distance-Awareness
arXiv:1811.00128 [cs.LG] (Published 2018-10-31)
Towards a Simple Approach to Multi-step Model-based Reinforcement Learning