arXiv Analytics

Sign in

arXiv:1905.00741 [cs.LG]AbstractReferencesReviewsResources

From Video Game to Real Robot: The Transfer between Action Spaces

Janne Karttunen, Anssi Kanervisto, Ville Hautamäki, Ville Kyrki

Published 2019-05-02Version 1

Training agents with reinforcement learning based techniques requires thousands of steps, which translates to long training periods when applied to robots. By training the policy in a simulated environment we avoid such limitation. Typically, the action spaces in a simulation and real robot are kept as similar as possible, but if we want to use a generic simulation environment, this strategy will not work. Video games, such as Doom (1993), offer a crude but multi-purpose environments that can used for learning various tasks. However, original Doom has four discrete actions for movement and the robot in our case has two continuous actions. In this work, we study the transfer between these two different action spaces. We begin with experiments in a simulated environment, after which we validate the results with experiments on a real robot. Results show that fine-tuning initially learned network parameters leads to unreliable results, but by keeping most of the neural network frozen we obtain above $90\%$ success rate in simulation and real robot experiments.

Comments: Two first authors contributed equally
Categories: cs.LG, cs.AI, cs.RO
Related articles: Most relevant | Search more
arXiv:2212.07946 [cs.LG] (Published 2022-12-15)
Combining information-seeking exploration and reward maximization: Unified inference on continuous state and action spaces under partial observability
arXiv:2211.16715 [cs.LG] (Published 2022-11-30)
Policy Optimization over General State and Action Spaces
arXiv:1906.11228 [cs.LG] (Published 2019-06-26)
Regularized Hierarchical Policies for Compositional Transfer in Robotics