arXiv Analytics

Sign in

arXiv:1805.04686 [cs.LG]AbstractReferencesReviewsResources

Adversarial Task Transfer from Preference

Xiaojian Ma, Mingxuan Jing, Fuchun Sun, Huaping Liu

Published 2018-05-12Version 1

Task transfer is extremely important for reinforcement learning, since it provides possibility for generalizing to new tasks. One main goal of task transfer in reinforcement learning is to transfer the action policy of an agent from the original basic task to specific target task. Existing work to address this challenging problem usually requires accurate hand-coded cost functions or rich demonstrations on the target task. This strong requirement is difficult, if not impossible, to be satisfied in many practical scenarios. In this work, we develop a novel task transfer framework which effectively performs the policy transfer using preference only. The hidden cost model for preference and adversarial training are elegantly combined to perform the task transfer. We give the theoretical analysis on the convergence about the proposed algorithm, and perform extensive simulations on some well-known examples to validate the theoretical results.

Related articles: Most relevant | Search more
arXiv:2403.11782 [cs.LG] (Published 2024-03-18, updated 2024-03-24)
A tutorial on learning from preferences and choices with Gaussian Processes
arXiv:2205.03699 [cs.LG] (Published 2022-05-07)
Rate-Optimal Contextual Online Matching Bandit
arXiv:1811.09751 [cs.LG] (Published 2018-11-24)
Characterizing and Avoiding Negative Transfer