arXiv Analytics

Sign in

arXiv:2102.09305 [cs.LG]AbstractReferencesReviewsResources

Boosting for Online Convex Optimization

Elad Hazan, Karan Singh

Published 2021-02-18Version 1

We consider the decision-making framework of online convex optimization with a very large number of experts. This setting is ubiquitous in contextual and reinforcement learning problems, where the size of the policy class renders enumeration and search within the policy class infeasible. Instead, we consider generalizing the methodology of online boosting. We define a weak learning algorithm as a mechanism that guarantees multiplicatively approximate regret against a base class of experts. In this access model, we give an efficient boosting algorithm that guarantees near-optimal regret against the convex hull of the base class. We consider both full and partial (a.k.a. bandit) information feedback models. We also give an analogous efficient boosting algorithm for the i.i.d. statistical setting. Our results simultaneously generalize online boosting and gradient boosting guarantees to contextual learning model, online convex optimization and bandit linear optimization settings.

Related articles: Most relevant | Search more
arXiv:2205.10927 [cs.LG] (Published 2022-05-22)
Fast ABC-Boost: A Unified Framework for Selecting the Base Class in Multi-Class Classification
arXiv:2009.14436 [cs.LG] (Published 2020-09-30)
Online Convex Optimization in Changing Environments and its Application to Resource Allocation
arXiv:1909.05207 [cs.LG] (Published 2019-09-07)
Introduction to Online Convex Optimization