arXiv Analytics

Sign in

arXiv:2402.08621 [cs.LG]AbstractReferencesReviewsResources

A Generalized Approach to Online Convex Optimization

Mohammad Pedramfar, Vaneet Aggarwal

Published 2024-02-13, updated 2024-05-13Version 2

In this paper, we analyze the problem of online convex optimization in different settings. We show that any algorithm for online linear optimization with fully adaptive adversaries is an algorithm for online convex optimization. We also show that any such algorithm that requires full-information feedback may be transformed to an algorithm with semi-bandit feedback with comparable regret bound. We further show that algorithms that are designed for fully adaptive adversaries using deterministic semi-bandit feedback can obtain similar bounds using only stochastic semi-bandit feedback when facing oblivious adversaries. We use this to describe general meta-algorithms to convert first order algorithms to zeroth order algorithms with comparable regret bounds. Our framework allows us to analyze online optimization in various settings, such full-information feedback, bandit feedback, stochastic regret, adversarial regret and various forms of non-stationary regret.

Related articles: Most relevant | Search more
arXiv:2102.03803 [cs.LG] (Published 2021-02-07)
Lazy OCO: Online Convex Optimization on a Switching Budget
arXiv:2310.11880 [cs.LG] (Published 2023-10-18)
Online Convex Optimization with Switching Cost and Delayed Gradients
arXiv:1804.04529 [cs.LG] (Published 2018-04-12)
Online convex optimization and no-regret learning: Algorithms, guarantees and applications