arXiv Analytics

Sign in

arXiv:2012.07348 [cs.LG]AbstractReferencesReviewsResources

Bandit Learning in Decentralized Matching Markets

Lydia T. Liu, Feng Ruan, Horia Mania, Michael I. Jordan

Published 2020-12-14, updated 2020-12-31Version 2

We study two-sided matching markets in which one side of the market (the players) does not have a priori knowledge about its preferences for the other side (the arms) and is required to learn its preferences from experience. Also, we assume the players have no direct means of communication. This model extends the standard stochastic multi-armed bandit framework to a decentralized multiple player setting with competition. We introduce a new algorithm for this setting that, over a time horizon $T$, attains $\mathcal{O}(\log(T))$ stable regret when preferences of the arms over players are shared, and $\mathcal{O}(\log(T)^2)$ regret when there are no assumptions on the preferences on either side.

Related articles: Most relevant | Search more
arXiv:1802.05693 [cs.LG] (Published 2018-02-15)
Bandit Learning with Positive Externalities
arXiv:2201.01902 [cs.LG] (Published 2022-01-06, updated 2022-01-30)
Gaussian Imagination in Bandit Learning
arXiv:1907.01287 [cs.LG] (Published 2019-07-02)
Bandit Learning Through Biased Maximum Likelihood Estimation