arXiv Analytics

Sign in

arXiv:1906.05363 [cs.LG]AbstractReferencesReviewsResources

Competing Bandits in Matching Markets

Lydia T. Liu, Horia Mania, Michael I. Jordan

Published 2019-06-12Version 1

Stable matching, a classical model for two-sided markets, has long been studied with little consideration for how each side's preferences are learned. With the advent of massive online markets powered by data-driven matching platforms, it has become necessary to better understand the interplay between learning and market objectives. We propose a statistical learning model in which one side of the market does not have a priori knowledge about its preferences for the other side and is required to learn these from stochastic rewards. Our model extends the standard multi-armed bandits framework to multiple players, with the added feature that arms have preferences over players. We study both centralized and decentralized approaches to this problem and show surprising exploration-exploitation trade-offs compared to the single player multi-armed bandits setting.

Related articles: Most relevant | Search more
arXiv:2204.12048 [cs.LG] (Published 2022-04-26)
Thompson Sampling for Bandit Learning in Matching Markets
arXiv:2210.11692 [cs.LG] (Published 2022-10-21)
Competing Bandits in Time Varying Matching Markets
arXiv:1902.04228 [cs.LG] (Published 2019-02-12)
Multi-objective Bayesian optimisation with preferences over objectives