Cooperative and Stochastic Multi-Player Multi-Armed Bandit: Optimal Regret With Neither Communication Nor Collisions

Sébastien Bubeck, Thomas Budzinski, Mark Sellke. Cooperative and Stochastic Multi-Player Multi-Armed Bandit: Optimal Regret With Neither Communication Nor Collisions. In Mikhail Belkin, Samory Kpotufe, editors, Conference on Learning Theory, COLT 2021, 15-19 August 2021, Boulder, Colorado, USA. Volume 134 of Proceedings of Machine Learning Research, pages 821-822, PMLR, 2021. [doi]

Abstract

Abstract is missing.