• Corpus ID: 221089750

# Cooperative Multi-Agent Bandits with Heavy Tails

@inproceedings{Dubey2020CooperativeMB,
title={Cooperative Multi-Agent Bandits with Heavy Tails},
author={Abhimanyu Dubey and Alex 'Sandy' Pentland},
booktitle={ICML},
year={2020}
}
• Published in ICML 14 August 2020
• Computer Science
We study the heavy-tailed stochastic bandit problem in the cooperative multi-agent setting, where a group of agents interact with a common bandit problem, while communicating on a network with delays. Existing algorithms for the stochastic bandit in this setting utilize confidence intervals arising from an averaging-based communication protocol known as~\textit{running consensus}, that does not lend itself to robust estimation for heavy-tailed settings. We propose \textsc{MP-UCB}, a…

## Figures from this paper

Distributed Bandits with Heterogeneous Agents
• Computer Science
IEEE INFOCOM 2022 - IEEE Conference on Computer Communications
• 2022
This paper proposes two learning algorithms, CO-UCB and CO-AAE, and proves that both algorithms achieve order-optimal regret, which is O(log T), where O is the minimum suboptimality gap between the reward mean of arm i and any local optimal arm.
Cooperative Stochastic Bandits with Asynchronous Agents and Constrained Feedback
• Computer Science
NeurIPS
• 2021
AAE-LCB is proposed, a two-stage algorithm that prioritizes pulling local arms following an active arm elimination policy, and switches to other arms only if all local arms are dominated by some external arms.
Cooperative Stochastic Multi-agent Multi-armed Bandits Robust to Adversarial Corruptions
• Computer Science
ArXiv
• 2021
This work proposes a new algorithm that not only achieves near-optimal regret in the stochastic setting, but also obtains a regret with an additive term of corruption in the corrupted setting, while maintaining efficient communication.
Multitask Bandit Learning through Heterogeneous Feedback Aggregation
• Computer Science
AISTATS
• 2021