Scalable Evaluation of Multi-Agent Reinforcement Learning with Melting Pot
In International Conference on Machine Learning 2021 (pp. 6187-6199). PMLR Existing evaluation suites for multi-agent reinforcement learning (MARL) do not assess generalization to novel situations as their primary objective (unlike supervised-learning benchmarks). Our contribution, Melting Pot, is a...
Gespeichert in:
Hauptverfasser: | , , , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | In International Conference on Machine Learning 2021 (pp.
6187-6199). PMLR Existing evaluation suites for multi-agent reinforcement learning (MARL) do
not assess generalization to novel situations as their primary objective
(unlike supervised-learning benchmarks). Our contribution, Melting Pot, is a
MARL evaluation suite that fills this gap, and uses reinforcement learning to
reduce the human labor required to create novel test scenarios. This works
because one agent's behavior constitutes (part of) another agent's environment.
To demonstrate scalability, we have created over 80 unique test scenarios
covering a broad range of research topics such as social dilemmas, reciprocity,
resource sharing, and task partitioning. We apply these test scenarios to
standard MARL training algorithms, and demonstrate how Melting Pot reveals
weaknesses not apparent from training performance alone. |
---|---|
DOI: | 10.48550/arxiv.2107.06857 |