Sample-Efficient Robust Multi-Agent Reinforcement Learning in the Face of Environmental Uncertainty

To overcome the sim-to-real gap in reinforcement learning (RL), learned policies must maintain robustness against environmental uncertainties. While robust RL has been widely studied in single-agent regimes, in multi-agent environments, the problem remains understudied—despite the fact that the problems posed by environmental uncertainties are often exacerbated by strategic interactions. This work focuses on learning in distributionally robust Markov games (RMGs), a robust variant of standard Markov games, wherein each agent aims to learn a policy that maximizes its own worst-case performance when the deployed environment deviates within its own prescribed uncertainty set. This results in a set of robust equilibrium strategies for all agents that align with classic notions of game-theoretic equilibria.

Data and Resources

Cite this as

Laixi Shi, Eric Mazumdar, Yuejie Chi, Adam Wierman (2024). Dataset: Sample-Efficient Robust Multi-Agent Reinforcement Learning in the Face of Environmental Uncertainty. https://doi.org/10.57702/f3z2d670

DOI retrieved: December 2, 2024

Additional Info

Field Value
Created December 2, 2024
Last update December 2, 2024
Defined In https://doi.org/10.48550/arXiv.2404.18909
Author Laixi Shi
More Authors
Eric Mazumdar
Yuejie Chi
Adam Wierman
Homepage https://arxiv.org/abs/2106.09453