Computing Nash Equilibria in Adversarial Stochastic Team Games

Computing Nash equilibrium policies in Multi-agent Reinforcement Learning (MARL) is a fundamental question that has received a lot of attention both in theory and in practice. However, beyond the single-agent setting, provable guarantees for computing stationary Nash equilibrium policies only exist...

Full description

Saved in:
Bibliographic Details
Main Author Stavroulakis, Stelios Andrew
Format Dissertation
LanguageEnglish
Published ProQuest Dissertations & Theses 01.01.2022
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Computing Nash equilibrium policies in Multi-agent Reinforcement Learning (MARL) is a fundamental question that has received a lot of attention both in theory and in practice. However, beyond the single-agent setting, provable guarantees for computing stationary Nash equilibrium policies only exist for limited settings, e.g., two-player zero-sum stochastic games and Markov Potential Games. This thesis investigates what happens when a team of players faces an adversary in the absence of coordination. It is an initial step for understanding non-cooperative two-team zero-sum stochastic games, even in this case, where one of the teams is composed of only one agent. In this thesis we consider the aforementioned setting. The contributions are twofold. We prove the existence of Nash equilibria in these settings and also design a decentralized algorithm for computing Nash equilibria in such games. One of the main technical challenges in the proof of correctness of the algorithm is the analysis of a non-linear program with non-convex constraints that arises when applying the Bellman equations; the analysis makes use of the so-called constraint qualifications and Lagrange duality.
ISBN:9798381107265