An actor–critic algorithm for multi-agent learning in queue-based stochastic games
We consider state-dependent pricing in a two-player service market stochastic game where state of the game and its transition dynamics are modeled using a semi-Markovian queue. We propose a multi-time scale actor-critic based reinforcement algorithm for multi-agent learning under self-play and provide experimental results on Nash convergence.
An actor–critic algorithm for multi-agent learning in queue-based stochastic games
We consider state-dependent pricing in a two-player service market stochastic game where state of the game and its transition dynamics are modeled using a semi-Markovian queue. We propose a multi-time scale actor-critic based reinforcement algorithm for multi-agent learning under self-play and provide experimental results on Nash convergence.
