3260 papers • 126 benchmarks • 313 datasets
Starcraft II is a RTS game; the task is to train an agent to play the game. ( Image credit: The StarCraft Multi-Agent Challenge )
(Image credit: Papersgraph)
These leaderboards are used to track progress in starcraft-ii-18
Use these libraries to find starcraft-ii-18 models and implementations
No subtasks available.
QMIX employs a network that estimates joint action-values as a complex non-linear combination of per-agent values that condition only on local observations, and structurally enforce that the joint-action value is monotonic in the per- agent values, which allows tractable maximisation of the jointaction-value in off-policy learning.
The StarCraft Multi-Agent Challenge (SMAC), based on the popular real-time strategy game StarCraft II, is proposed as a benchmark problem and an open-source deep multi-agent RL learning framework including state-of-the-art algorithms is opened.
This paper introduces SC2LE (StarCraft II Learning Environment), a reinforcement learning environment based on the StarCraft II game that offers a new and challenging environment for exploring deep reinforcement learning algorithms and architectures and gives initial baseline results for neural networks trained from this data to predict game outcomes and player actions.
This work proposes Perceiver IO, a general-purpose architecture that handles data from arbitrary settings while scaling linearly with the size of inputs and outputs and augments the Perceiver with a flexible querying mechanism that enables outputs of various sizes and semantics, doing away with the need for task-specific architecture engineering.
A novel MARL approach, called duPLEX dueling multi-agent Q-learning (QPLEX), which takes a duplex dueling network architecture to factorize the joint value function and encodes the IGM principle into the neural network architecture and thus enables efficient value function learning.
FACMAC is a new method for cooperative multi-agent reinforcement learning in both discrete and continuous action spaces that uses a centralised but factored critic, which combines per-agent utilities into the joint action-value function via a non-linear monotonic function.
This is the first public work to investigate AI agents that can defeat the built-in AI in the StarCraft II full game, and the AI agent TStarBot1 is based on deep reinforcement learning over a flat action structure and theAI agent T starBot2 isbased on hard-coded rules over a hierarchical action structure.
Adding a benchmark result helps the community track progress.