3260 papers • 126 benchmarks • 313 datasets
This task has no description! Would you like to contribute one?
(Image credit: Papersgraph)
These leaderboards are used to track progress in game-of-poker-1
No benchmarks available.
Use these libraries to find game-of-poker-1 models and implementations
No datasets available.
No subtasks available.
An overview of the key components in RLCard is provided, a discussion of the design principles, a brief introduction of the interfaces, and comprehensive evaluations of the environments are provided.
This paper introduces the first scalable end-to-end approach to learning approximate Nash equilibria without prior domain knowledge, and combines fictitious self-play with deep reinforcement learning.
DeepStack is introduced, a new algorithm for imperfect information settings such as poker that combines recursive reasoning to handle information asymmetry, decomposition to focus computation on the relevant decision, and a form of intuition about arbitrary poker situations that is automatically learned from self-play games using deep learning
Through building DouZero, it is shown that classic Monte-Carlo methods can be made to deliver strong results in a hard domain with a complex action space, and the code and an online demo are released with the hope that this insight could motivate future work.
Adding a benchmark result helps the community track progress.