Monte-carlo tree search (mcts) is a best-first search method guided by the results of monte-carlo simulations. It is based on randomized exploration of the search space. Using the results of previous explorations, the method gradually builds up a game tree in memory and successively becomes better at accurately estimating the values of the most promising moves. Mcts has substantially advanced the state of the art in board games such as go, amazons, hex, chinese checkers, kriegspiel, and lines of action.this chapter gives an overview of popular and effective enhancements for board game playing mcts agents. First, it starts by describing the structure of mcts and giving pseudocode. It also addresses how to adjust mcts to prove the game-theoretic value of a board position. Next, popular enhancements such as rave, progressive bias, progressive widening, and prior knowledge, which improve the simulation in the tree part of mcts, are discussed in detail. Subsequently, enhancements such as mast, n-grams, and evaluation function-based strategies are explained for improving the simulation outside the tree. As modern computers have nowadays multiple cores, this chapter mentions techniques to parallelize mcts in a straightforward but effective way. Finally, approaches to deal with imperfect information and stochasticity in an mcts context are discussed as well.
|Title of host publication||Handbook of Digital Games and Entertainment Technologies|
|Editors||Ryohei Nakatsu, Matthias Rauterberg, Paolo Ciancarini|
|Publication status||Published - 2017|