Abstract
Adaptive techniques tend to converge to a single optimum. For adaptive game AI, such convergence is often undesirable, as repetitive game AI is considered to be uninteresting for players. In this paper, we propose a method for automatically learning diverse but effective macros that can be used as components of adaptive game AI scripts. Macros are learned by a cross-entropy method (CEM). This is a selection-based optimization method that, in our experiments, maximizes an interestingness measure. We demonstrate the approach in a computer role-playing game (CRPG) simulation with two duelling wizards, one of which is controlled by an adaptive game AI technique called "dynamic scripting." Our results show that the macros that we learned manage to increase both adaptivity and diversity of the scripts generated by dynamic scripting, while retaining playing strength.
Original language | English |
---|---|
Pages (from-to) | 16-27 |
Number of pages | 12 |
Journal | IEEE Transactions on Computational Intelligence and AI in Games |
Volume | 1 |
Issue number | 1 |
DOIs | |
Publication status | Published - Mar 2009 |
Keywords
- AI
- cross-entropy method (CEM)
- dynamic scripting
- game
- reinforcement learning