Skip to main content

Artificial Intelligence for Games

  • Chapter
  • First Online:
A Guided Tour of Artificial Intelligence Research

Abstract

This chapter presents the classical alpha-beta algorithm and several variants, Monte Carlo Tree Search which is at the origin of recent progresses in many games, techniques used in video games and puzzles, and retrograde analysis which performs perfect play in endgames.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 149.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Hardcover Book
USD 199.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  • Allis L (1994) Searching for solutions in games and artificial intelligence. PhD thesis, Vrije Universitat Amsterdam

    Google Scholar 

  • Allis L, van der Meulen M, van den Herik H (1994) Proof-number search. Artif Intell 66:91–124

    Google Scholar 

  • Anantharaman T, Campbell M, Hsu F (1989) Singular extensions: adding selectivity to brute force searching. Artif Intell 43(1):99–109

    Google Scholar 

  • Andrade G, Ramalho G, Gomes AS, Corruble V (2006) Dynamic game balancing: an evaluation of user satisfaction. In: AAAI conference on artificial intelligence and interactive digital entertainment, pp 3–8

    Google Scholar 

  • Arneson B, Hayward RB, Henderson P (2010) Monte Carlo tree search in hex. IEEE Trans Comput Intell AI Games 2(4):251–258

    Google Scholar 

  • Beal D (1990) A generalised quiescence search algorithm. Artif Intell 43:85–98

    Google Scholar 

  • Berliner H (1979) The B* tree search algorithm: a best-first proof procedure. Artif Intell 12:23–40

    Google Scholar 

  • Bouzy B (2005) Associating domain-dependent knowledge and Monte Carlo approaches within a Go program. Inf Sci 175(4):247–257

    Google Scholar 

  • Bouzy B, Cazenave T (2001) Computer Go: an AI oriented survey. Artif Intell 132(1):39–103

    Google Scholar 

  • Bouzy B, Helmstetter B (2003) Monte-Carlo Go developments. In: ACG, vol 263 of IFIP. Kluwer, pp 159–174

    Google Scholar 

  • Bruegmann B (1993) Monte Carlo Go. Unpublished

    Google Scholar 

  • Bulitko V, Lustrek M, Schaeffer J, Bjornsson Y, Sigmundarson S (2008) Dynamic control in real-time heuristic search. J Artif Intell Res 32(1):419–452

    Google Scholar 

  • Campbell M, Marsland T (1983) A comparison of minimax tree search algorithms. Artif Intell 20:347–367

    Google Scholar 

  • Campbell M, Hoane A, Hsu F-H (2002) Deep blue. Artif Intell 134:57–83

    Google Scholar 

  • Cazenave T (2003) Metarules to improve tactical Go knowledge. Inf Sci 154(3–4):173–188

    Google Scholar 

  • Cazenave T (2006a) Optimizations of data structures, heuristics and algorithms for path-finding on maps. In: CIG, pp 27–33

    Google Scholar 

  • Cazenave T (2006b) A phantom-Go program. Advances in computer games 2005. Lecture notes in computer science, vol. 4250. Springer, Berlin, pp 120–125

    Google Scholar 

  • Cazenave T (2009) Nested Monte Carlo search. In: IJCAI 2009, Pasadena, USA, pp 456–461

    Google Scholar 

  • Cazenave T, Helmstetter B (2005) Combining tactical search and Monte-Carlo in the game of Go. IEEE CIG 2005:171–175

    Google Scholar 

  • Cazenave T, Jouandeau N (2007) On the parallelization of UCT. In: Proceedings of CGW07, pp 93–101

    Google Scholar 

  • Cazenave T, Teytaud F (2012) Application of the nested rollout policy adaptation algorithm to the traveling salesman problem with time windows. In: Learning and intelligent optimization, pp 42–54

    Google Scholar 

  • Cazenave T, Saffidine A, Schofield MJ, Thielscher M (2016) Nested Monte Carlo search for two-player games. In: Proceedings of the thirtieth AAAI conference on artificial intelligence, Phoenix, Arizona, USA, 12–17 Feb 2016, pp 687–693

    Google Scholar 

  • Chang HS, Fu MC, Hu J, Marcus SI (2005) An adaptive sampling algorithm for solving Markov decision processes. Oper Res 53(1):126–139

    Google Scholar 

  • Chaslot G, Saito J, Bouzy B, Uiterwijk JWHM, van den Herik HJ (2006) Monte-Carlo strategies for computer Go. In: Schobbens P-Y, Vanhoof W, Schwanen G (eds) Proceedings of the 18th BeNeLux conference on artificial intelligence, Namur, Belgium, pp 83–91

    Google Scholar 

  • Chaslot G, Winands M, Uiterwijk J, van den Herik H, Bouzy B (2008) Progressive strategies for Monte-Carlo tree search. New Math Nat Comput 4(3):343–357

    Google Scholar 

  • Chou C-W, Teytaud O, Yen S-J (2011) Revisiting Monte Carlo tree search on a normal form game: Nogo. In: European conference on the applications of evolutionary computation. Springer, Berlin, pp 73–82

    Google Scholar 

  • Corruble V, Ramalho G (2009) Jeux vidéo et Systèmes Multi-Agents. IC2 Series. Hermès Lavoisier, pp 235–264. ISBN: 978-2-7462-1785-0

    Google Scholar 

  • Couetoux A (2013) Monte Carlo tree search for continuous and stochastic sequential decision making problems. PhD thesis, Université Paris Sud-Paris XI

    Google Scholar 

  • Coulom R (2006) Efficient selectivity and backup operators in Monte-Carlo tree search. In: Ciancarini P, van den Herik HJ (eds) Proceedings of the 5th international conference on computers and games, Turin, Italy

    Google Scholar 

  • Coulom R (2007) Computing Elo ratings of move patterns in the game of Go. ICGA J 4(30):198–208

    Google Scholar 

  • Csikszentmihalyi M (1975) Beyond boredom and anxiety. Jossey-Bass San Francisco

    Google Scholar 

  • Csikszentmihalyi M (1990) Flow: the psychology of optimal experience. Harper and Row, New York

    Google Scholar 

  • Culberson JC, Schaeffer J (1998) Pattern databases. Comput Intell 4(14):318–334

    Google Scholar 

  • De Mesmay F, Rimmel A, Voronenko Y, Püschel M (2009) Bandit-based optimization on graphs with application to library performance tuning. In: ICML, Montréal Canada

    Google Scholar 

  • Degris T, Sigaud O, Wuillemin P (2009) Apprentissage par renforcement factorisé pour le comportement de personnages non joueurs. Revue d’Intelligence Artificielle 23(2):221–251

    Google Scholar 

  • Donninger C (1993) Null move and deep search: selective search heuristics for obtuse chess programs. ICCA J 16(3):137–143

    Google Scholar 

  • Felner A, Korf RE, Hanan S (2004) Additive pattern database heuristics. J Artif Intell Res (JAIR) 22:279–318

    Google Scholar 

  • Felner A, Korf RE, Meshulam R, Holte RC (2007) Compressed pattern databases. J Artif Intell Res (JAIR) 30:213–247

    Google Scholar 

  • Finley L (2016) Nested Monte Carlo tree search as applied to Samurai Sudoku

    Google Scholar 

  • Finnsson H, Björnsson Y (2008a) Simulation-based approach to general game playing. AAAI 8:259–264

    Google Scholar 

  • Finnsson H, Björnsson Y (2008b) Simulation-based approach to general game playing. In: AAAI, pp 259–264

    Google Scholar 

  • Flórez-Puga G, Gómez-Martın M, Dıaz-Agudo B, González-Calero P (2008) Dynamic expansion of behaviour trees. In: AAAI conference on artificial intelligence and interactive digital entertainment, pp 36–41

    Google Scholar 

  • Gelly S, Wang Y, Munos R, Teytaud O (2006) Modification of UCT with patterns in Monte Carlo Go. Rapport de recherche INRIA RR-6062

    Google Scholar 

  • Gelly S, Hoock JB, Rimmel A, Teytaud O, Kalemkarian Y (2008) The parallelization of Monte-Carlo planning. In: Proceedings of the international conference on informatics in control, automation and robotics (ICINCO 2008), pp 198–203. To appear

    Google Scholar 

  • Greenblatt R, Eastlake D, Crocker S (1967) The Greenblatt chess program. In: Fall joint computing conference, vol 31. ACM, New York, pp 801–810

    Google Scholar 

  • Harel D (1987) Statecharts: a visual formalism for complex systems

    Google Scholar 

  • Hart P, Nilsson N, Raphael B (1968) A formal basis for the heuristic determination of minimum cost paths. IEEE Trans Syst Sci Cybernet 4(2):100–107

    Google Scholar 

  • Hoang H, Lee-Urban S, Muñoz-Avila H (2005) Hierarchical plan representations for encoding strategic game AI. In: Proceedings of artificial intelligence and interactive digital entertainment conference (AIIDE-05)

    Google Scholar 

  • Ierusalimschy R, de Figueiredo LH, Celes W (2007) The evolution of LUA. In: HOPL III: proceedings of the third ACM SIGPLAN conference on history of programming languages, ACM, New York, NY, USA, pp 2-1-2-26

    Google Scholar 

  • Junghanns A (1998) Are there practical alternatives to alpha-beta? ICCA J 21(1):14–32

    Google Scholar 

  • Junghanns A, Schaeffer J (2001) Sokoban: enhancing general single-agent search methods using domain knowledge. Artif Intell 129(1–2):219–251

    Google Scholar 

  • Kendall G, Parkes A, Spoerer K (2008) A survey of NP-complete puzzles. ICGA J 31(1):13–34

    Google Scholar 

  • Knuth D, Moore R (1975) An analysis of alpha-beta pruning. Artif Intell 6:293–326

    Google Scholar 

  • Kocsis L, Szepesvari C (2006) Bandit-based Monte-Carlo planning. In: ECML’06, pp 282–293

    Google Scholar 

  • Korf R (1985a) Depth-first iterative deepening: an optimal admissible tree search. Artif Intell 27:97–109

    Google Scholar 

  • Korf RE (1985b) Depth-first iterative-deepening: an optimal admissible tree search. Artif Intell 27(1):97–109

    Google Scholar 

  • Korf RE (1997) Finding optimal solutions to Rubik’s cube using pattern databases. In: AAAI-97, pp 700–705

    Google Scholar 

  • Korf R, Chickering D (1994) Best-first search. Artif Intell 84:299–337

    Google Scholar 

  • Kozelek T (2009) Methods of MCTS and the game arimaa

    Google Scholar 

  • Laird JE (2002) Research in human-level AI using computer games. Commun ACM 45(1):32–35

    Google Scholar 

  • Lee C-S, Wang M-H, Chaslot G, Hoock J-B, Rimmel A, Teytaud O, Tsai S-R, Hsu S-C, Hong T-P (2009) The computational intelligence of MOGO revealed in Taiwan’s computer Go tournaments. In: IEEE transactions on computational intelligence and AI in games

    Google Scholar 

  • Madeira C, Corruble V (2009) Strada: une approche adaptative pour les jeux de stratégie modernes. Revue d’Intelligence Artificielle 23(2):293–326

    Google Scholar 

  • Maîtrepierre R, Mary J, Munos R (2008) Adaptative play in texas hold’em poker. In: European conference on artificial intelligence-ECAI

    Google Scholar 

  • Marsland T (1986) A review of game-tree pruning. ICCA J 9(1):3–19

    Google Scholar 

  • McAllester D (1988) Conspiracy numbers for min-max search. Artif Intell 35:287–310

    Google Scholar 

  • Müller M (2002) Computer Go. Artif Intell 134(1–2):145–179

    Google Scholar 

  • Natkin S (2004) Jeux vidéo et médias du XXIe siècle: quels modèles pour les nouveaux loisirs numériques? Vuibert

    Google Scholar 

  • Nijssen P, Winands MH (2012) Monte Carlo tree search for the hide-and-seek game Scotland yard. IEEE Trans Comput Intell AI Games 4(4):282–294

    Google Scholar 

  • Nijssen J, Winands MH (2013) Search policies in multi-player games 1. Icga J 36(1):3–21

    Google Scholar 

  • Ochs M, Sabouret N, Corruble V (2009) Simulation de la dynamique des emotions et des relations sociales de personnages virtuels. Revue d’Intelligence Artificielle 23(2):327–358

    Google Scholar 

  • Paxton C, Raman V, Hager GD, Kobilarov M (2017) Combining neural networks and tree search for task and motion planning in challenging environments. arXiv:1703.07887

  • Pearl J (1980a) Asymptotic properties of minimax trees and game-searching procedures. Artif Intell 14:113–138

    Google Scholar 

  • Pearl J (1980b) SCOUT: a simple game-searching algorithm with proven optimal properties. In: Proceedings of the first annual national conference on artificial intelligence, pp 143–145

    Google Scholar 

  • Perlin K (2005) Toward interactive narrative. In: International conference on virtual storytelling. Springer, Berlin, pp 135–147

    Google Scholar 

  • Pitrat J (1968) Realization of a general game-playing program. IFIP Congr 2:1570–1574

    Google Scholar 

  • Plaat A, Schaeffer J, Pils W, de Bruin A (1996) Best-first fixed depth minimax algorithms. Artif Intell 87:255–293

    Google Scholar 

  • Rabin S (2002) AI game programming wisdom. Charles River Media, USA

    Google Scholar 

  • Rickel J, Marsella S, Gratch J, Hill R, Traum D, Swartout W (2002) Toward a new generation of virtual humans for interactive experiences. In: IEEE intelligent systems, pp 32–38

    Google Scholar 

  • Rivest R (1988) Game-tree searching by min-max approximation. Artif Intell 34(1):77–96

    Google Scholar 

  • Rolet P, Sebag M, Teytaud O (2009a) Optimal active learning through billiards and upper confidence trees in continuous domains. In: Proceedings of the ECML conference

    Google Scholar 

  • Rolet P, Sebag M, Teytaud O (2009b) Optimal robust expensive optimization is tractable. In: Gecco 2009, ACM, Montréal Canada, p 8

    Google Scholar 

  • Romein JW, Bal HE (2003) Solving awari with parallel retrograde analysis. IEEE Comput 36(10):26–33

    Google Scholar 

  • Rosin CD (2011) Nested rollout policy adaptation for Monte Carlo tree search. In: Ijcai, pp 649–654

    Google Scholar 

  • Schadd MP, Winands MH, Van Den Herik HJ, Chaslot GM-B, Uiterwijk JW (2008) Single-player Monte Carlo tree search. In: International conference on computers and games. Springer, Berlin, pp 1–12

    Google Scholar 

  • Schaeffer J (1989) The history heuristic and alpha-beta search enhancements in practice. IEEE Trans Pattern Anal Mach Intell 11(11):1203–1212

    Google Scholar 

  • Schaeffer J (1990) Conspiracy numbers. Artif Intell 43:67–84

    Google Scholar 

  • Schaeffer J, van den Herik J (2002) Games, computers, and artificial intelligence. Artif Intell 134:1–7

    Google Scholar 

  • Schaeffer J, Burch N, Bjornsson Y, Kishimoto A, Muller M, Lake R, Lu P, Sutphen S (2007) Checkers is solved. In: Science

    Google Scholar 

  • Shannon C (1950) Programming a computer to play Chess. Philos Mag 41:256–275

    Google Scholar 

  • Sheppard B (2002) World-championship-caliber scrabble scrabble® is a registered trademark. All intellectual property rights in and to the game are owned in the USA by Hasbro Inc. in Canada by Hasbro Canada Corporation, and throughout the rest of the world by JW Spear & Sons Limited of Maidenhead, Berkshire, England, a subsidiary of Mattel Inc. Artif Intell 134(1-2):241–275

    Google Scholar 

  • Silver D, Huang A, Maddison CJ, Guez A, Sifre L, Van Den Driessche G, Schrittwieser J, Antonoglou I, Panneershelvam V, Lanctot M et al (2016) Mastering the game of Go with deep neural networks and tree search. Nature 529(7587):484–489

    Google Scholar 

  • Silver D, Hubert T, Schrittwieser J, Antonoglou I, Lai M, Guez A, Lanctot M, Sifre L, Kumaran D, Graepel T et al (2017a) Mastering chess and shogi by self-play with a general reinforcement learning algorithm. arXiv:1712.01815

  • Silver D, Schrittwieser J, Simonyan K, Antonoglou I, Huang A, Guez A, Hubert T, Baker L, Lai M, Bolton A et al (2017b) Mastering the game of go without human knowledge. Nature 550(7676):354

    Google Scholar 

  • Slate D, Atkin L (1977) Chess 4.5 - the northwestern university chess program. In: Frey P (ed) Chess skill in man and machine. Springer, Berlin, pp 82–118

    Google Scholar 

  • Spronck P, Ponsen M, Sprinkhuizen-Kuyper I, Postma E (2006) Adaptive game AI with dynamic scripting. Mach Learn 63(3):217–248

    Google Scholar 

  • Stockman G (1979) A minimax algorithm better than alpha-beta? Artif Intell 12:179–196

    Google Scholar 

  • Sturtevant NR, Felner A, Barrer M, Schaeffer J, Burch N (2009) Memory-based heuristics for explicit state spaces. In: IJCAI, pp 609–614

    Google Scholar 

  • Teytaud F, Teytaud O (2009) Creating an upper-confidence-tree program for havannah. In: Advances in computer games. Springer, Berlin, pp 65–74

    Google Scholar 

  • Thompson K (1996) 6-Piece endgames. ICCA J 19(4):215–226

    Google Scholar 

  • Tian Y, Jerry Ma*, Qucheng Gong*, Sengupta S, Chen Z, Zitnick CL (2018) Elf opengo. https://github.com/pytorch/ELF

  • von Neumann J, Morgenstern O (1944) Theory of games and economic behavior. Princeton University Press, Princeton

    Google Scholar 

  • Williams RJ (1992) Simple statistical gradient-following algorithms for connectionist reinforcement learning. Mach Learn 8(3–4):229–256

    Google Scholar 

  • Zobrist A (1990) A new hashing method with application for game playing. ICCA J 13(2):69–73

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Bruno Bouzy .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this chapter

Check for updates. Verify currency and authenticity via CrossMark

Cite this chapter

Bouzy, B., Cazenave, T., Corruble, V., Teytaud, O. (2020). Artificial Intelligence for Games. In: Marquis, P., Papini, O., Prade, H. (eds) A Guided Tour of Artificial Intelligence Research. Springer, Cham. https://doi.org/10.1007/978-3-030-06167-8_11

Download citation

Publish with us

Policies and ethics