Skip to main content

Evolution of Reactive Rules in Multi Player Computer Games Based on Imitation

  • Conference paper
Advances in Natural Computation (ICNC 2005)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 3611))

Included in the following conference series:

Abstract

Observing purely reactive situations in modern computer games, one can see that in many cases few, simple rules are sufficient to perform well in the game. In spite of this, the programming of an artificial opponent is still a hard and time consuming task in the way it is done for the most games today. In this paper we propose a system in which no direct programming of the behaviour of the opponents is necessary. Instead, rules are gained by observing human players and then evaluated and optimised by an evolutionary algorithm to optimise the behaviour. We will show that only little learning effort is required to be competitive in reactive situations. In the course of our experiments our system proved to generate better artificial players than the original ones supplied with the game.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 119.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Bakkes, S., Spronck, P., Postma, E.: TEAM: The Team-Oriented Evolutionary Adaptability Mechanism. In: Rauterberg, M. (ed.) ICEC 2004. LNCS, vol. 3166, pp. 273–282. Springer, Heidelberg (2004)

    Chapter  Google Scholar 

  2. Beyer, H.-G., Schwefel, H.-P.: Evolution strategies – A comprehensive introduction. Natural Computing 1, 3–52 (2002)

    Article  MATH  MathSciNet  Google Scholar 

  3. Gonzalez, R.C., Wintz, P.A.: Digital Image Processing. Addison-Wesley, Reading (1992)

    Google Scholar 

  4. Hawes, N.: An Anytime Planning Agent For Computer Game Worlds. In: Schaeffer, J., Müller, M., Björnsson, Y. (eds.) CG 2002. LNCS, vol. 2883, pp. 1–14. Springer, Heidelberg (2003)

    Chapter  Google Scholar 

  5. Kaminka, G., Go, J., Vu, T.: Context-dependent joint-decision arbitration for computer games (2002)

    Google Scholar 

  6. Laird, J.: It Knows What You’re Going to Do: Adding Anticipation to a Quakebot. In: AAAI 2000 Spring Symposium Series: Artificial Intelligence and Interactive Entertainment: AAAI Technical Report SS-00-02 (2000)

    Google Scholar 

  7. Nareyek, A.: A Planning Model for Agents in Dynamic and Uncertain Real-Time Environments. In: Proceedings of the Workshop on Integrating Planning, Scheduling and Execution in Dynamic and Uncertain Environments at the Fourth International Conference on Artificial Intelligence Planning Systems, pp. 7–14. AAAI Press, Menlo Park (1998)

    Google Scholar 

  8. Nareyek, A.: Constraint-Based Agents - An Architecture for Constraint-Based Modeling and Local-Search-Based Reasoning for Planning and Scheduling in Open and Dynamic Worlds. Künstliche Intelligenz 2, 51–53 (2002)

    Google Scholar 

  9. Nason, S., Laird, J.: Soar-RL: Integrating Reinforcement Learning with Soar. In: International Conference on Cognitive Modelling (2004)

    Google Scholar 

  10. Norling, E.: Capturing the Quake Player: Using a BDI Agent to Model Human Behaviour. In: Proceedings of the Second International Joint Conference on Autonomous Agents and Multiagent Systems, pp. 1080–1081 (2003)

    Google Scholar 

  11. Priesterjahn, S., Goebels, A., Weimer, A.: Stigmergetic Communication for Cooperative Agent Routing in Virtual Environments. In: Proceedings of the International Conference on Artificial Intelligence and the Simulation of Behaviour (April 2005)

    Google Scholar 

  12. Schwefel, H.-P.: Evolution and Optimum Seeking. Sixth-Generation Computer Technology. Wiley Interscience, New York (1995)

    Google Scholar 

  13. Thurau, C., Bauckhage, C., Sagerer, G.: Imitation learning at all levels of game-AI. In: Proceedings of the International Conference on Computer Games, Artificial Intelligence, Design and Education, pp. 402–408 (2004)

    Google Scholar 

  14. Thurau, C., Bauckhage, C., Sagerer, G.: Learning Human-Like Movement Behavior for Computer Games. In: Proceedings of the 8th International Conference on the Simulation of Adaptive Behavior (SAB 2004) (2004)

    Google Scholar 

  15. Thurau, C., Bauckhauge, C., Sagerer, G.: Combining Self Organizing Maps and Multilayer Perceptrons to Learn Bot-Behavior for a Commercial Game. In: Proceedings of the GAME-ON 2003 Conference, pp. 119–123 (2003)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2005 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Priesterjahn, S., Kramer, O., Weimer, A., Goebels, A. (2005). Evolution of Reactive Rules in Multi Player Computer Games Based on Imitation. In: Wang, L., Chen, K., Ong, Y.S. (eds) Advances in Natural Computation. ICNC 2005. Lecture Notes in Computer Science, vol 3611. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11539117_105

Download citation

  • DOI: https://doi.org/10.1007/11539117_105

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-28325-6

  • Online ISBN: 978-3-540-31858-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics