Skip to main content

A Note on Working Memory in Agent Learning

  • Chapter
  • 511 Accesses

Part of the International Handbooks on Information Systems book series (INFOSYS)

Abstract

An important dimension of system and mechanism design, working memory, has been paid insufficient attention by scholars. Existing literature reports mixed findings on the effects of the amount of working memory on system efficiency. In this note, we investigate this relationship with a computational approach. We design an intelligent agent system in which three agents, one buyer and two bidders, play an Exchange Game repeatedly. The buyer agent decides whether to list a request for proposal, while the bidders bid for it independently. Only one bidder can win on a given round of play. Once the winning bidder is chosen by the buyer, a transaction takes place. The two parties of the trade can either cooperate or defect at this point. The decisions are made simultaneously and the payoffs essentially follow the Prisoner’s Dilemma game. We find that the relationship between working memory and the efficiency of the system has an inverted U-shape, i.e., there seems to be an optimal memory size. When we mixed agents with different memory sizes together, agents with the same amount of working memory generate the most efficient outcome in terms of total payoffs.

Keywords

These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (Canada)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   169.00
Price excludes VAT (Canada)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   219.99
Price excludes VAT (Canada)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD   219.99
Price excludes VAT (Canada)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Robert Axelrod and W.D. Hamilton, The evolution of cooperation, Science 211 (1981), 1390.

    CrossRef  ADS  PubMed  CAS  ISI  MathSciNet  Google Scholar 

  2. Adam M. Brandenburger and Harborne (Gus) W. Stuart, Biform games, Working paper No. 00-06, Harvard NOM Research Paper, http://ssrn.com/abstract=264199, November 2000.

    Google Scholar 

  3. Caroline Claus and Craig Boutilier, The dynamics of reinforcement learning in cooperative multiagent systems, Proceedings of the Fifteenth National Conference on Artificial Intelligence (Menlo Park, CA), AAAI Press/MIT Press, 1998, pp. 746–752.

    Google Scholar 

  4. Chrysanthos N. Dellarocas, Efficiency and robustness of binary feedback mechanisms in trading environments with moral hazard, Working Paper No. 4297-03, MIT, Sloan School, Cambridge, MA, January 2003, http://ssrn.com/abstract=393043.

    Google Scholar 

  5. J. Hu and M. P. Wellman, Multiagent reinforcement learning: Theoretical framework and an algorithm, Fifteenth International Conference on Machine Learning, July 1998, pp. 242–250.

    Google Scholar 

  6. Steven O. Kimbrough and Ronald M. Lee, Formal aspects of electronic commerce: Research issues and challenges, International Journal of Electronic Commerce 1 (1997), no. 4, 11–30.

    Google Scholar 

  7. Steven O. Kimbrough, Ming Lu, and Ann Kuo, A note on strategic learning in policy space, Formal Modelling in Electronic Commerce: Representation, Inference, and Strategic Interaction (Steven O. Kimbrough and D. J. Wu, eds.), Springer, 2004.

    Google Scholar 

  8. Leslie Pack Kaelbling, Michael L. Littman, and Andrew W. Moore, Reinforcement learning: A survey, Journal of Artificial Intelligence Research 4 (1996), 237–285.

    ISI  Google Scholar 

  9. Steven O. Kimbrough, D. J. Wu, and Fang Zhong, Computers play the beer game: Can artificial agents manage supply chains?, Decision Support Systems 33 (2002), no. 3, 323–333.

    CrossRef  ISI  Google Scholar 

  10. Richar S. Sutton and Andrew G. Barto, Reinforcement learning: An introduction, The MIT Press, Cambridge, MA, 1998.

    Google Scholar 

  11. T. Sandholm and R. Crites, Multiagent reinforcement learning in iterated prisoner’s dilemma, Biosystems 37 (1995), 147–166, Special Issue on the Prisoner’s Dilemma.

    CrossRef  ISI  Google Scholar 

  12. C. Shapiro, The theory of business strategy, Rand Journal of Economics 20 (1989), 125–137.

    CrossRef  PubMed  CAS  ISI  Google Scholar 

  13. J. Sutton, Sunk costs and market structure: Price competition, advertising, and the evolution of concentration, The MIT Press, Cambridge, MA, 1991.

    Google Scholar 

  14. C.J.C.H. Watkins and P. Dayan, Q-learning, Machine Learning 8 (1992), 279–292.

    ISI  Google Scholar 

  15. Fang Zhong, Steven O. Kimbrough, and D. J. Wu, Cooperative agent systems: Artificial agents play the ultimatum game, Journal of Group Decision and Negotiation 11 (November 2002), no. 6, 433–447.

    CrossRef  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2005 Springer-Verlag Berlin Heidelberg

About this chapter

Cite this chapter

Zhong, F. (2005). A Note on Working Memory in Agent Learning. In: Kimbrough, S.O., Wu, D. (eds) Formal Modelling in Electronic Commerce. International Handbooks on Information Systems. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-26989-4_20

Download citation

Publish with us

Policies and ethics