skip to main content
10.1145/1160633.1160759acmconferencesArticle/Chapter ViewAbstractPublication PagesaamasConference Proceedingsconference-collections
Article

Efficient agents for cliff-edge environments with a large set of decision options

Published:08 May 2006Publication History

ABSTRACT

This paper proposes an efficient agent for competing in Cliff Edge (CE) environments, such as sealed-bid auctions, dynamic pricing and the ultimatum game. The agent competes in one-shot CE interactions repeatedly, each time against a different human opponent, and its performance is evaluated based on all the interactions in which it participates. The agent, which learns the general pattern of the population's behavior, does not apply any examples of previous interactions in the environment, neither of other competitors nor its own. We propose a generic approach which competes in different CE environments under the same configuration, with no knowledge about the specific rules of each environment. The underlying mechanism of the proposed agent is a new meta-algorithm, Deviated Virtual Learning (DVL), which extends existing methods to efficiently cope with environments comprising a large number of optional decisions at each decision point. Experiments comparing the performance of the proposed algorithm with algorithms taken from the literature, as well as another intuitive meta-algorithm, reveal a significant superiority of the former in average payoff and stability. In addition, the agent performed better than human competitors executing the same task.

References

  1. P. Bourgine and B. Leloup. May learning explain the ultimatum game paradox? Technical Report GRID Working Paper No. 00--03, Ecole Polytechnique, 2000.Google ScholarGoogle Scholar
  2. T. Brenner and N. Vriend. On the behavior of proposers in ultimatum games. J. Econ. Behav. Organ. Forthcoming.Google ScholarGoogle Scholar
  3. A. Byde, C. Preist, and N. Jennings. Decision procedures for multiple auctions. In AAMAS, pages 613--620, 2002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. J. DiMicco, A. Greenwald, and P. Maes. Dynamic pricing strategies under a finite time horizon. In EC'01, 2001. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. S. Fatima, M. Wooldridge, and N. Jennings. Sequential auctions for objects with common and private values. In AAMAS, pages 635--642, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Y. Gal, A. Pfeffer, F. Marzo, and B. Grosz. Learning social preferences in games. In AAAI, pages 226--231, 2004. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. J. Gittins. Multiarmed Bandits Allocation Indices. Wiley, New York, 1989.Google ScholarGoogle Scholar
  8. B. Grosskopf. Reinforcement and directional learning in the ultimatum game with responder competition. Experimental Economics, 6(2):141--158, 2003.Google ScholarGoogle ScholarCross RefCross Ref
  9. W. Guth and S. Huck. From ultimatum bargaining to dictatorship - an experimental study of four games varying in veto power. Metroeconomica, 48(3):262--279, 1997.Google ScholarGoogle ScholarCross RefCross Ref
  10. V. Krishna and J. Morgan. An analysis of the war of attrition and the all-pay auction. J. Econ. Theory, 72:343--362, 1997.Google ScholarGoogle ScholarCross RefCross Ref
  11. B. Leloup and L. Deveaux. Dynamic pricing on the internet: Theory and simulations. JECR, 1(3):265--276, 2001. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. L. Niklasson, H. Engstrom, and U. Johansson. An adaptive rock, scissors and paper player based on a tapped delay neural network. In ADCOG21, pages 130--136, 2001.Google ScholarGoogle Scholar
  13. A. Roth and I. Erev. Learning in extensive form games: Experimental data and simple dynamic models in the intermediate term. Games Econ. Behav., 8:164--212, 1995.Google ScholarGoogle ScholarCross RefCross Ref
  14. R. Sutton and A. Barto. An Introduction to Reinforcement Learning. MIT Press, 1998. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. P. Todd and B. Borges. Designing socially intelligent agents for the ultimatum game. In K. Dautenhahn, editor, Socially intelligent agents-Papers from the 1997 Fall Symposium, pages 134--136. AAAI Press, Menlo Park, CA, 1997.Google ScholarGoogle Scholar
  16. N. Vreind. Will reasoning improve learning? Econ. Lett., 55(1):9--18. 1997.Google ScholarGoogle ScholarCross RefCross Ref
  17. F. Zhong, D. Wu, and S. Kimbrough. Cooperative agent systems: Artificial agents play the ultimatum game. Journal of Group Decision and Negotiation, 11(6):433--447, 2002.Google ScholarGoogle ScholarCross RefCross Ref

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image ACM Conferences
    AAMAS '06: Proceedings of the fifth international joint conference on Autonomous agents and multiagent systems
    May 2006
    1631 pages
    ISBN:1595933034
    DOI:10.1145/1160633

    Copyright © 2006 ACM

    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    • Published: 8 May 2006

    Permissions

    Request permissions about this article.

    Request Permissions

    Check for updates

    Qualifiers

    • Article

    Acceptance Rates

    Overall Acceptance Rate1,155of5,036submissions,23%

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader