- 1.Sachiyo Arai, Kazuteru Miyazaki, and Shigenobu Kobayashi. Generating cooperative behavior by multi-agent reinforcement learning. In Sixth European Workshop on Learning Robots, Brighton, UK, August 1997.Google Scholar
- 2.Minoru Asada, Shoichi Nods, Sukoya Tawaratumida, and Koh Hosoda. Purposive behavior acquisition for a real robot by vision-based reinforcement learning. Machine Learning, 23:279-303, 1996. Google ScholarDigital Library
- 3.J. A. Boyan and M. L. Littman. Packet routing in dynamically changing networks: A reinforcement learning approach. In J. D. Cowan, G. Tesauro, and J. Alspector, editors, Advances In Neural Information Processing Systems 6. Morgan Kaufmann Publishers, 1994.Google Scholar
- 4.Leslie Pack Kaelbling, Anthony R. Cassandra, and Michael L. Littman. Acting optimally in partially observable stochastic domains, in Proceedings of the Twelfth National Conference on Artificial Intelligence, 1994. Google ScholarDigital Library
- 5.Leslie Pack Kaelbling, Michael L. Littman, and Andrew W. Moore. Reinforcement learning: A survey. Journal of Artificial Intelligence Research, 4:237-285, May 1996. Google ScholarDigital Library
- 6.Hiroaki Kitano, Yasuo Kuniyoshi, Itsuki Nods, Minoru Asada, Hitoshi Matsubara, and Eiichi Osawa. RoboCup: A challenge problem for AI. AI Magazine, 18(1):73-85, Spring 1997.Google Scholar
- 7.Michael L. Littman. Markov games as a framework for multiagent reinforcement learning. In Proceedings of the Eleventh International Conference on Machine Learning, pages 157- 163, San Mateo, CA, 1994. Morgan Kaufman.Google ScholarDigital Library
- 8.Seen Luke, Charles Hohn, Jonathan Farris, Gary Jackson, and James Hendler. Co-evolving soccer softbot team coordination with genetic programming. In Hiroaki Kitano, editor, RoboCup-97: Robot Soccer World Cup I, pages398-411, Berlin, 1998. Springer Verlag. Google ScholarDigital Library
- 9.Maja J. Mataric. Interaction and intelligent behavior. MIT EECS PhD Thesis AITR-1495, MIT AI Lab, August 1994. Google ScholarDigital Library
- 10.Itsuki Nods, Hitoshi Matsubara, and Kazuo Hiraki. Learning cooperative behavior in multi-agent environment: a case study of choice of play-plans in soccer. In PRICAi'96: Topics in Artificial Intelligence (Proc. of dth Pacific Rim International Conference on Artificial Intelligence, Cairns, Australia), pages 570-579, Cairns,Australia, August 1996. Google Scholar
- 11.J. Ross Quinlan. C,t.5: Programs for Machine Learning. Morgan Kaufmann, San Mateo, CA, 1993. Google ScholarDigital Library
- 12.Rafal P. Salustowicz, Marco A. Wiering, and Jurgen Schmidhuber. Learning team strategies: Soccer case studies. Machine Learning, 1998. Google ScholarDigital Library
- 13.Peter Stone and Manuela Veloso. Multiagent systems: A survey from a machine learning perspective. Technical Report CMU-CS-97-193, Computer Science Department, Carnegie Mellon University, Pittsburgh, PA, December 1997.Google ScholarCross Ref
- 14.Peter Stone and Manuela Veloso. A layered approach to learning client behaviors in the RoboCup soccer server. Applied Artificial Intelligence, 12:165-188, 1998.Google ScholarCross Ref
- 15.Peter Stone and Manuela Veloso. Towards collaborative and adversarial learning: A case study in robotic soccer. International Journal of Human-Computer Studies, 48(1):83-104, January 1998. Google ScholarDigital Library
- 16.Peter Stone and Manuela Veloso. Using decision tree confidence factors for multiagent control. In Hiroaki Kitaao, editor, RoboCup-97: Robot Soccer World Cup I, pages 99-111. Springer Verlag, Berlin, 1998. Also in Proceedings of the Second International Conference on Autonomous Agents, 1998. Google ScholarDigital Library
- 17.Ming Tan. Multi-agent reinforcement learning: Independent vs. cooperative agents. In Proceedings of the Tenth International Conference on Machine Learning, pages 330-337, 1993.Google ScholarDigital Library
- 18.Manuela Veloso, Peter Stone, Kwun Hen, and Sorin Achim. The CMUnited-97 small-robot team. In Hiroaki Kitano, editor, RoboCup-97: Robot Soccer World Cup I~ pages 242-256. Springer Verlag, Berlin, 1998. Google ScholarDigital Library
- 19.Jieyu Zhao and Jurgen Schmidhuber. Incremental selfimprovement for life-time multi-agent reinforcement learning. In Proceedings of the 4th International Conference o! Simulation of Adaptive Behavior, pages 363-372.MIT Press, 1996.Google Scholar
Index Terms
- Team-partitioned, opaque-transition reinforcement learning
Recommendations
Team-Partitioned, Opaque-Transition Reinforced Learning
RoboCup-98: Robot Soccer World Cup IIWe present a novel multi-agent learning paradigm called team-partitioned, opaque-transition reinforcement learning (TPOT-RL). TPOT-RL introduces the use of action-dependent features to generalize the state space. In our work, we use a learned action-...
Reward Shaping in Episodic Reinforcement Learning
AAMAS '17: Proceedings of the 16th Conference on Autonomous Agents and MultiAgent SystemsRecent advancements in reinforcement learning confirm that reinforcement learning techniques can solve large scale problems leading to high quality autonomous decision making. It is a matter of time until we will see large scale applications of ...
Comments