2001 ConvergenceofGradientDynamicswi
- (Bowling & Veloso, 2001) ⇒ Michael H. Bowling, and Manuela M. Veloso. (2001). “Convergence of Gradient Dynamics with a Variable Learning Rate.” In: Proceedings of the Eighteenth International Conference on Machine Learning. ISBN:1-55860-778-1
Subject Headings: Multi-Agent Learning Algorithm; WoLF Algorithm.
Notes
Cited By
- http://scholar.google.com/scholar?q=%222001%22+Convergence+of+Gradient+Dynamics+with+a+Variable+Learning+Rate
- http://dl.acm.org/citation.cfm?id=645530.655659&preflayout=flat#citedby
Quotes
Abstract
As multiagent environments become more prevalent we need to understand how this changes the agent-based paradigm. One aspect that is heavily affected by the presence of multiple agents is learning. Traditional learning algorithms have core assumptions, such as Markovian transitions, which are violated in these environments. Yet, understanding the behavior of learning algorithms in these domains is critical. Singh, Kearns, and Mansour (2000) examine gradient ascent learning, specifically within a restricted class of repeated matrix games. They prove that when using this technique the average of expected payoffs over time converges. On the other hand, they also show that neither the players’ strategies nor their expected payoffs themselves are guaranteed to converge. In this paper we introduce a variable learning rate for gradient ascent, along with the WoLF (“Win or Learn Fast”) principle for regulating the learning rate. We then prove that this modification to gradient ascent has the stronger notion of convergence, that is, strategies and payoffs converge to a Nash equilibrium.
References
;
Author | volume | Date Value | title | type | journal | titleUrl | doi | note | year | |
---|---|---|---|---|---|---|---|---|---|---|
2001 ConvergenceofGradientDynamicswi | Michael H. Bowling Manuela M. Veloso | Convergence of Gradient Dynamics with a Variable Learning Rate | 2001 |