Potential game
Potential game
In game-theory, a game is said to be a potential game if the incentive of all players to change their strategy can be expressed using a single global function called the potential function. The concept originated in a 1996 paper by Dov Monderer and lloyd-shapley.
The properties of several types of potential games have since been studied. Games can be either ordinal or cardinal potential games. In cardinal games, the difference in individual payoffs for each player from individually changing one's strategy, other things equal, has to have the same value as the difference in values for the potential function. In ordinal games, only the signs of the differences have to be the same.
The potential function is a useful tool to analyze equilibrium properties of games, since the incentives of all players are mapped into one function, and the set of pure Nash equilibria can be found by locating the local optima of the potential function. Convergence and finite-time convergence of an iterated game towards a Nash equilibrium can also be understood by studying the potential function.
Potential games can be studied as repeated-games with state so that every round played has a direct consequence on game's state in the next round. This approach has applications in distributed control such as distributed resource allocation, where players without a central correlation mechanism can cooperate to achieve a globally optimal resource distribution.
Definition Let <math>N</math> be the number of players, <math>A</math> the set of action profiles over the action sets <math>A_{i}</math> of each player and <math>u_i:A \to \mathbb{R}</math> be the payoff function for player <math>1\le i\le N</math>.
Given a game <math>G=(N,A=A_{1}\times\ldots\times A_{N}, u: A \rightarrow \reals^N) </math>, we say that <math>G</math> is a potential game with an exact (weighted, ordinal, generalized ordinal, best response) potential function if <math>\Phi: A \rightarrow \reals</math> is an exact (weighted, ordinal, generalized ordinal, best response, respectively) potential function for <math>G</math>. We define these notions below.
<math>\Phi</math> is called an exact potential function if <math> \forall i, \forall {a_{-i}\in A_{-i}},\ \forall {a'_{i},\ a_{i}\in A_{i}}</math>, ::<math> \Phi(a'_{i},a_{-i})-\Phi(a_{i},a_{-i}) = u_{i}(a'_{i},a_{-i})-u_{i}(a_{i},a_{-i})</math>
::That is: when player <math>i</math> switches from action <math>a'</math> to action <math>a*</math>, the change in the potential <math>\Phi</math> equals the change in the utility of that player.
<math>\Phi</math> is called a weighted potential function if there is a vector <math>w \in \reals_{++}^N</math> such that <math> \forall i,\forall {a_{-i}\in A_{-i}},\ \forall {a'_{i},\ a_{i}\in A_{i}}</math>, ::<math> \Phi(a'_{i},a_{-i})-\Phi(a_{i},a_{-i}) = w_{i}(u_{i}(a'_{i},a_{-i})-u_{i}(a_{i},a_{-i}))</math> That is: when a player switches action, the change in <math>\Phi</math> equals the change in the player's utility, times a positive player-specific weight. Every exact PF is a weighted PF with wi=1 for all i.
<math>\Phi</math> is called an ordinal potential function if <math> \forall i,\forall {a_{-i}\in A_{-i}},\ \forall {a'_{i},\ a_{i}\in A_{i}}</math>, ::<math> u_{i}(a'_{i},a_{-i})-u_{i}(a_{i},a_{-i})>0 \Leftrightarrow \Phi(a'_{i},a_{-i})-\Phi(a_{i},a_{-i})>0</math> That is: when a player switches action, the sign of the change in <math>\Phi</math> equals the sign of the change in the player's utility, whereas the magnitude of change may differ. Every weighted PF is an ordinal PF.
<math>\Phi</math> is called a generalized ordinal potential function if <math>\forall i, \forall {a_{-i}\in A_{-i}},\ \forall {a'_{i},\ a_{i}\in A_{i}}</math>, ::<math> u_{i}(a'_{i},a_{-i})-u_{i}(a_{i},a_{-i})>0 \Rightarrow \Phi(a'_{i},a_{-i})-\Phi(a_{i},a_{-i}) >0 </math> That is: when a player switches action, if the player's utility increases, then the potential increases (but the opposite is not necessarily true). Every ordinal PF is a generalized-ordinal PF.
<math>\Phi</math> is called a best-response potential function if <math>\forall i\in N,\ \forall {a_{-i}\in A_{-i}}</math>, ::<math>\arg\max_{a_i\in A_i} u_i(a_i,a_{-i}) =\arg\max_{a_i\in A_i} \Phi(a_i,a_{-i})</math>. That is: for each player i*, maximizing the common potential function leads to the same outcome as maximizing his own utility.
- <math>\Phi</math> is called a **pseudo-potential function proved that every [[congestion-game]] has an exact potential; Monderer and Shapley proved that, for every problem in the complexity class PLS (essentially, every local search problem), there exists an ordinal potential game with polynomially many players, such that the set of pure Nash equilibria equals the set of local optima.
Potential games and improvement paths An improvement path (also called Nash dynamics) is a sequence of strategy-vectors, in which each vector is attained from the previous vector by a single player switching his strategy to a strategy that strictly increases his utility. If a game has a generalized-ordinal-potential function <math>\Phi</math>, then <math>\Phi</math> is strictly increasing in every improvement path, so every improvement path is acyclic. If, in addition, the game has finitely many strategies, then every improvement path must be finite. This property is called the finite improvement property (FIP). We have just proved that every finite generalized-ordinal-potential game has the FIP. The opposite is also true: every finite game has the FIP has a generalized-ordinal-potential function. The terminal state in every finite improvement path is a Nash equilibrium, so FIP implies the existence of a pure-strategy Nash equilibrium. Moreover, it implies that a Nash equilibrium can be computed by a distributed process, in which each agent only has to improve his own strategy.
A best-response path is a special case of an improvement path, in which each vector is attained from the previous vector by a single player switching his strategy to a best-response strategy. The property that every best-response path is finite is called the finite best-response property (FBRP). FBRP is weaker than FIP, and it still implies the existence of a pure-strategy Nash equilibrium. It also implies that a Nash equilibrium can be computed by a distributed process, but the computational burden on the agents is higher than with FIP, since they have to compute a best-response.
An even weaker property is **weak-acyclicity (WA)*'. It means that, for any initial strategy-vector, *there exists* a finite best-response path starting at that vector. Weak-acyclicity is not sufficient for existence of a potential function (since some improvement-paths may be cyclic), but it is sufficient for the existence of pure-strategy Nash equilibrium. It implies that a Nash equilibrium can be computed almost-surely by a stochastic distributed process, in which at each point, a player is chosen at random, and this player chooses a best-strategy at random. prove:
- Any game of weak strategic substitutes or strategic-complements with aggregation is a pseudo-potential game.
- Any pseudo-potential game has a pure-strategy Nash equilibrium.
Correlated equilibria abraham-neyman studied potential games in which (a) the potential is a smooth and concave function, (b) the strategy sets are convex, (c) the utilities are bounded. He proves that, in such games, any correlated-equilibrium is a mixture of pure strategy profiles which maximize the potential.
If, in addition, (d) the strategy sets are compact, (e) the potential is a strictly concave function, then the correlated equilibrium is unique.