Nash Equilibria

A game can be represented by:

  • Players N={1,β‹―n}N = \{1, \cdots n \}

  • Actions β†’ players can do something to affect the world

  • Outcomes β†’ players’ actions lead to outcomes

  • Preferences over outcomes β†’ each player pip_i has a preferences / ranking of the outcomes (in terms of utility)

This provides a general, abstract framework for strategic interaction.

Pure Nash Equilibrium

Given everyone else’s actions aβƒ—βˆ’i\vec a_{-i}, the best response set of player ii is:

BRi(aβƒ—βˆ’i)={b∈Ai∣b∈arg max⁑ui(aβƒ—βˆ’i,b)}BR_i(\vec a_{-i}) = \{ b \in A_i | b \in \argmax u_i(\vec a_{-i}, b) \}

That is, a player ii’s best response can be any of the actions that give him the maximum possible utility, given everyone else’s actions.

An action profile a⃗\vec a is a (pure) Nash equilibrium if:

βˆ€i∈N,ai∈BRi(aβƒ—βˆ’i)\forall i \in N, a_i \in BR_i(\vec a_{-i})

That is, every player has picked their best response given everyone else’s actions. You can also think of it as: each player picks the action as if they were the last one to do so, and they knew everyone else’s actions.

Every player thinks β€œI’m doing the best I can, given everyone else’s actions.”

Mixed Nash Equilibrium

Instead of choosing a single action, one can play a random mix of them (to be unpredictable, otherwise if someone knows you well): pβƒ—iβˆˆΞ”(Ai)\vec p_i \in \Delta(A_i) is a probability distribution (from the set of all possible probability distributions) over player ii’s actions.

(Note: In general, we use Ξ”(X)\Delta(X) to represent the set of all possible distributions over X)

A (not necessarily pure) strategy profile can be represented as:

pβƒ—=(pβƒ—1,⋯ ,pβƒ—n)βˆˆΞ”(A1)Γ—β‹―Γ—Ξ”(An)\vec p = (\vec p_1, \cdots, \vec p_n) \in \Delta(A_1) \times \cdots \times \Delta(A_n)

Then, a player’s utility (for a given mixed strategy) is given by:

ui(pβƒ—)=βˆ‘aβƒ—βˆˆAui(aβƒ—)Pr⁑[aβƒ—]=Eaβƒ—βˆΌpβƒ—[ui(aβƒ—)]u_i(\vec p) = \sum_{\vec a \in A} u_i(\vec a)\Pr[\vec a] = E_{\vec a \sim \vec p}[u_i(\vec a)]

In particular, note that the player’s ONLY care about the expected value of their strategy, NOT the variance / risk that their strategy runs. e.g. they would be indifferent to a strategy that guarantees them $50 vs. flipping a coin for $100 β€” both have the same expected value, and being β€œrisk-neutral”, they wouldn’t care which one they did. That is, they are neutral / indifferent to risk β†’ they don’t care if they take risk or don’t (they don’t avoid it - by taking the gauranteed money - nor do they seek it - by taking the coin flip; they’re just indifferent).

Then, pβƒ—\vec p is a (not necessarily pure) nash equilibrium is when: for all players i∈Ni \in N, and all possible strategy (possibly containing mixed actions) of each player qiβˆˆΞ”(Ai)q_i \in \Delta(A_i), ui(pβƒ—)β‰₯ui(pβƒ—βˆ’i,qβƒ—i)u_i(\vec p) \geq u_i(\vec p_{-i}, \vec q_i) β†’ their current strategy is at least as good as any other.

In these notes, we use β€œaction” to refer to a pure strategy (only one possible action, played with probability = 1), and the general term β€œstrategy” to refer to possibly mixed actions (i.e., playing multiple actions, each with some probability)

Nash’s Theorem: Unlike a pure Nash Equilibrium, a (not necessarily pure) Nash Equilibrium ALWAYS EXISTS.

The way to compute nash equilibria in a 2x2 games (2 players, each have 2 actions) is by considering 2 possible cases:

  1. Case 1 β†’ Compute all NE (if any) in which at least one player plays a pure strategy (do this exhaustively, e.g. β€œrow player plays action X while col player mixes between A and B with probability q and (1-q), then find the value(s) of q for which the best response is X β€” then you’ve found a NE)

  2. Case 2 β†’ Compute all NE (if any) in which both players mix between both strategies

Nash’s theorem says that the union of the set of NE’s found in case 1 and case 2 is non-empty β†’ it doesn’t say whether the NE is pure / not.

The key idea behind case 2 is that the only time that a player would mix between 2 strategies is when (given the opponent’s strategy), he is indifferent to his 2 actions. That is, no matter what action he picks, his utility is the same given the opponent’s strategy. This allows us to equate the expected utilities of the two actions for player 1, to find the β€œmixing probabilities” for player 2 😲 (and vice versa).

Dominant Strategies

We say that a strategy pβƒ—βˆˆΞ”(Ai)\vec p \in \Delta(A_i) (weakly) dominates qβƒ—βˆˆΞ”(Ai)\vec q \in \Delta(A_i) if:

βˆ€pβƒ—βˆ’iβˆˆΞ”(Aβˆ’i):ui(pβƒ—βˆ’i,pβƒ—)β‰₯ui(pβƒ—βˆ’i,qβƒ—)\forall \vec p_{-i} \in \Delta(A_{-i}) : u_i(\vec p_{-i}, \vec p) \geq u_i(\vec p_{-i}, \vec q)

In words: no matter what everyone else does, picking p⃗\vec p is at least as good as picking q⃗\vec q.

Strict domination is when ui(pβƒ—βˆ’i,pβƒ—)>ui(pβƒ—βˆ’i,qβƒ—)u_i(\vec p_{-i}, \vec p) > u_i(\vec p_{-i}, \vec q) for every pβƒ—βˆ’i\vec p_{-i} β†’ no matter what everyone else does, strategy pβƒ—\vec p is STRICTLY better than qβƒ—\vec q (and so, you should never even consider picking qβƒ—\vec q !!).

Theorem: If an action a∈Aia \in A_i is STRICTLY dominated by some strategy pβƒ—βˆˆΞ”(Ai)\vec p \in \Delta(A_i), then action aa is NEVER PLAYED with any positive probability in ANY Nash equilibrium.

Intuitively, if you would move all the probability out of the lamer / weaker actions into the ones that dominate it, and you strictly increase your utility. Hence, picking aa (with any non-zero probability) can never be a β€œbest” response 🀯

Iterated Removal of Strictly Dominated Strategies

If you’re trying to find Nash Equilibria, then you can remove action profiles that contain a strictly dominated strategy (from either player’s side).

Note:

  • This ONLY works if the action is strictly dominated by another strategy (can be a strategy involving multiple actions β€” using mixed probabilities too)

  • If you’re left with a single action / strategy profile at the end, then by Nash Theorem, it must be a NE.

  • The reason this works is simply because those cells which contain a strictly dominated strategy (from either player’s side) are never going to happen, i.e., they can never be an equilibrium (because at least one player would change their strategy to the one which strictly dominates it)

  • When considrering whether a strategy SS (may be mixed) strictly dominates an action PP for the row player, we don’t have to consider all possible strategies of the col player β€” we can just look at the individual actions and ensure that SS strictly dominates PP for every action. Then, since any strategy is a linear combination of actions, if SS strictly beats PP for every action of the col player, then SS strictly beats PP for every strategy involving those actions too. (Moreover, we only have to lok at the actions of the col players that are not already strictly dominated by other strategies β€” since col player would not play the weaker actions otherwise).

Last updated