Strategic Game Theory: The Math of Choice
Table of Contents
- π The Core Framework: Nash Equilibrium
- π’ The Prisonerβs Dilemma β Payoff Matrix
- π Classic Games in Game Theory
- π Strategies in Repeated Games
- π Real-World Applications of Game Theory
- π How to Analyze a Strategic Situation
- βοΈ Cooperative vs Non-Cooperative Game Theory
- Frequently Asked Questions
- What is a Nash Equilibrium and why does it matter?
- Why doesnβt rational self-interest always produce the best collective outcome?
- How does the number of repetitions affect cooperation?
- What is the difference between pure and mixed strategies?
- How is game theory used in auction design?
- What is the Shapley Value?
- Related Courses
Game Theory β The Math Behind Strategic Choice
Game theory is the mathematical study of strategic decision-making between rational agents. Developed by John von Neumann and Oskar Morgenstern (1944) and extended by John Nash (1950), it provides a rigorous framework for analyzing any situation where the outcome depends on the choices of multiple players β from business competition to international diplomacy to evolutionary biology.
Nash Equilibrium
No Unilateral Gain
Stable strategy profile β no player benefits from deviating alone
Prisoner's Dilemma
Both Betray
Rational individuals produce collectively worse outcome
Zero-Sum Games
Win = Loss
Chess, poker β one player's gain is exactly the other's loss
Tit-for-Tat
Best in Repeats
Start cooperative, mirror last move β wins tournament play
π The Core Framework: Nash Equilibrium
John Nash proved that every finite game has at least one equilibrium β a stable point no player wants to deviate from.
s_i^* \in \arg\max_{s_i} \; u_i(s_i, s_{-i}^*) \quad \forall i
π’ The Prisonerβs Dilemma β Payoff Matrix
The most famous game in game theory. Two suspects face identical choices with no communication.
| Player A β / Player B β | B Cooperates (Stay Silent) | B Betrays (Confess) |
|---|---|---|
| A Cooperates (Stay Silent) | (1 year, 1 year) β Mutual optimum | (5 years, 0 years) β A's worst outcome |
| A Betrays (Confess) | (0 years, 5 years) β B's worst outcome | (3 years, 3 years) β Nash Equilibrium |
From Aβs perspective: if B cooperates, betraying is better (0 vs 1 year). If B betrays, betraying is still better (3 vs 5 years). So A always betrays β and B reasons identically. The Nash Equilibrium (Betray, Betray) gives 3 years each, even though (Cooperate, Cooperate) gives only 1 year each. Rational individual logic leads to collective irrationality.
π Classic Games in Game Theory
| Game | Structure | Nash Equilibrium | Key Insight | Real-World Analog |
|---|---|---|---|---|
| Prisoner's Dilemma | Non-zero-sum, simultaneous | (Betray, Betray) | Dominant strategy produces socially inferior outcome | Price wars, arms races, carbon emissions |
| Battle of the Sexes | Non-zero-sum, coordination | Two equilibria: (Opera, Opera) or (Football, Football) | Multiple equilibria β communication and commitment matter | Industry standards (VHS vs Betamax, USB-C) |
| Stag Hunt | Non-zero-sum, coordination | (Stag, Stag) or (Hare, Hare) | Payoff-dominant equilibrium exists but requires trust | Team projects, international treaties |
| Chicken (Hawk-Dove) | Non-zero-sum, anti-coordination | One swerves, one continues β mixed strategy | Commitment and credibility determine outcome | Labor strikes, nuclear standoffs |
| Matching Pennies | Zero-sum, simultaneous | Mixed strategy: each plays 50/50 | No pure strategy equilibrium β randomization is optimal | Rock-paper-scissors, penalty kicks in soccer |
| Ultimatum Game | Non-zero-sum, sequential | Theoretically: offer minimum, accept anything | Actual behavior: offers near 50/50 β fairness matters | Wage negotiations, VC term sheets |
π Strategies in Repeated Games
When the same game is played repeatedly, cooperation becomes sustainable through reputation and retaliation.
| Strategy | First Move | Subsequent Moves | Tournament Performance | Key Property |
|---|---|---|---|---|
| Always Defect | Betray | Always betray | Wins single encounters; loses repeated play | Exploitative β destroys long-term value |
| Always Cooperate | Cooperate | Always cooperate | Exploited by defectors; loses overall | Naive β cannot punish betrayal |
| Tit-for-Tat (TFT) | Cooperate | Mirror opponent's last move | Won both Axelrod tournaments | Nice, retaliatory, forgiving, clear |
| Tit-for-Tat with forgiveness | Cooperate | Mirror last move; occasionally cooperate despite betrayal | Robust against noise/mistakes | Breaks mutual defection spirals |
| Grim Trigger | Cooperate | Cooperate until first betrayal, then always defect forever | Strong deterrent; fragile to errors | Maximum punishment β lacks forgiveness |
| Pavlov (Win-Stay, Lose-Shift) | Cooperate | Repeat if outcome was good; switch if bad | Outperforms TFT in noisy environments | Self-correcting β avoids permanent defection spirals |
TFT never beats any opponent in a single game β at best it ties. Yet it won the tournament by consistently achieving near-optimal mutual cooperation. The lesson: in repeated interaction, being cooperative and predictable outperforms being exploitative. This explains why reputation, trust, and reciprocity are so economically valuable.
π Real-World Applications of Game Theory
| Domain | Game Type | Key Model | Application |
|---|---|---|---|
| Auction Design | Non-zero-sum | Vickrey (second-price) auction | Google AdWords, spectrum auctions β bidders reveal true valuations; no advantage to overbidding |
| Oligopoly Competition | Non-zero-sum | Cournot (quantity) / Bertrand (price) | Airlines, oil companies β tacit coordination without explicit agreement |
| International Trade | Repeated Prisoner's Dilemma | Tariff retaliation game | WTO as commitment device β countries cooperate because repeat play makes defection costly |
| Labor Negotiations | Bargaining game | Nash Bargaining Solution | Split of surplus proportional to each side's outside options (BATNA) |
| Platform Competition | Coordination game | Two-sided market entry | Which platform becomes standard (iOS vs Android, Slack vs Teams) depends on early coordination |
| Climate Policy | Global commons game | Tragedy of the Commons | Each country has incentive to free-ride on others' emissions reductions β classic multi-player Prisoner's Dilemma |
| Sports Strategy | Zero-sum | Mixed strategy equilibrium | Penalty kick direction, serve placement in tennis β optimal play requires randomization to be unpredictable |
π How to Analyze a Strategic Situation
Step-by-step game theory analysis framework
Identify Players, Actions, and Payoffs
Who are the decision-makers? What can each player do? What does each player receive for each combination of actions? Build the payoff matrix.
Check for Dominant Strategies
Does any player have a strategy that is best regardless of what others do? Eliminate dominated strategies β rational players never use them. This simplifies the analysis.
Find Nash Equilibria
For each cell in the payoff matrix, check: can either player improve their payoff by switching? If no player can improve, it is a Nash Equilibrium. Multiple equilibria may exist.
Consider the Game Type
Is this one-shot or repeated? Simultaneous or sequential? Zero-sum or cooperative? Repeated play enables cooperation. Sequential play creates first-mover or follower advantages.
βοΈ Cooperative vs Non-Cooperative Game Theory
| κ΅¬λΆ | Non-Cooperative Game Theory | Cooperative Game Theory |
|---|---|---|
| Focus | Individual player strategies and self-interest | Coalition formation and fair division of collective gains |
| Key solution concept | Nash Equilibrium β stable individual strategies | Shapley Value β fair distribution of coalition surplus |
| Communication | Binding agreements not enforceable | Binding contracts and side payments allowed |
| Examples | Oligopoly competition, auctions, international trade | Joint ventures, labor unions, vote trading in legislatures |
| Limitation | May reach inefficient equilibria (Prisoner's Dilemma) | Requires enforceable contracts; ignores strategic manipulation |
Frequently Asked Questions
What is a Nash Equilibrium and why does it matter?
A Nash Equilibrium is a stable configuration where no player can improve their outcome by unilaterally changing their strategy. It matters because it predicts where rational strategic interaction will settle. Nash proved that every finite game has at least one equilibrium (possibly in mixed strategies). However, equilibria are not always efficient β the Prisonerβs Dilemma has a Nash Equilibrium that is worse for everyone than the cooperative outcome.
Why doesnβt rational self-interest always produce the best collective outcome?
The Prisonerβs Dilemma demonstrates this precisely: each player follows the individually rational dominant strategy (betray), yet both end up worse than if they had cooperated. This conflict between individual and collective rationality underlies many social problems: carbon emissions, overfishing, price wars, arms races. Solutions include binding agreements, reputation mechanisms, regulation, or repeated interaction (where future cooperation becomes valuable enough to outweigh short-term defection gains).
How does the number of repetitions affect cooperation?
In a finitely repeated game with a known end point, backward induction logic unravels cooperation: both players know they will defect on the last round, so they defect on the second-to-last, and so on. In infinitely repeated games (or games with uncertain endpoints), the βfolk theoremβ shows that cooperation can be sustained if players value future payoffs enough β if the discount factor is high enough, the threat of future punishment makes cooperation individually rational.
What is the difference between pure and mixed strategies?
A pure strategy is a deterministic choice (always cooperate, always betray). A mixed strategy assigns probabilities to different actions (e.g., betray with 60% probability). In zero-sum games like matching pennies, no pure strategy equilibrium exists β the optimal solution requires randomization so opponents cannot predict and exploit your pattern. Mixed strategy equilibria always exist when pure strategy equilibria do not.
How is game theory used in auction design?
Auction mechanism design is a direct application of game theory. The Vickrey (second-price sealed-bid) auction is strategy-proof: bidding your true value is a dominant strategy regardless of othersβ bids. Google AdWords uses a generalized second-price auction. Spectrum auctions (for mobile phone bandwidth) use simultaneous ascending auctions designed by game theorists to allocate resources efficiently and prevent collusion.
What is the Shapley Value?
The Shapley Value (Lloyd Shapley, Nobel 2012) is a fair allocation method for cooperative games. It assigns each player their average marginal contribution across all possible coalition orderings. It is used in cost allocation (how much each partner contributes to a joint project), voting power analysis, and machine learning (SHAP values for feature importance β a direct application of the Shapley Value concept).