甲鱼和什么食物相克| 茯苓有什么功效和作用| 一毛不拔是什么生肖| instagram是什么| kap是什么意思| 先天性心脏病最怕什么| 宝宝为什么喜欢趴着睡| 香菜炒什么好吃| 屈光是什么意思| saa是什么检查| 背锅侠是什么意思| 12月31号什么星座| 水柔棉是什么面料| 风声鹤唳是什么意思| 爱妃是什么意思| 1996年是什么命| 肠胃炎引起的发烧吃什么药| 眼睛总是干涩是什么原因| 怀孕牙龈出血是什么原因| 什么东西补肾| 胃ca是什么意思| 刘备是一个什么样的人| 经常手瘾吃什么药| 英文为什么怎么写| 吃什么不会长胖| 什么是钓鱼执法| 吃什么可以降尿酸| 1103是什么星座| 46岁属什么| 梧桐树长什么样子| 逍遥丸治什么| 山谷念什么| 眼睛散光是什么意思| 三晋是什么意思| 什么花不能浇水| 拔完牙不能吃什么| 籍贯填什么| 凤毛麟角是什么生肖| 男人喝什么茶壮阳| 蟑螂长什么样子| 肚子里有虫子会有什么症状| 吃莲雾有什么好处| 牡丹花像什么| 哀鸿遍野什么意思| 去冰和常温有什么区别| 欠缺是什么意思| 舌苔发白什么原因| sc1是什么意思| 九月什么花开| 德字五行属什么| 吃李子有什么好处和坏处| 肺鳞癌是什么意思| 为什么会突然晕倒| 他喵的什么意思| 这是什么猫| vd是什么| 洋地黄是什么药| 梦见家里死人了代表什么预兆| 什么三迁| 蝎子泡酒有什么功效| 冲正是什么意思| 扩心病是什么病| 流汗有什么好处| 小分子肽能治什么病| 水浒传什么朝代| 18岁是什么生肖| gy是什么意思| 送男生什么生日礼物好| 马栗是什么植物| 解痉镇痛酊有什么功效| 微量元素六项是什么检查| 勾魂是什么意思| 来大姨妈肚子疼是什么原因| 寻麻疹不能吃什么| 金灿灿的什么| 久之的之是什么意思| 白带发绿是什么原因| 什么食物高蛋白含量高| 985和211有什么区别| 木隶念什么| 吃西红柿有什么好处和坏处| 荷花又什么又什么| 什么时候闰三月| 68属什么生肖| 小脑延髓池是什么意思| 地中海贫血有什么症状| cob是什么意思| 为什么一直打嗝| 胃镜能检查出什么| 梦见给别人剪头发是什么意思| 奄奄一息的息是什么意思| 小孩子肚子痛吃什么药| 修容是什么意思| 拉杆箱什么材质好| 梦见吃梨是什么意思| aep是什么意思| 3月16日什么星座| 爱是什么偏旁| 什么是心律失常| 1009是什么星座| 牙周炎是什么| 瞌睡多什么原因| 骨挫伤是什么意思| 苋菜是什么菜| 什么窃什么盗| 戒手淫有什么好处| 初中毕业可以考什么证| 牙齿痛用什么药| 鲨鱼为什么怕海豚| 出栏是什么意思| 眩晕症什么症状| 碧霄是什么意思| 猹是什么| 长期失眠吃什么药| 右侧中耳乳突炎是什么意思| 洋桔梗花语是什么| 胎盘成熟度1级是什么意思| 血色素低吃什么补得快| 最好的止疼药是什么药| 腿毛多是什么原因| 被是什么偏旁怎么读| 爱啃指甲是什么原因| 小孩上火吃什么药| 吉兰巴雷综合征是什么病| 脚底发麻是什么病的前兆| 五十坐地能吸土是什么意思| 哪吒妈妈叫什么| 什么日什么里| 鸡肉配什么菜好吃| 西铁城手表属于什么档次| 早上起床口苦是什么原因| 皮炎是什么原因引起的| 过的第五笔是什么| 坐车晕车是什么原因| 缪在姓氏中读什么| 香奶奶是什么牌子| 鞭挞是什么意思| 刮痧用的油是什么油| 喝普洱茶有什么好处| 胜字五行属什么| 非那雄胺片是什么药| 中指长痣代表什么| 发难是什么意思| 阿迪达斯和三叶草有什么区别| 晨勃是什么意思啊| 喘是什么意思| 优生优育检查什么项目| 拉肚子挂什么科室| 乳房疼痛应该挂什么科| 小孩抽动症是什么引起的| 十一月七号是什么星座| 坦诚相待下一句是什么| 为什么空调外机会滴水| ppsu是什么材质| 怀孕周期是从什么时候开始算的| 17号来月经什么时候是排卵期| 梦见一个人代表什么| 桃子和什么相克| 客服是什么工作| 血压低什么症状| 直辖市市长是什么级别| 什么是爱情| 9月10日什么星座| plg是什么意思| 为什么吃鸽子刀口长得快| 什么东西能解酒| 叶酸有什么好处| 咳嗽发烧吃什么药| iwc手表是什么档次| 独善其身是什么意思啊| 跳票什么意思| 物质是由什么组成的| 口腔医学专业学什么| 几月初几是叫什么历| 晕车药什么时候吃最好| 岁月匆匆是什么意思| 当归不能和什么一起吃| 看胸部挂什么科| 床垫什么材质的好| 骨科什么意思| 琬字五行属什么| 噩耗是什么意思| 第一次要注意什么| 麻椒和花椒有什么区别| 镜子是用什么做的| 经常吃莲子有什么好处| 烧心吃什么| 油嘴滑舌指什么生肖| 黄连泡水喝能治什么病| 腔梗是什么| 孩子打喷嚏流鼻涕吃什么药| otc是什么药| 一张纸可以折什么| 司命星君掌管什么| 螃蟹吃什么| 雷人是什么意思| hoka是什么牌子| 阴历六月十三是什么日子| 家里为什么不能放假花| 哪吒妈妈叫什么名字| 头颈出汗多是什么原因| 心率快吃什么中成药| 什么什么斜斜| 产妇吃什么下奶快又多| 第二天叫什么日| 六月二十五号是什么星座| 撒野是什么意思| 西汉后面是什么朝代| 53岁属什么| 产后吃什么水果好| 0706是什么星座| 做爱是什么感觉| 光明磊落是什么生肖| 取关是什么意思| 王火火念什么| 月经不停吃什么药| 眼皮为什么会跳| 舌苔发苦是什么原因| 随心所欲的欲什么意思| 为什么身上痒一抓就起疙瘩| 内分泌失调吃什么药| 尿潜血是什么原因造成的| fast什么意思| 甲状腺做什么检查最准确| cml是什么意思| 公鸭嗓是什么声音| 系带断裂有什么影响吗| 舐犊是什么意思| 老人吃什么水果对身体好| hospital是什么意思| 进是什么结构| amazon是什么意思| 小苏打是什么成分| 农历10月是什么月| 婴儿掉头发是什么原因| 梦到老公被蛇咬是什么意思| 公章一般是什么字体| 灰配什么颜色好看| 996是什么| 庞统为什么叫凤雏| 尿液中有白色沉淀物是什么原因| mrsa医学上是什么意思| 6月15是什么星座| 什么叫扁平疣长什么样| 旭五行属性是什么| 笑气是什么气体| 他叫什么名字| 通告是什么意思| 花团锦簇什么意思| 合肥古代叫什么| 睡觉为什么会磨牙| 高姓和什么姓是世仇| 神戳戳是什么意思| 乱伦是什么| 脱毛膏是什么原理| 纳纹女装属于什么档次| spo2是什么意思| 梦见白蛇是什么预兆| 普拉提和瑜伽有什么区别| 孕激素低吃什么补得快| 始祖鸟什么档次| 属虎是什么命| 呻吟是什么意思| 酒后头疼吃什么| 百度Jump to content

速拓服装鞋帽管理系统(迷你版) V16.0928官方最新版

From Wikipedia, the free encyclopedia
Nash equilibrium
Solution concept in game theory
Relationship
Subset ofRationalizability, Epsilon-equilibrium, Correlated equilibrium
Superset ofEvolutionarily stable strategy, Subgame perfect equilibrium, Perfect Bayesian equilibrium, Trembling hand perfect equilibrium, Stable Nash equilibrium, Strong Nash equilibrium
Significance
Proposed byJohn Forbes Nash Jr.
Used forAll non-cooperative games
百度 清东陵龙门湖上沙鸥嬉戏,黎河岸边白鹭栖息,大鸨鸟、大雁等国家级野生保护鸟类来遵化“安家落户”……这些都是最好见证。

In game theory, a Nash equilibrium is a situation where no player could gain more by changing their own strategy (holding all other players' strategies fixed) in a game. Nash equilibrium is the most commonly used solution concept for non-cooperative games.[1]

If each player has chosen a strategy – an action plan based on what has happened so far in the game – and no one can increase one's own expected payoff by changing one's strategy while the other players keep theirs unchanged, then the current set of strategy choices constitutes a Nash equilibrium.

If two players Alice and Bob choose strategies A and B, (A, B) is a Nash equilibrium if Alice has no other strategy available that does better than A at maximizing her payoff in response to Bob choosing B, and Bob has no other strategy available that does better than B at maximizing his payoff in response to Alice choosing A. In a game in which Carol and Dan are also players, (A, B, C, D) is a Nash equilibrium if A is Alice's best response to (B, C, D), B is Bob's best response to (A, C, D), and so forth.

The idea of Nash equilibrium dates back to the time of Cournot, who in 1838 applied it to his model of competition in an oligopoly.[2] John Nash showed that there is a Nash equilibrium, possibly in mixed strategies, for every finite game.[3]

Applications

[edit]

Game theorists use Nash equilibrium to analyze the outcome of the strategic interaction of several decision makers. In a strategic interaction, the outcome for each decision-maker depends on the decisions of the others as well as their own. The simple insight underlying Nash's idea is that one cannot predict the choices of multiple decision makers if one analyzes those decisions in isolation. Instead, one must ask what each player would do taking into account what the player expects the others to do. Nash equilibrium is achieved when no player can improve their outcome by changing their decision, assuming the other players' choices remain unchanged.

The concept has been used to analyze hostile situations such as wars and arms races[4] (see prisoner's dilemma), and also how conflict may be mitigated by repeated interaction (see tit-for-tat). It has also been used to study to what extent people with different preferences can cooperate (see battle of the sexes), and whether they will take risks to achieve a cooperative outcome (see stag hunt). It has been used to study the adoption of technical standards,[citation needed] and also the occurrence of bank runs and currency crises (see coordination game). Other applications include traffic flow (see Wardrop's principle), how to organize auctions (see auction theory), the outcome of efforts exerted by multiple parties in the education process,[5] regulatory legislation such as environmental regulations (see tragedy of the commons),[6] natural resource management,[7] analysing strategies in marketing,[8] penalty kicks in football (I.e. soccer; see matching pennies),[9] robot navigation in crowds,[10] energy systems, transportation systems, evacuation problems[11] and wireless communications.[12]

History

[edit]

Nash equilibrium is named after American mathematician John Forbes Nash Jr. The same idea was used in a particular application in 1838 by Antoine Augustin Cournot in his theory of oligopoly.[13] In Cournot's theory, each of several firms choose how much output to produce to maximize its profit. The best output for one firm depends on the outputs of the others. A Cournot equilibrium occurs when each firm's output maximizes its profits given the output of the other firms, which is a pure-strategy Nash equilibrium. Cournot also introduced the concept of best response dynamics in his analysis of the stability of equilibrium. Cournot did not use the idea in any other applications, however, or define it generally.

The modern concept of Nash equilibrium is instead defined in terms of mixed strategies, where players choose a probability distribution over possible pure strategies (which might put 100% of the probability on one pure strategy; such pure strategies are a subset of mixed strategies). The concept of a mixed-strategy equilibrium was introduced by John von Neumann and Oskar Morgenstern in their 1944 book The Theory of Games and Economic Behavior, but their analysis was restricted to the special case of zero-sum games. They showed that a mixed-strategy Nash equilibrium will exist for any zero-sum game with a finite set of actions.[14] The contribution of Nash in his 1951 article "Non-Cooperative Games" was to define a mixed-strategy Nash equilibrium for any game with a finite set of actions and prove that at least one (mixed-strategy) Nash equilibrium must exist in such a game. The key to Nash's ability to prove existence far more generally than von Neumann lay in his definition of equilibrium. According to Nash, "an equilibrium point is an n-tuple such that each player's mixed strategy maximizes [their] payoff if the strategies of the others are held fixed. Thus each player's strategy is optimal against those of the others." Putting the problem in this framework allowed Nash to employ the Kakutani fixed-point theorem in his 1950 paper to prove existence of equilibria. His 1951 paper used the simpler Brouwer fixed-point theorem for the same purpose.[15]

Game theorists have discovered that in some circumstances Nash equilibrium makes invalid predictions or fails to make a unique prediction. They have proposed many solution concepts ('refinements' of Nash equilibria) designed to rule out implausible Nash equilibria. One particularly important issue is that some Nash equilibria may be based on threats that are not 'credible'. In 1965 Reinhard Selten proposed subgame perfect equilibrium as a refinement that eliminates equilibria which depend on non-credible threats. Other extensions of the Nash equilibrium concept have addressed what happens if a game is repeated, or what happens if a game is played in the absence of complete information. However, subsequent refinements and extensions of Nash equilibrium share the main insight on which Nash's concept rests: the equilibrium is a set of strategies such that each player's strategy is optimal given the choices of the others.

Definitions

[edit]

A strategy profile is a set of strategies, one for each player. Informally, a strategy profile is a Nash equilibrium if no player can do better by unilaterally changing their strategy. To see what this means, imagine that each player is told the strategies of the others. Suppose then that each player asks themselves: "Knowing the strategies of the other players, and treating the strategies of the other players as set in stone, can I benefit by changing my strategy?"

For instance if a player prefers "Yes", then that set of strategies is not a Nash equilibrium. But if every player prefers not to switch (or is indifferent between switching and not) then the strategy profile is a Nash equilibrium. Thus, each strategy in a Nash equilibrium is a best response to the other players' strategies in that equilibrium.[16]

Formally, let be the set of all possible strategies for player , where . Let be a strategy profile, a set consisting of one strategy for each player, where denotes the strategies of all the players except . Let be player i's payoff as a function of the strategies. The strategy profile is a Nash equilibrium if

A game can have more than one Nash equilibrium. Even if the equilibrium is unique, it might be weak: a player might be indifferent among several strategies given the other players' choices. It is unique and called a strict Nash equilibrium if the inequality is strict so one strategy is the unique best response:

The strategy set can be different for different players, and its elements can be a variety of mathematical objects. Most simply, a player might choose between two strategies, e.g. Or the strategy set might be a finite set of conditional strategies responding to other players, e.g. Or it might be an infinite set, a continuum or unbounded, e.g. such that is a non-negative real number. Nash's existing proofs assume a finite strategy set, but the concept of Nash equilibrium does not require it.

Variants

[edit]

Pure/mixed equilibrium

[edit]

A game can have a pure-strategy or a mixed-strategy Nash equilibrium. In the latter, not every player always plays the same strategy. Instead, there is a probability distribution over different strategies.

Strict/non-strict equilibrium

[edit]

Suppose that in the Nash equilibrium, each player asks themselves: "Knowing the strategies of the other players, and treating the strategies of the other players as set in stone, would I suffer a loss by changing my strategy?"

If every player's answer is "Yes", then the equilibrium is classified as a strict Nash equilibrium.[17]

If instead, for some player, there is exact equality between the strategy in Nash equilibrium and some other strategy that gives exactly the same payout (i.e. the player is indifferent between switching and not), then the equilibrium is classified as a weak[note 1] or non-strict Nash equilibrium[citation needed][clarification needed].

Equilibria for coalitions

[edit]

The Nash equilibrium defines stability only in terms of individual player deviations. In cooperative games such a concept is not convincing enough. Strong Nash equilibrium allows for deviations by every conceivable coalition.[18] Formally, a strong Nash equilibrium is a Nash equilibrium in which no coalition, taking the actions of its complements as given, can cooperatively deviate in a way that benefits all of its members.[19] However, the strong Nash concept is sometimes perceived as too "strong" in that the environment allows for unlimited private communication. In fact, strong Nash equilibrium has to be weakly Pareto efficient. As a result of these requirements, strong Nash is too rare to be useful in many branches of game theory. However, in games such as elections with many more players than possible outcomes, it can be more common than a stable equilibrium.

A refined Nash equilibrium known as coalition-proof Nash equilibrium (CPNE)[18] occurs when players cannot do better even if they are allowed to communicate and make "self-enforcing" agreement to deviate. Every correlated strategy supported by iterated strict dominance and on the Pareto frontier is a CPNE.[20] Further, it is possible for a game to have a Nash equilibrium that is resilient against coalitions less than a specified size, k. CPNE is related to the theory of the core.

Existence

[edit]

Nash proved that if mixed strategies (where a player chooses probabilities of using various pure strategies) are allowed, then every game with a finite number of players in which each player can choose from finitely many pure strategies has at least one Nash equilibrium, which might be a pure strategy for each player or might be a probability distribution over strategies for each player.

Nash equilibria need not exist if the set of choices is infinite and non-compact. For example:

  • A game where two players simultaneously name a number and the player naming the larger number wins does not have a NE, as the set of choices is not compact because it is unbounded.
  • Each of two players chooses a real number strictly less than 5 and the winner is whoever has the biggest number; no biggest number strictly less than 5 exists (if the number could equal 5, the Nash equilibrium would have both players choosing 5 and tying the game). Here, the set of choices is not compact because it is not closed.

However, a Nash equilibrium exists if the set of choices is compact with each player's payoff continuous in the strategies of all the players.[21]

Generalizations

[edit]

Nash's existence theorem has been extended to more general classes of games:

  • Concave games: the space of possible strategy profiles is an arbitrary convex set. This means that the space of strategies available to a player may depend on the strategies chosen by other players (in Nash's theorem, the space of strategy profiles is a Cartesian product of simplices; the space of strategies available to each player is a simplex representing all mixed strategies, and it does not depend on other players' actions). Moreover, the payoff of each player may be an arbitrary function of the actions of the players, as long as it is a concave function of the player's own action. Rosen[22] proved that an equilibrium exists under certain conditions; see concave game.
  • Non-atomic games: the set of players is infinite - there is a continuum of players (in Nash's theorem, the set is finite). David Schmeidler[23] proved that an equilibrium exists under certain conditions; see non-atomic game.

Rationality

[edit]

The Nash equilibrium may sometimes appear non-rational in a third-person perspective. This is because a Nash equilibrium is not necessarily Pareto optimal.

Nash equilibrium may also have non-rational consequences in sequential games because players may "threaten" each other with threats they would not actually carry out. For such games the subgame perfect Nash equilibrium may be more meaningful as a tool of analysis.

Examples

[edit]

Coordination game

[edit]
A coordination game showing payoffs for player 1 (row) and player 2 (column)
Player 1
strategy
Player 2 strategy
A B
A
4
4
3
1
B
1
3
2
2

The coordination game is a classic two-player, two-strategy game, as shown in the example payoff matrix to the right. There are two pure-strategy equilibria, (A,A) with payoff 4 for each player and (B,B) with payoff 2 for each. The combination (B,B) is a Nash equilibrium because if either player unilaterally changes their strategy from B to A, their payoff will fall from 2 to 1.

The stag hunt
Player 1
strategy
Player 2 strategy
Hunt stag Hunt rabbit
Hunt stag
2
2
1
0
Hunt rabbit
0
1
1
1

A famous example of a coordination game is the stag hunt. Two players may choose to hunt a stag or a rabbit, the stag providing more meat (4 utility units, 2 for each player) than the rabbit (1 utility unit). The caveat is that the stag must be cooperatively hunted, so if one player attempts to hunt the stag, while the other hunts the rabbit, the stag hunter will totally fail, for a payoff of 0, whereas the rabbit hunter will succeed, for a payoff of 1. The game has two equilibria, (stag, stag) and (rabbit, rabbit), because a player's optimal strategy depends on their expectation on what the other player will do. If one hunter trusts that the other will hunt the stag, they should hunt the stag; however if they think the other will hunt the rabbit, they too will hunt the rabbit. This game is used as an analogy for social cooperation, since much of the benefit that people gain in society depends upon people cooperating and implicitly trusting one another to act in a manner corresponding with cooperation.

Driving on a road against an oncoming car, and having to choose either to swerve on the left or to swerve on the right of the road, is also a coordination game. For example, with payoffs 10 meaning no crash and 0 meaning a crash, the coordination game can be defined with the following payoff matrix:

The driving game
Player 1 strategy Player 2 strategy
Drive on the left Drive on the right
Drive on the left
10
10
0
0
Drive on the right
0
0
10
10

In this case there are two pure-strategy Nash equilibria, when both choose to either drive on the left or on the right. If we admit mixed strategies (where a pure strategy is chosen at random, subject to some fixed probability), then there are three Nash equilibria for the same case: two we have seen from the pure-strategy form, where the probabilities are (0%, 100%) for player one, (0%, 100%) for player two; and (100%, 0%) for player one, (100%, 0%) for player two respectively. We add another where the probabilities for each player are (50%, 50%).

Network traffic

[edit]
Sample network graph. Values on edges are the travel time experienced by a "car" traveling down that edge. is the number of cars traveling via that edge.

An application of Nash equilibria is in determining the expected flow of traffic in a network. Consider the graph on the right. If we assume that there are "cars" traveling from A to D, what is the expected distribution of traffic in the network?

This situation can be modeled as a "game", where every traveler has a choice of 3 strategies and where each strategy is a route from A to D (one of ABD, ABCD, or ACD). The "payoff" of each strategy is the travel time of each route. In the graph on the right, a car travelling via ABD experiences travel time of , where is the number of cars traveling on edge AB. Thus, payoffs for any given strategy depend on the choices of the other players, as is usual. However, the goal, in this case, is to minimize travel time, not maximize it. Equilibrium will occur when the time on all paths is exactly the same. When that happens, no single driver has any incentive to switch routes, since it can only add to their travel time. For the graph on the right, if, for example, 100 cars are travelling from A to D, then equilibrium will occur when 25 drivers travel via ABD, 50 via ABCD, and 25 via ACD. Every driver now has a total travel time of 3.75 (to see this, a total of 75 cars take the AB edge, and likewise, 75 cars take the CD edge).

Notice that this distribution is not, actually, socially optimal. If the 100 cars agreed that 50 travel via ABD and the other 50 through ACD, then travel time for any single car would actually be 3.5, which is less than 3.75. This is also the Nash equilibrium if the path between B and C is removed, which means that adding another possible route can decrease the efficiency of the system, a phenomenon known as Braess's paradox.

Competition game

[edit]
A competition game
Player 1
strategy
Player 2 strategy
Choose "0" Choose "1" Choose "2" Choose "3"
Choose "0" 0, 0 2, ?2 2, ?2 2, ?2
Choose "1" ?2, 2 1, 1 3, ?1 3, ?1
Choose "2" ?2, 2 ?1, 3 2, 2 4, 0
Choose "3" ?2, 2 ?1, 3 0, 4 3, 3

This can be illustrated by a two-player game in which both players simultaneously choose an integer from 0 to 3 and they both win the smaller of the two numbers in points. In addition, if one player chooses a larger number than the other, then they have to give up two points to the other.

This game has a unique pure-strategy Nash equilibrium: both players choosing 0 (highlighted in light red). Any other strategy can be improved by a player switching their number to one less than that of the other player. In the adjacent table, if the game begins at the green square, it is in player 1's interest to move to the purple square and it is in player 2's interest to move to the blue square. Although it would not fit the definition of a competition game, if the game is modified so that the two players win the named amount if they both choose the same number, and otherwise win nothing, then there are 4 Nash equilibria: (0,0), (1,1), (2,2), and (3,3).

Nash equilibria in a payoff matrix

[edit]

There is an easy numerical way to identify Nash equilibria on a payoff matrix. It is especially helpful in two-person games where players have more than two strategies. In this case formal analysis may become too long. This rule does not apply to the case where mixed (stochastic) strategies are of interest. The rule goes as follows: if the first payoff number, in the payoff pair of the cell, is the maximum of the column of the cell and if the second number is the maximum of the row of the cell – then the cell represents a Nash equilibrium.

We can apply this rule to a 3×3 matrix:

A payoff matrix – Nash equilibria in bold
Player 1
strategy
Player 2 strategy
Option A Option B Option C
Option A 0, 0 25, 40 5, 10
Option B 40, 25 0, 0 5, 15
Option C 10, 5 15, 5 10, 10

Using the rule, we can very quickly (much faster than with formal analysis) see that the Nash equilibria cells are (B,A), (A,B), and (C,C). Indeed, for cell (B,A), 40 is the maximum of the first column and 25 is the maximum of the second row. For (A,B), 25 is the maximum of the second column and 40 is the maximum of the first row; the same applies for cell (C,C). For other cells, either one or both of the duplet members are not the maximum of the corresponding rows and columns.

This said, the actual mechanics of finding equilibrium cells is obvious: find the maximum of a column and check if the second member of the pair is the maximum of the row. If these conditions are met, the cell represents a Nash equilibrium. Check all columns this way to find all NE cells. An N×N matrix may have between 0 and N×N pure-strategy Nash equilibria.

Stability

[edit]

The concept of stability, useful in the analysis of many kinds of equilibria, can also be applied to Nash equilibria.

A Nash equilibrium for a mixed-strategy game is stable if a small change (specifically, an infinitesimal change) in probabilities for one player leads to a situation where two conditions hold:

  1. the player who did not change has no better strategy in the new circumstance
  2. the player who did change is now playing with a strictly worse strategy.

If these cases are both met, then a player with the small change in their mixed strategy will return immediately to the Nash equilibrium. The equilibrium is said to be stable. If condition one does not hold then the equilibrium is unstable. If only condition one holds then there are likely to be an infinite number of optimal strategies for the player who changed.

In the "driving game" example above there are both stable and unstable equilibria. The equilibria involving mixed strategies with 100% probabilities are stable. If either player changes their probabilities slightly, they will be both at a disadvantage, and their opponent will have no reason to change their strategy in turn. The (50%,50%) equilibrium is unstable. If either player changes their probabilities (which would neither benefit or damage the expectation of the player who did the change, if the other player's mixed strategy is still (50%,50%)), then the other player immediately has a better strategy at either (0%, 100%) or (100%, 0%).

Stability is crucial in practical applications of Nash equilibria, since the mixed strategy of each player is not perfectly known, but has to be inferred from statistical distribution of their actions in the game. In this case unstable equilibria are very unlikely to arise in practice, since any minute change in the proportions of each strategy seen will lead to a change in strategy and the breakdown of the equilibrium.

Finally in the eighties, building with great depth on such ideas Mertens-stable equilibria were introduced as a solution concept. Mertens stable equilibria satisfy both forward induction and backward induction. In a game theory context stable equilibria now usually refer to Mertens stable equilibria.[citation needed]

Occurrence

[edit]

If a game has a unique Nash equilibrium and is played among players under certain conditions, then the NE strategy set will be adopted. Sufficient conditions to guarantee that the Nash equilibrium is played are:

  1. The players all will do their utmost to maximize their expected payoff as described by the game.
  2. The players are flawless in execution.
  3. The players have sufficient intelligence to deduce the solution.
  4. The players know the planned equilibrium strategy of all of the other players.
  5. The players believe that a deviation in their own strategy will not cause deviations by any other players.
  6. There is common knowledge that all players meet these conditions, including this one. So, not only must each player know the other players meet the conditions, but also they must know that they all know that they meet them, and know that they know that they know that they meet them, and so on.

Where the conditions are not met

[edit]

Examples of game theory problems in which these conditions are not met:

  1. The first condition is not met if the game does not correctly describe the quantities a player wishes to maximize. In this case there is no particular reason for that player to adopt an equilibrium strategy. For instance, the prisoner's dilemma is not a dilemma if either player is happy to be jailed indefinitely.
  2. Intentional or accidental imperfection in execution. For example, a computer capable of flawless logical play facing a second flawless computer will result in equilibrium. Introduction of imperfection will lead to its disruption either through loss to the player who makes the mistake, or through negation of the common knowledge criterion leading to possible victory for the player. (An example would be a player suddenly putting the car into reverse in the game of chicken, ensuring a no-loss no-win scenario).
  3. In many cases, the third condition is not met because, even though the equilibrium must exist, it is unknown due to the complexity of the game, for instance in Chinese chess.[24] Or, if known, it may not be known to all players, as when playing tic-tac-toe with a small child who desperately wants to win (meeting the other criteria).
  4. The criterion of common knowledge may not be met even if all players do, in fact, meet all the other criteria. Players wrongly distrusting each other's rationality may adopt counter-strategies to expected irrational play on their opponents’ behalf. This is a major consideration in "chicken" or an arms race, for example.

Where the conditions are met

[edit]

In his Ph.D. dissertation, John Nash proposed two interpretations of his equilibrium concept, with the objective of showing how equilibrium points can be connected with observable phenomenon.

(...) One interpretation is rationalistic: if we assume that players are rational, know the full structure of the game, the game is played just once, and there is just one Nash equilibrium, then players will play according to that equilibrium.

This idea was formalized by R. Aumann and A. Brandenburger, 1995, Epistemic Conditions for Nash Equilibrium, Econometrica, 63, 1161-1180 who interpreted each player's mixed strategy as a conjecture about the behaviour of other players and have shown that if the game and the rationality of players is mutually known and these conjectures are commonly known, then the conjectures must be a Nash equilibrium (a common prior assumption is needed for this result in general, but not in the case of two players. In this case, the conjectures need only be mutually known).

A second interpretation, that Nash referred to by the mass action interpretation, is less demanding on players:

[i]t is unnecessary to assume that the participants have full knowledge of the total structure of the game, or the ability and inclination to go through any complex reasoning processes. What is assumed is that there is a population of participants for each position in the game, which will be played throughout time by participants drawn at random from the different populations. If there is a stable average frequency with which each pure strategy is employed by the average member of the appropriate population, then this stable average frequency constitutes a mixed strategy Nash equilibrium.

For a formal result along these lines, see Kuhn, H. and et al., 1996, "The Work of John Nash in Game Theory", Journal of Economic Theory, 69, 153–185.

Due to the limited conditions in which NE can actually be observed, they are rarely treated as a guide to day-to-day behaviour, or observed in practice in human negotiations. However, as a theoretical concept in economics and evolutionary biology, the NE has explanatory power. The payoff in economics is utility (or sometimes money), and in evolutionary biology is gene transmission; both are the fundamental bottom line of survival. Researchers who apply games theory in these fields claim that strategies failing to maximize these for whatever reason will be competed out of the market or environment, which are ascribed the ability to test all strategies. This conclusion is drawn from the "stability" theory above. In these situations the assumption that the strategy observed is actually a NE has often been borne out by research.[25]

NE and non-credible threats

[edit]
Extensive and normal-form illustrations that show the difference between SPNE and other NE. The blue equilibrium is not subgame perfect because player two makes a non-credible threat at 2(2) to be unkind (U).

The Nash equilibrium is a superset of the subgame perfect Nash equilibrium. The subgame perfect equilibrium in addition to the Nash equilibrium requires that the strategy also is a Nash equilibrium in every subgame of that game. This eliminates all non-credible threats, that is, strategies that contain non-rational moves in order to make the counter-player change their strategy.

The image to the right shows a simple sequential game that illustrates the issue with subgame imperfect Nash equilibria. In this game player one chooses left(L) or right(R), which is followed by player two being called upon to be kind (K) or unkind (U) to player one, However, player two only stands to gain from being unkind if player one goes left. If player one goes right the rational player two would de facto be kind to her/him in that subgame. However, The non-credible threat of being unkind at 2(2) is still part of the blue (L, (U,U)) Nash equilibrium. Therefore, if rational behavior can be expected by both parties the subgame perfect Nash equilibrium may be a more meaningful solution concept when such dynamic inconsistencies arise.

Proof of existence

[edit]

Proof using the Kakutani fixed-point theorem

[edit]

Nash's original proof (in his thesis) used Brouwer's fixed-point theorem (e.g., see below for a variant). This section presents a simpler proof via the Kakutani fixed-point theorem, following Nash's 1950 paper (he credits David Gale with the observation that such a simplification is possible).

To prove the existence of a Nash equilibrium, let be the best response of player i to the strategies of all other players.

Here, , where , is a mixed-strategy profile in the set of all mixed strategies and is the payoff function for player i. Define a set-valued function such that . The existence of a Nash equilibrium is equivalent to having a fixed point.

Kakutani's fixed point theorem guarantees the existence of a fixed point if the following four conditions are satisfied.

  1. is compact, convex, and nonempty.
  2. is nonempty.
  3. is upper hemicontinuous
  4. is convex.

Condition 1. is satisfied from the fact that is a simplex and thus compact. Convexity follows from players' ability to mix strategies. is nonempty as long as players have strategies.

Condition 2. and 3. are satisfied by way of Berge's maximum theorem. Because is continuous and compact, is non-empty and upper hemicontinuous.

Condition 4. is satisfied as a result of mixed strategies. Suppose , then . i.e. if two strategies maximize payoffs, then a mix between the two strategies will yield the same payoff.

Therefore, there exists a fixed point in and a Nash equilibrium.[26]

When Nash made this point to John von Neumann in 1949, von Neumann famously dismissed it with the words, "That's trivial, you know. That's just a fixed-point theorem." (See Nasar, 1998, p. 94.)

Alternate proof using the Brouwer fixed-point theorem

[edit]

We have a game where is the number of players and is the action set for the players. All of the action sets are finite. Let denote the set of mixed strategies for the players. The finiteness of the s ensures the compactness of .

We can now define the gain functions. For a mixed strategy , we let the gain for player on action be

The gain function represents the benefit a player gets by unilaterally changing their strategy. We now define where for . We see that

Next we define:

It is easy to see that each is a valid mixed strategy in . It is also easy to check that each is a continuous function of , and hence is a continuous function. As the cross product of a finite number of compact convex sets, is also compact and convex. Applying the Brouwer fixed point theorem to and we conclude that has a fixed point in , call it . We claim that is a Nash equilibrium in . For this purpose, it suffices to show that

This simply states that each player gains no benefit by unilaterally changing their strategy, which is exactly the necessary condition for a Nash equilibrium.

Now assume that the gains are not all zero. Therefore, and such that . Then

So let

Also we shall denote as the gain vector indexed by actions in . Since is the fixed point we have:

Since we have that is some positive scaling of the vector . Now we claim that

To see this, first if then this is true by definition of the gain function. Now assume that . By our previous statements we have that

and so the left term is zero, giving us that the entire expression is as needed.

So we finally have that

where the last inequality follows since is a non-zero vector. But this is a clear contradiction, so all the gains must indeed be zero. Therefore, is a Nash equilibrium for as needed.

Computing Nash equilibria

[edit]

If a player A has a dominant strategy then there exists a Nash equilibrium in which A plays . In the case of two players A and B, there exists a Nash equilibrium in which A plays and B plays a best response to . If is a strictly dominant strategy, A plays in all Nash equilibria. If both A and B have strictly dominant strategies, there exists a unique Nash equilibrium in which each plays their strictly dominant strategy.

In games with mixed-strategy Nash equilibria, the probability of a player choosing any particular (so pure) strategy can be computed by assigning a variable to each strategy that represents a fixed probability for choosing that strategy. In order for a player to be willing to randomize, their expected payoff for each (pure) strategy should be the same. In addition, the sum of the probabilities for each strategy of a particular player should be 1. This creates a system of equations from which the probabilities of choosing each strategy can be derived.[16]

Examples

[edit]
Matching pennies
Player A
plays
Player B plays
H T
H ?1, +1 +1, ?1
T +1, ?1 ?1, +1

In the matching pennies game, player A loses a point to B if A and B play the same strategy and wins a point from B if they play different strategies. To compute the mixed-strategy Nash equilibrium, assign A the probability of playing H and of playing T, and assign B the probability of playing H and of playing T.

Thus, a mixed-strategy Nash equilibrium in this game is for each player to randomly choose H or T with and .

Oddness of equilibrium points

[edit]
Free money game
Player A
votes
Player B votes
Yes No
Yes 1, 1 0, 0
No 0, 0 0, 0

In 1971, Robert Wilson came up with the "oddness theorem",[27] which says that "almost all" finite games have a finite and odd number of Nash equilibria. In 1993, Harsanyi published an alternative proof of the result.[28] "Almost all" here means that any game with an infinite or even number of equilibria is very special in the sense that if its payoffs were even slightly randomly perturbed, with probability one it would have an odd number of equilibria instead.

The prisoner's dilemma, for example, has one equilibrium, while the battle of the sexes has three—two pure and one mixed, and this remains true even if the payoffs change slightly. The free money game is an example of a "special" game with an even number of equilibria. In it, two players have to both vote "yes" rather than "no" to get a reward and the votes are simultaneous. There are two pure-strategy Nash equilibria, (yes, yes) and (no, no), and no mixed strategy equilibria, because the strategy "yes" weakly dominates "no". "Yes" is as good as "no" regardless of the other player's action, but if there is any chance the other player chooses "yes" then "yes" is the best reply. Under a small random perturbation of the payoffs, however, the probability that any two payoffs would remain tied, whether at 0 or some other number, is vanishingly small, and the game would have either one or three equilibria instead.

See also

[edit]

Notes

[edit]
  1. ^ This term is dispreferred, as it can also mean the opposite of a "strong" Nash equilibrium (i.e. a Nash equilibrium that is vulnerable to manipulation by groups).

References

[edit]
  1. ^ Osborne, Martin J.; Rubinstein, Ariel (12 Jul 1994). A Course in Game Theory. Cambridge, MA: MIT. p. 14. ISBN 9780262150415.
  2. ^ Kreps D.M. (1987) "Nash Equilibrium." In: Palgrave Macmillan (eds) The New Palgrave Dictionary of Economics. Palgrave Macmillan, London.
  3. ^ Nash, John F. (1950). "Equilibrium points in n-person games". PNAS. 36 (1): 48–49. Bibcode:1950PNAS...36...48N. doi:10.1073/pnas.36.1.48. PMC 1063129. PMID 16588946.
  4. ^ Schelling, Thomas, The Strategy of Conflict, copyright 1960, 1980, Harvard University Press, ISBN 0-674-84031-3.
  5. ^ De Fraja, G.; Oliveira, T.; Zanchi, L. (2010). "Must Try Harder: Evaluating the Role of Effort in Educational Attainment". Review of Economics and Statistics. 92 (3): 577. doi:10.1162/REST_a_00013. hdl:2108/55644. S2CID 57072280.
  6. ^ Ward, H. (1996). "Game Theory and the Politics of Global Warming: The State of Play and Beyond". Political Studies. 44 (5): 850–871. doi:10.1111/j.1467-9248.1996.tb00338.x. S2CID 143728467.,
  7. ^ Thorpe, Robert B.; Jennings, Simon; Dolder, Paul J. (2017). "Risks and benefits of catching pretty good yield in multispecies mixed fisheries". ICES Journal of Marine Science. 74 (8): 2097–2106. doi:10.1093/icesjms/fsx062.,
  8. ^ "Marketing Lessons from Dr. Nash - Andrew Frank". 2025-08-14. Retrieved 2025-08-14.
  9. ^ Chiappori, P. -A.; Levitt, S.; Groseclose, T. (2002). "Testing Mixed-Strategy Equilibria when Players Are Heterogeneous: The Case of Penalty Kicks in Soccer" (PDF). American Economic Review. 92 (4): 1138. CiteSeerX 10.1.1.178.1646. doi:10.1257/00028280260344678.
  10. ^ Muchen Sun; Francesca Baldini; Katie Hughes; Peter Trautman; Todd Murphey (2024). "Mixed-Strategy Nash Equilibrium for Crowd Navigation". arXiv:2403.01537 [cs.RO].
  11. ^ Djehiche, B.; Tcheukam, A.; Tembine, H. (2017). "A Mean-Field Game of Evacuation in Multilevel Building". IEEE Transactions on Automatic Control. 62 (10): 5154–5169. doi:10.1109/TAC.2017.2679487. ISSN 0018-9286. S2CID 21850096.
  12. ^ Djehiche, Boualem; Tcheukam, Alain; Tembine, Hamidou (2025-08-14). "Mean-Field-Type Games in Engineering". AIMS Electronics and Electrical Engineering. 1: 18–73. arXiv:1605.03281. doi:10.3934/ElectrEng.2017.1.18. S2CID 16055840.
  13. ^ Cournot, Augustin (1897) [1838]. Researches on the Mathematical Principles of the Theory of Wealth. Translated by Bacon, Nathaniel T. New York: The Macmillan Company.
  14. ^ J. Von Neumann, O. Morgenstern, Theory of Games and Economic Behavior, copyright 1944, 1953, Princeton University Press
  15. ^ Carmona, Guilherme; Podczeck, Konrad (2009). "On the Existence of Pure Strategy Nash Equilibria in Large Games" (PDF). Journal of Economic Theory. 144 (3): 1300–1319. doi:10.1016/j.jet.2008.11.009. hdl:10362/11577. SSRN 882466. Archived from the original (PDF) on May 21, 2009.
  16. ^ a b von Ahn, Luis. "Preliminaries of Game Theory" (PDF). Science of the Web. Archived from the original (PDF) on 2025-08-14. Retrieved 2025-08-14.
  17. ^ Robert Wyttenbach. "Nash Equilibria". hoylab.cornell.edu. Archived from the original on Jun 16, 2019. Retrieved 2025-08-14.
  18. ^ a b B. D. Bernheim; B. Peleg; M. D. Whinston (1987), "Coalition-Proof Equilibria I. Concepts", Journal of Economic Theory, 42 (1): 1–12, doi:10.1016/0022-0531(87)90099-8.
  19. ^ Aumann, R. (1959). "Acceptable points in general cooperative n-person games". Contributions to the Theory of Games. Vol. IV. Princeton, N.J.: Princeton University Press. ISBN 978-1-4008-8216-8. {{cite book}}: ISBN / Date incompatibility (help)
  20. ^ D. Moreno; J. Wooders (1996), "Coalition-Proof Equilibrium" (PDF), Games and Economic Behavior, 17 (1): 80–112, doi:10.1006/game.1996.0095, hdl:10016/4408.
  21. ^ MIT OpenCourseWare. 6.254: Game Theory with Engineering Applications, Spring 2010. Lecture 6: Continuous and Discontinuous Games.
  22. ^ Rosen, J. B. (1965). "Existence and Uniqueness of Equilibrium Points for Concave N-Person Games". Econometrica. 33 (3): 520–534. doi:10.2307/1911749. hdl:2060/19650010164. ISSN 0012-9682. JSTOR 1911749.
  23. ^ Schmeidler, David (2025-08-14). "Equilibrium points of nonatomic games". Journal of Statistical Physics. 7 (4): 295–300. doi:10.1007/BF01014905. ISSN 1572-9613.
  24. ^ T. L. Turocy, B. Von Stengel, Game Theory, copyright 2001, Texas A&M University, London School of Economics, pages 141-144. Nash proved that a perfect NE exists for this type of finite extensive form game[citation needed] – it can be represented as a strategy complying with his original conditions for a game with a NE. Such games may not have unique NE, but at least one of the many equilibrium strategies would be played by hypothetical players having perfect knowledge of all 10150 game trees[citation needed].
  25. ^ J. C. Cox, M. Walker, Learning to Play Cournot Duoploy Strategies Archived 2025-08-14 at the Wayback Machine, copyright 1997, Texas A&M University, University of Arizona, pages 141-144
  26. ^ Fudenburg, Drew; Tirole, Jean (1991). Game Theory. MIT Press. ISBN 978-0-262-06141-4.
  27. ^ Wilson, Robert (2025-08-14). "Computing Equilibria of N-Person Games". SIAM Journal on Applied Mathematics. 21 (1): 80–87. doi:10.1137/0121011. ISSN 0036-1399.
  28. ^ Harsanyi, J. C. (2025-08-14). "Oddness of the Number of Equilibrium Points: A New Proof". International Journal of Game Theory. 2 (1): 235–250. doi:10.1007/BF01737572. ISSN 1432-1270. S2CID 122603890.

Bibliography

[edit]

Game theory textbooks

[edit]

Original Nash papers

[edit]

Other references

[edit]
[edit]
三个火念什么 最高的山是什么山 太阳穴疼痛是什么原因 红色尿液是什么原因 jk制服是什么意思
手爆皮是什么原因 猪心炖什么好吃又营养 颈椎病挂什么科 白细胞2加号什么意思 鸭蛋炒什么好吃
低盐饮食有利于预防什么 商标r是什么意思 中性粒细胞高说明什么 木加一笔变成什么字 皮肤起小水泡很痒是什么原因
贫血吃什么食物 拔牙后吃什么恢复快 同什么协什么 女生胸痛什么原因 鬼剃头是什么病
半套什么意思hcv8jop9ns6r.cn 鱼水之欢是什么意思hcv9jop1ns8r.cn 今天是什么甲子hcv9jop3ns5r.cn 维生素b2是什么hcv9jop8ns3r.cn 香港有什么东西值得买hcv9jop3ns3r.cn
左眼皮跳是什么预兆hcv7jop7ns1r.cn 洗衣机脱水是什么意思hcv8jop0ns4r.cn 胡萝卜不能和什么食物一起吃xinjiangjialails.com 除体内湿热最好的中成药是什么hcv8jop4ns0r.cn 雪莲菌泡牛奶有什么功效hcv8jop1ns1r.cn
低血压低是什么原因hcv8jop5ns4r.cn 医者仁心是什么意思hcv8jop6ns4r.cn 动物的尾巴有什么用处hcv9jop2ns9r.cn 薰衣草什么时候开花hcv9jop1ns4r.cn 敖是什么意思hcv8jop5ns1r.cn
病毒的繁殖方式是什么hcv7jop6ns3r.cn 红景天有什么功效hcv8jop7ns1r.cn 手指为什么会脱皮xianpinbao.com 视网膜为什么会脱落hcv8jop7ns8r.cn 十二指肠霜斑样溃疡是什么意思xinjiangjialails.com
百度