少点错误 10月27日 14:07
无限效用函数面临的数学证明与现实考量
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文探讨了无限效用函数的数学证明及其在现实决策中的局限性。Vann McGee的数学证明表明,具有无限效用函数且具备合理认知假设的代理人容易受到“荷兰式赌局”的剥削。该证明指出,在存在无限状态的世界中,无限效用函数可能导致代理人在一系列特定赌局中遭受净效用损失。文章进一步分析了这一证明所揭示的“三难困境”:代理人要么需要拥有不合理的认知,要么拥有有界的效用函数,要么采取其他“怪异”的策略。作者认为,虽然数学证明存在理论上的重要性,但在实际应用中,无限效用函数会引发诸如帕斯卡赌注、帕斯卡抢劫和圣彼得堡悖论等一系列悖论,这些悖论都指向了有界效用函数的合理性。最终,文章强调了效用函数有界性在提供“本体论清晰度”方面的价值,并认为大多数个体在构想理想状态时,其效用函数会趋于一个最大值,而非无限增长。

💡 数学证明揭示无限效用函数的脆弱性:Vann McGee的数学证明表明,在无限状态和合理认知假设下,具有无限效用函数的代理人可以通过一系列精心设计的赌局(荷兰式赌局)被剥削,导致净效用损失。这表明理论上无限效用函数并非总是最优选择,尤其是在存在可利用性的情况下。

⚖️ 现实悖论凸显有界效用函数的必要性:文章通过帕斯卡赌注、帕斯卡抢劫和圣彼得堡悖论等经典决策理论问题,说明了无限效用函数在实际情境中会产生荒谬的结论。例如,帕斯卡赌注导致所有行为的期望效用都为无穷大,而帕斯卡抢劫则使得代理人极易被勒索。这些悖论都暗示了效用函数需要有界。

🧠 效用函数有界性的直观合理性:作者认为,即使不考虑数学证明和悖论,从直观感受上,大多数人构想理想世界(乌托邦)或最差世界(反乌托邦)时,其效用函数也并非无限增长或无限下降,而是趋于一个最大值和最小值。这种“本体论清晰度”使得有界效用函数更符合个体对价值的内在认知,意味着在追求更好结果时会遇到收益递减。

🔄 决策理论中的三难困境:数学证明引出了一个三难困境:代理人要么拥有不合理的认知(例如将所有概率集中在有限结果上),要么拥有有界的效用函数,要么采取其他未知策略。作者倾向于认为,拥有有界效用函数是解决这一困境的合理途径,因为它既规避了数学证明的风险,也符合直观的价值判断。

Published on October 27, 2025 5:06 AM GMT

Previous Post

For context on how I discuss utilitarianism in this sequence, read the first post.

The Proof

There is a mathematical proof that is a compelling case for bounded utility functions, but isn't the whole story.

tl;dr: Vann McGee proves that agents with unbounded utility functions and under reasonable assumptions about their epistemics are consistently vulnerable to Dutch Books which exploit their willingness to seek out high-utility low-probability outcomes in some contexts.

Proof Outline

Consider an agent in a world with infinitely many states, and the agent believes that some infinite (not necessarily strict) subset of those states is possible (although they can have zero probability). 

If the agent has an unbounded utility function, then you can subject it to a Dutch Book using an infinite sequence of bets about truth values of the propositions (A1,A2,...,An):

Bet 1: You lose one util if A1 is true and gain 1+P(A1)P(¬A1) utils if A1 is false.

 

Bet 2: You lose 2 utils if A1 is false, and you gain 3P(¬A1)P(A1¬A2) utils if A1 is true and A2 is false; otherwise, the bet is called off.


Bet n+1: You lose n+1P(A1A2...An1¬An) utils if A1,A2,...,An1 are all true and An is false. You gain n+2P(A1A2An¬An+1) utils if A1,A2,,An1,An are all true and An+1 is false. Otherwise, the bet is called off.

Each of these bets has an expected utility of 1, making it advantageous to take them, assuming that the casino has unbounded utility to hand out. However, only finitely many of the bets will be won with any reality-measure, so this scheme leads to the agent always losing net utility if it chooses to take the whole infinite bundle of bets.

McGee goes into more detail on the mathematical nuances of this. Peter De Blanc investigates a more general and abstract angle on this problem.

Beyond the Proof

The above proof implies the trilemma:

    Agents should defy those reasonable epistemic assumptions and instead have unreasonable epistemics (by concentrating all their probability mass into finitely many outcomes in the infinite outcome space).Agents should have bounded utility functions.Agents should do some other weird thing, like whatever this post is hinting at. 

I won't be responding to that post here, and I think we can agree to not do the thing labeled "unreasonable epistemics", since 0 and 1 are not probabilities

This leaves us with option 2, but does it really make sense to have a bounded utility function? I'm going to try and come at this from several different angles, in the hopes of conveying why bounding utility makes sense as a property of coherent agents.

The Proof is Limitedly Useful

The proof assumes an infinite sequence of bets, which nobody ever has time to execute, so there's some question as to whether this conclusion holds up in real life. For that, I reference this passage from McGee's paper which gives a reason to care about this argument, albeit one that is more poetic than satisfying:

Even a simply infinite sequence of bets is something in which we mortals never have a chance to participate, so as long as our interest in decision theory is purely practical, we needn’t fret over the example, whether it’s presented to us in the static or dynamic version. There is, however, some theoretical interest in trying to devise a standard of rationality so steadfast that one could safely rely on it even if confronted with an infinite array of choices. We only have occasion to make finitely many choices, but it would be surprising if the finitude of the set of choices we make were a prerequisite for rationality.

I don't have an argument that is properly satisfying, but I do have several different suggestive intuition pumps that constitute much of the reason that I think of myself as having a bounded utility function.

Other Paradoxes of Unbounded Utility

Pascal

There is a classic situation in decision theory called Pascal's Wager:

Some people claim there is a God who will send you to heaven the land of infinite utility, if and only if you Believe in Him and perform some specific rituals. You're a good Bayesian, and you don't assign literally zero probability to this God being real. Infinite utility multiplied by a positive probability is infinity, so you should pick up the infinite expected utility on the ground and join this religion. 

This argument is generally considered around LessWrong to be Wrong, and if you haven't seen it before I encourage you to look for flaws yourself.




Done looking for flaws? Okay. Here is mine:

This argument proves far too many things to be jointly the optimal action. For every possible action, there is a logical possibility of a god that rewards that specific thing with infinite utility, and so the expected utility of every action is infinite, even the ones we class as obviously stupid, like donating all of your money to Effective Evil. This is an argument from absurdity against infinite utility functions, but not quite against unbounded ones.

Pascal's Mugging is a variant of this that classically goes as follows:

There is a Guy on the street who asks you for $5, threatening that otherwise He will step outside The Matrix and simulate 33[1] suffering humans.

This steps around the problem of all the infinities being the same, but in doing so it creates several more minor problems and one major one:

If your utility function is bounded, then you aren't exploitable in this way. 

St. Peter

The St. Petersburg paradox proposes a game:

I flip a fair coin until it comes up tails for the first time, and I note down the total number of flips (including the final tails) as n. After, I pay you $2n.

The question is, how much should you pay to enter into this game? Once again, I encourage you to work it out if you haven't seen this before.




We can calculate the expected payout as follows:

P(n=1)×$21+P(n=2)×$22+P(n=3)×$23+P(n=4)×$24+...

=12×$2+14×$4+18×$8+116×$16+...

=$1+$1+$1+$1+...

=$

Hmmmm. 

That's weird.

If we're measuring payouts in money, then this runs into the finite funds of St. Peter's Casino, as well as the fact that many agents have sublinear utility in money. If payouts are given in utility, then this runs into the same infinity problem as Pascal's Wager. Among the solutions for both the money version and the utility version is, of course, bounded utility functions.

Maxima and Minima

Outside the realm of thought experiments with mysterious Russian games and interdimensional wizards, it still makes sense to me to bound my utility function. I won't say this is quite "practical" in the sense that I use it for real decisions in my real life, but it does genuinely provide ontological clarity.

One equivalent rephrasing of "my utility function is bounded" is "my utility function has a maximum and minimum value". 

This fact isn't quite explained by my concept Sublinear Utility in Population despite that addressing utility functions asymptoting to some fixed value, as mhampton (correctly) notes in a comment (lightly reformatted and truncated):

This applies only to interpersonal aggregation, and so if you can have unboundedly high utility in one individual, your utility function is not truly bounded, right? I.e., it would get you out of Pascal's muggings of the form, "Pay me five dollars and I will create 33 happy copies of Alice" but not of the form "Pay me five dollars and I will create one copy of Alice and give her 33 utils."

When I try to think about getting 33 utils, this doesn't really seem well defined. When I try to construct an ideal utopia or anti-ideal dystopia in my head, this doesn't intuitively look like the scale of goodness keeps going up or down infinitely, it feels like any changes I make are asymptoting towards an optimum, whether it's in making there be more happy people or making one person really really happy, or any of the other things I value. 

I model that, if my utility function were actually unbounded, then the process of finding a utopia would feel like always going "hmmm, I could make this world a substantial amount better by just changing this set of details" and always getting returns in utility of similar sizes, rather than running into diminishing returns. 

To overuse an example, if my utility function was exactly the number of paper clips in the universe, then each time I would try to imagine a concrete utopia, I would always be able to add more paper clips to get a better world, without ever asymptoting or plateauing in utility, and without reaching an optimum.

Maybe some people's intuitive values are structured like this, in which case they can work things out between them and the unbounded utility paradoxes I've listed here. I don't have a good sense of how many people have introspective assessments of their values matching mine in this respect, but I'm curious to find out.

  1. ^

    very very very large number, see Knuth's up-arrow notation for the definition

  2. ^

    Not to say that I'm not interested in Very Big Numbers categorically, I'm just not inclined to care when the number of people is so much larger than the number of atoms in the universe.

  3. ^

    "But the casino must have finite funds available" and "But I have sublinear utility in money", respectively.



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

无限效用函数 决策理论 数学证明 荷兰式赌局 帕斯卡赌注 圣彼得堡悖论 有界效用函数 Unbounded Utility Decision Theory Mathematical Proof Dutch Book Pascal's Wager St. Petersburg Paradox Bounded Utility
相关文章