少点错误 10月22日 03:21
“分层乌托邦”:一种未来资源分配的可能性
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

文章探讨了“分层乌托邦”这一未来可能出现的理想状态,即平凡的价值观(如周末聚会、赢得同伴赞赏)能够获得近处的资源(时空上靠近地球),而更加超凡的价值观(如最大化整体福祉、实现至善)则获得遥远处的资源(遥远的星系、漫长的未来)。作者深入分析了这种分层乌托邦的形成机制,包括高效的资源分配、价值观的构成以及资源与价值观的兼容性,并探讨了其发生的可能性和潜在的顾虑,如早期灭绝或分层反乌托邦的出现。

🌟 **分层乌托邦的概念与构成**:文章提出“分层乌托邦”是一种理想化的未来图景,其中宇宙资源根据价值观的“平凡”与“超凡”属性进行分配。平凡价值观(如享受生活、个人成就)获得时空上靠近地球的资源,而超凡价值观(如最大化整体福祉、追求终极真理)则获得遥远星系和漫长未来的资源。这种分配模式得益于平凡价值观的“范围不敏感性”和“短期主义”,以及超凡价值观的“范围敏感性”和“长期主义”。

🚀 **分层乌托邦的形成机制**:文章认为,分层乌托邦的出现基于四个核心前提:1. **高效分配**:资源会流向最看重它的价值观;2. **价值观构成**:平凡与超凡价值观在资源分配时都具有相当的影响力;3. **资源兼容性**:平凡价值观偏好近端资源,超凡价值观偏好远端资源;4. **持久性**:一旦形成,这种分配模式会因路径依赖或难以逆转的协议而保持稳定。文章分析了市场、战争、谈判、一致外推意愿(CEV)和独裁等多种可能的分配机制。

🤔 **分层乌托托邦的现实性与挑战**:尽管作者认为分层乌托邦是可能的,但也指出了潜在的挑战。这些挑战包括:资源分配可能走向“悲观化”而非“优化化”,导致分层反乌托邦;早期灭绝事件可能阻止任何分配的发生;以及如果某种单一价值观占据主导地位,可能导致资源集中而非分层;此外,外星文明的出现或错误对齐的AI也可能改变资源分配的格局。文章还探讨了平凡价值观是否会随着时间推移而变得更具“范围敏感性”,从而改变其对资源的需求。

⏳ **平凡价值观的特质与演变**:文章详细阐述了平凡价值观的几个关键特征,如范围不敏感、短期主义、基底特定(重视物质本身)等,并与超凡价值观(范围敏感、长期主义、基底独立)形成对比。作者认为,宗教对特定地点的依恋、早期价值观锁定、指导性理想化等因素可能有助于维持平凡价值观对近端资源的偏好。然而,超级智能代理、社会比较心理、以及通过认知接触进行的价值观修改,都可能促使平凡价值观向更远端、更基底独立的方向演变,从而影响分层乌托邦的最终形态。

Published on October 21, 2025 7:09 PM GMT

Summary: "Stratified utopia" is a possible outcome where mundane values get proximal resources (near Earth in space and time) and exotic values get distal resources (distant galaxies and far futures). I discuss whether this outcome is likely or desirable.

1. Introduction

1.1. Happy Coincidence

I hold mundane values, such as partying on the weekend, the admiration of my peers, not making a fool of myself, finishing this essay, raising children, etc. I also have more exotic values, such as maximizing total wellbeing, achieving The Good, and bathing in the beatific vision. These values aren't fully-compatible, i.e. I won't be partying on the weekend in any outcome which maximizes total wellbeing[1].

But I think these values are nearly-compatible. My mundane values can get 99% of what they want (near Earth in space and time) and my exotic values can get get 99% of what they want (distant galaxies and far futures). This is a happy coincidence.

I call this arrangement "stratified utopia." In this essay, I discuss two claims:

    Stratified utopia is likely.Stratified utopia is desirable.

The simple case in favour: when different values care about different kinds of resources, it's both likely and desirable that resources are allocated to the values which care about them the most. I discuss considerations against both claims.

1.2. Values and Resources

1.2.1. Mundane vs Exotic Values

Humans hold diverse values. I'm using a thick notion of "values" that encompasses not only a utility function over world states but also ontologies, decision procedures, non-consequentialist principles, and meta-values about value idealizations.[2] You could use the synonyms "normative worldviews", "conceptions of goodness", "moral philosophies", and "policies about what matters".  Each human holds multiple values to different degrees in different contexts, and this value profile varies across individuals.

Values have many structural properties, and empirically certain properties cluster together. There is a cluster whose two poles I will call 'mundane' and 'exotic':

Mundane ValuesExotic ValuesScope-Insensitive: These values saturate at human-scale quantities.Scope-Sensitive: These values scale linearly with resources, without a saturation point. Shorttermist: These values weight the near future more heavily than the far future.Longtermist: These values weight all times equally, with no temporal discounting.Substrate-Specific: The atoms matter, not just the computational pattern.Substrate-Independent: Only the computation matters, not the physical implementation.Causal-focused: These values care about things reachable through normal physical causation.Acausal-focused: These values embrace acausal cooperation with distant civilisations we'll never meet.

1.2.2. Proximal vs Distal Resources

Cosmic resources similarly possess many structural properties, and empirically certain properties cluster together. There is a cluster whose two poles I will call 'proximal' and 'distal':

Proximal ResourcesDistal ResourcesSpatial Proximity: Near Earth in space. The solar system. Maybe the local galaxy cluster.Spatially Distant: Galaxies billions of light-years away. The vast majority of the observable universe.Temporal Proximity: The near future. The next million years.Temporally Distant: The far future, trillions of years from now, after the last stars have burned out.Substrate Proximity: Physical reality. Atoms and matter you can touch.Substrate Distant: Simulated worlds. Computation rather than physical matter.Causal Proximity: Resources we can influence through normal causation.Causally Distant: Distant civilizations which are correlated with our actions but cannot be physically reached.

It follows that mundane values want proximal resources. Exotic values want distal resources. This is the happy coincidence making these values near-compatible.

2. Stratified Utopia: A sketch

2.1. Spatial Stratification

The cosmos gets divided into shells expanding outward from Earth-2025, like layers of an onion. Each shell is a stratum. The innermost shell Earth satisfies our most mundane values. Moving outward, each shell satisfies increasingly exotic values.

The innermost shells contain a tiny proportion of the total volume, but fortunately the mundane values are scope-insensitive. The outermost shells cannot be reached for billions of years, because of the speed of light, but fortunately the exotic values are longtermist.

Spatial Stratification: 

2.2. What Each Stratum Looks Like

Level 0: The Humane World. This is our world but without atrocities such as factory farming and extreme poverty. People are still driving to school in cars, working ordinary jobs, and partying on the weekend.

Level 1: Better Governance. Everything from Level 0, plus better individual and collective rationality, improved economic, scientific, and governmental institutions, and the absence of coordination failures. Think of Yudkowsky's Dath Ilan.

Level 2: Longevity. Everything from Level 1, plus people live for centuries instead of decades. Different distributions of personality traits and cognitive capabilities, but nothing unprecedented. 

Level 3: Post-Scarcity. Everything from Level 2, plus material abundance. Think of Carlsmith's "Concrete Utopia"[3].

Level 4: Biological Transhumanism. Everything from Level 3, plus fully-customizable biological anatomy. 

Level 5: Mechanical Transhumanism. Everything from Level 4, but including both biological and mechanical bodies. Imagine a utopian version of Futurama or Star Trek, or the future Yudkowksy describes in Fun Theory.

Level 6: Virtual Worlds. Similar to Level 5, except everything runs on a computer simulation, possibly distant in the future. The minds here still have recognizably human psychology; they value science, love, achievement, meaning. But the physics is whatever works best.

Level 7: Non-Human Minds. There are agents optimizing for things we about, such as scientific discovery or freedom. But they aren't human in any meaningful sense. This is closer to what Carlsmith calls "Sublime Utopia".

Level 8: Optimized Welfare. Moral patients in states of optimized welfare. This might be vast numbers of small minds experiencing constant euphoria, i.e. "shrimp on heroin". Or it might be a single supergalactic utility monster. Whatever satisfies our mature population ethics. There might not be any moral agents here.

Level 9: Sludge. In this stratum, there might be not even be moral patients. It is whatever configuration of matter optimizes for our idealized values. This might look like classical -oniums, pure substances maximizing for recognizable moral concepts[4]. Or it might be optimizing for something incomprehensible, e.g. "enormously complex patterns of light ricocheting off intricate, nano-scale, mirror-like machines" computing IGJC #4[5].

3. Is Stratified Utopia Likely?

3.0. Four Premises

Predicting the longterm future of the cosmos is tricky, but I think something like stratified utopia is likely[6]. The argument rests on four premises:

    Efficient Allocation: When different values care about different resources, those resources get allocated to whoever values them most.Value Composition: At allocation time, mundane and exotic values will dominate with comparable influence.Resource Compatibility: At allocation time, mundane and exotic values will still want different resources (proximal vs distal).Persistence: Once established, this allocation persists, either because values remain stable over cosmic timescales, or because initial allocation creates path dependencies that resist reversal.

It follows from these premises that proximal resources satisfy mundane values and distal resources satisfy exotic values. I'll examine each premise in turn.

3.1. Efficient Allocation

The cosmos will be divided among value systems through some mechanism. Here are the main possibilities:

    Market: Values are traders with initial endowments. Mundane values would trade away distal resources (which they don't care about) in exchange for guarantees about proximal resources (which they care about). Both parties benefit from a coincidence of wants.[7]War: Values are militaries competing through force. But war is costly, and those willing to pay higher costs for specific resources will control them. Mundane values will exhaust their military capacity defending proximal resources while exotic values will exhaust their military capacity defending distal resources.Bargaining: Values negotiate with comparable bargaining power. Standard bargaining solutions produce stratification when parties value resources differently. Both Nash bargaining (maximizing the product of gains) and Kalai-Smorodinsky (equalizing proportional gains) lead to the same outcome: mundane values get proximal resources, exotic values get distal resources[8]. Each party trades what they value less for what they value more[9].Coherent Extrapolated Volition: Values are terminal considerations of a singleton implementing humanity’s CEV. If the singleton discovers that after extrapolation different values want different resources, then it will use resources for the values that care the most about it. The critical question is whether different values will still care about different resources after extrapolation, or instead whether mundane values become more scope-sensitive, longtermist, substrate-neutral, or acausally-focused[10].Autocracy: Values are preferences held by an autocrat with absolute power. The autocrat holds both mundane and exotic values with similar weighting. When they maximize their preferences, they will use resources according to their most relevant preference.

I expect resources will be allocated through some mixture of these mechanisms and others. Stratification emerges across this range because these mechanisms share a common feature: when different values care about different resources, those resources get allocated to whoever values them most.

Some considerations against this premise:

    Pessimization: The allocation might pessimize for the constituent values, rather than optimizing them, producing stratified dystopia (see Appendix B). A stratified dystopia has the same spatial structure, with inner strata pessimizing mundane values and outer strata pessimizing exotic values. See Richard Ngo's survey of different pessimization mechanisms.Early Extinction: The allocation never happens because there are no surviving values. Pandemic, nuclear war, simulation shutdown kills everyone before we reach the stars.

3.2. Value Composition

At allocation time, mundane and exotic values will dominate with comparable weighting.

Several factors push toward this premise:

    Population diversity: The population contains humans who hold exotic values (effective altruists, longtermists, utilitarians) and humans who hold mundane values. In CEV, this ensures both value types receive consideration.Exotic values have disproportionate influence: Humans with exotic values are overrepresented among those with wealth, political power, and influence over AI training and deployment. In market allocation, bargaining, or war, this gives them resources, leverage, or strategic advantage.Mixed values within individuals: Most humans hold both exotic and mundane values simultaneously. Hence, if resources are allocated according to an autocrat's preferences, both value types will be weighted in the allocation.

Several factors push against this premise:

    Single Value Dominance: If one value dominates the others, then none of these allocation mechanism produces stratification. By dominance, I mean initial equity (market), strategic advantage (war), bargaining power (negotiation), consideration (CEV), or preferential weighting (autocracy). All the resources will be used by the value which dominates the allocation process.Paperclippers: We might build misaligned AIs with alien exotic values, e.g. maximizing paperclips. This would effectively drive up the price of the stars; literally in market allocation, figuratively in other allocation mechanisms.Aliens: If alien civilizations exists and becomes grabby (expanding rapidly through space), they would influence the resource allocation. Hanson's grabby aliens model suggests that humanity would encounter aliens after 200 million to 2 billion years, and we would control between 100,000 and 30 million galaxies (approximately 0.005% to 1.5% of the observable universe's 2 trillion galaxies).[11] This could limit available resources to below the saturation point of our mundane values.

3.2. Resource Compatibility

Will mundane values remain proximal-focused? Several factors push toward yes:

    Some values inherently can't be scope-sensitive. Religious people care about sacred sites on Earth, e.g. Jerusalem, Mecca, the Ganges. These commitments are structurally location-specific. You can't satisfy them with simulated versions or equivalent sites elsewhere.Early value lock-in. Values might get “locked-in” early, before they have time to change. Mundane might deliberately preserve current values, recognizing that if they drift toward scope-sensitivity they’d be competing for resources we don't currently care about.Guided idealization. Even without explicit lock-in, values might undergo idealization in ways that preserve the mundane/exotic distinction. Recall that "values" here includes higher-order preferences about value idealization. My mundane values are less interested in extreme idealizations, e.g. my protect-my-family values would resist any process of reflection that disrupted that value.[12] But my exotic values would be more willing to undergo such processes.Early allocation. The resource allocation might occur very early (perhaps as soon as we have self-replicating probes or achieve basic coordination) before mundane values have a chance to modify into distal-seeking.Parallel evolution. Even if mundane values modify to become more scope-sensitive, exotic values might modify at the same rate, maintaining a constant gap. Both value systems might scale up proportionally while preserving their relative difference.

Several factors push toward no:

    Superintelligent agents: Mundane values might give open-ended instructions to AIs such as "make me as much money as possible", creating distal-seeking proxies of mundane values, even though the original mundane values were scope-insensitive.Keeping-Up-with-the-Joneses: Some mundane values are comparative, e.g. status, relative wealth, competitive dominance. These care about ranking, not absolute levels. If proximal resources are fixed and I want to outrank you, I'm incentivized to grab distal resources. However, two reasons this might not threaten resource compatibility:
      First, comparative mundane values are mostly about price, not stars. Distal resources are cheap per unit, whereas proximal resources command high prices per unit due to scarcity. As an analogy, people today do not signal wealth by maximizing the total mass of their assets, but the total price.Second, comparative values competing for resources is a negative-sum game: both parties expend resources without changing relative rankings. Therefore, many allocation mechanisms would prevent this, e.g. CEV, autocracy, and negotiation. Only uncoordinated markets or war might allow it, and even in markets parties have incentives to coordinate against waste.
    Value modification through epistemic engagement. Mundane values might engage with superintelligence about epistemic matters, and this might modify their values. Their ontology might shift to fundamental physics, where there's no privileged "here" and "now" and macroscopic objects (like sacred sites) and persons have flimsier identity conditions. This makes distal-seeking more natural. Or they interact extensively with AIs, learn to care about them, and become more substrate-neutral about moral patients. Or they learn they're in a simulation and become more substrate-neutral.Selection effects. If the resource allocation occurs quite late, you might have value modification purely from selection effects. Robin Hanson argues that competition will eventually select for units capable of taking longer views, and these units will dominate. If competition continues for long enough, eventually our world should contain units that do care about the distant future and are capable of planning effectively over long timescales. These units would have enough competitive advantage to dominate, potentially eliminating scope-insensitive values before the allocation[13].Mundane Values are Proximal-Seeking: Perhaps mundane values don't actually have the properties I claim, i.e. scope-insensitivity, shorttermism, substrate-specificity. Introspectively, this seems false: when I imagine learning the universe is 10x larger, the part of me that values partying on weekends doesn't leap for joy. When I imagine it's 10% the size, that part doesn't weep. It's genuinely indifferent to total cosmic resources, suggesting scope-insensitivity. But this introspection might be unreliable. I'm thinking about "the universe" in far-mode (abstract and distant), but the mundane values operate in near-mode (concrete and immediate).[14] If they were actually engaging with concrete decisions about distal resources rather than abstract thought experiments, they might behave differently than introspection suggests.

3.3. Persistence

Several factors push toward persistence:

    Longterm contracts and commitments. The allocation could be locked in through binding agreements, constitutional structures, or coordination mechanisms that are difficult to reverse. This is especially likely if the resources are controlled by an AI that can make credible commitments.Defense-dominated conflict. Once someone controls resources, it might be much easier to defend them than to capture them. Mundane values controlling Earth can fortify it against conversion, similarly with exotic values controlling distance galaxies.[15]

Several factors push against persistence:

    Contracts unenforceable. Agreements made today might not bind entities in the far future. Once a value changes how much it cares about a resource, it will attempt to capture it.Offense-dominated conflict. Defense might not be easier than offense. Perhaps exotic values could conquer proximal resources cheaply, if they began to care about them. Similarly, perhaps mundane values could conquer distal resources.

4. Is Stratified Utopia Desirable?

Here are some moral intuitions that endorse stratified utopia:

    The 99/1 Split is Fair: For the exotic values, getting 99% of the cosmos is 99% as good as getting 100%, even though they'd prefer to convert Earth into hedonium too. For mundane values, getting 1% of the cosmos is 99% as good as getting 100%, even though they'd prefer to convert the distant galaxies into… I’m not even sure what. From a normative perspective, this makes stratified utopia seem like a reasonable compromise. Neither side is being asked to make an unreasonable sacrifice.The Nothing-To-Mourn Principle: When we imagine the best future, no one should find that they mourn the loss of something. If we converted the proximal resources with exotic values, then I would have something to mourn. Stratified utopia satisfies the Nothing-To-Mourn Principle because no values mourn any loss.[16]Moral uncertainty: How should we behave if we are uncertain about which moral theory is correct, our mundane or our exotic values? Several approaches have been proposed. We might maximize expected choiceworthiness across values,[17] implement a bargaining solution between them,[18] convene a parliament where delegates representing values vote on outcomes,[19] or create a moral marketplace where values trade resources.[20] These approaches lead to stratified utopia.

But I have some countervailing moral intuitions:

    Grabbing almost all the stars isn't nice. There's something troubling about one value system getting 99% of stars and calling this a fair division. That said, objective theories of value, which equate value with easily measured quantities (e.g. gold, land, or labour), have worse track records than subjective theories, both descriptively and normatively. I suspect a "stellar theory of value" would fare no better.Idealized mundane values might seek distal resources. Perhaps under idealization, mundane values would become more scope-sensitive, more longtermist, more substrate-neutral. If so, then stratified utopia rejects CEV as the allocation mechanism. It satisfies the mundane values in their unidealized form, but not in their idealized form.Exotic values might have objectively greater range. Perhaps exotic values are genuinely more important, i.e. they can realize more value, produce more good, create more that matters. Even if mundane and exotic values have equal weighting initially, exotic values might deserve more resources because they can do more with them. A utilitarian might argue that we should maximize total wellbeing, which means giving nearly everything to the value system that scales.Simple approaches to moral uncertainty don't produce stratification: Two straightforward approaches to moral uncertainty fail to generate stratified utopia. Under "My Favorite Theory", you choose the moral theory you're most confident in and follow its prescriptions exclusively. Under "My Favorite Option", you choose the option most likely to be right across all theories you give credence to.[21] But, if you have 51% credence in exotic values and 49% credence in mundane values, then both approaches would allocate all resources to the exotic values.

Appendices

Appendix A: Spatial vs Temporal Stratification

If we want to split the cosmos between mundane and exotic values, we have two basic options. We could stratify temporally, saying the first era belongs to mundane values and later eras belong to exotic values. Or we could stratify spatially, saying the inner regions belong to mundane values and the outer regions belong to exotic values.

I think that spatial stratification is better than temporal stratification. Under temporal stratification, the first million years of the cosmos belong to mundane values, but after that deadline passes, exotic values take over everywhere, including Earth and the nearby stars.

Spatial stratification has several moral advantages over temporal stratification

    Longtermist scope-insensitive values: Under temporal stratification, your children's children get to live flourishing lives, enjoying proximal resources. But their descendants, born after the million-year cutoff, get converted into hedonium or utility substrate or whatever configuration the exotic values prefer. And every sacred site you care about eventually gets handed over to be optimized for exotic values. I think this violates something that my mundane values care about. I think that my mundane values are longtermist in some sense, even though they aren’t scope-sensitive.Avoids forced handover problem: Under temporal stratification, someone eventually shows up saying "the million years are over, time to convert you to hedonium"[22]. Under spatial stratification, there's no lease that expires, and the exotic values stay in their distant galaxies.Operationally cleaner: "This galaxy is mundane, that galaxy is exotic" is straightforward to implement. Spatial boundaries have clear precedent in human governance, temporal boundaries do not.

Appendix B: Stratified Dystopia

If stratification optimally satisfies mixed values, it also pessimizes them. The strata might look like:

Level 0: Ordinary suffering of people alive today

Levels 1-4: Enhanced biological torture

Levels 5-7: Simulated hells

Levels 8-9: Pure suffering substrate, maximally efficient negative utility

Appendix C: Firewalls

Stratified utopia requires firewalls, which blocks the flow of information from higher strata to lower ones. This has some advantages:

    Genre Preservation: Each stratum has its own genre, its own aesthetic. Inner strata don't want to live in a sci-fi world---they want mundane life, recognizable human experience. Without firewalls, the genre of lower strata gets corrupted by the strangeness of higher strata.Meaning Preservation: Firewalls preserve the ability to find meaning in mundane things. Without firewalls, it would be psychologically difficult to care about ordinary human concerns, as the scale of cosmic optimization would make ordinary concerns feel trivial.Protection Against Exotic Interference: Firewalls prevent inhabitants or systems in outer strata from influencing inner strata. Without firewalls, entities optimizing for exotic values in distant galaxies could interfere with inner strata, either intentionally or accidentally. They might offer trades, present arguments, or simply broadcast their existence in ways that shift values. Firewalls prevent this kind of interference.

Appendix D: Death and Uplifting

Lower strata inhabitants don't know about upper strata because firewalls prevent that knowledge. They lead lives that are, by most welfare standards, worse than lives in higher strata. This creates a moral problem: we sacrifice individual welfare for non-welfarist values like tradition and freedom. This seems unjust to those in inner strata.

To mitigate this injustice, when people die in Levels 0-6, they can be uplifted into higher strata if they would endorse this upon reflection. Someone whose worldview includes "death should be final" gets their wish: no uplifting. But everyone else would be uplifted, either physically (to Levels 1-5 which are substrate-specific utopias) or uploaded (to Levels 6-7 which are substate-neutral). Alternatively, we could uplift someone at multiple points in their life, forking them into copies: one continues in the inner stratum, another moves to a higher stratum.

  1. ^

     Utilitarianism does not love you, nor does it hate you, but you’re made of atoms that it can use for something else. In particular: hedonium (that is: optimally-efficient pleasure, often imagined as running on some optimally-efficient computational substrate). —Being nicer than Clippy (Joe Carlsmith, Jun 2021)

  2. ^

    Second, the idealising procedure itself, if subjective, introduces its own set of free parameters. How does an individual or group decide to resolve internal incoherencies in their preferences, if they even choose to prioritize consistency at all? How much weight is given to initial intuitions versus theoretical virtues like simplicity or explanatory power? Which arguments are deemed persuasive during reflection? How far from one's initial pre-reflective preferences is one willing to allow the idealization process to take them? — Better Futures (William MacAskill, August 2025)

    For a defence of the subjectivity of idealization procedure, see On the limits of idealized values (Joe Carlsmith, Jun 2021).

  3. ^

     Glancing at various Wikipedias, my sense is that literary depictions of Utopia often involve humans in some slightly-altered political and material arrangement: maybe holding property in common, maybe with especially liberated sexual practices, etc. And when we imagine our own personal Utopias, it can be easy to imagine something like our current lives, but with none of the problems, more of the best bits, a general overlay of happiness and good-will, and some favored aesthetic — technological shiny-ness, pastoralness, punk rock, etc — in the background. — Actually possible: thoughts on Utopia (Joe Carlsmith, Jan 2021)

  4. ^

     Classical -oniums include: alethonium (most truthful), areteonium (most virtuous), axionium (most valuable), dikaionium (most just), doxonium (most glorious), dureonium (most enduring), dynamonium (most powerful), eirenium (most peaceful), eleutheronium (most free), empathonium (most compassionate), eudaimonium (most flourishing), harmonium (most harmonious), hedonium (most pleasurable), holonium (most complete), kalionium (most beautiful), magnanimium (most generous), philonium (most loving), pneumonium (most spiritual), praxonium (most righteous), psychonium (most mindful), sophonium (most wise), teleonium (most purposeful), timonium (most honourable).

  5. ^

    Suppose, for example, that a candidate galaxy Joe---a version of myself created by giving original me 'full information' via some procedure involving significant cognitive enhancement---shows me his ideal world. It is filled with enormously complex patterns of light ricocheting off of intricate, nano-scale, mirror-like machines that appear to be in some strange sense 'flowing.' These, he tells me, are computing something he calls [incomprehensible galaxy Joe concept (IGJC) #4], in a format known as [IGJC #5], undergirded and 'hedged' via [IGJC #6]. He acknowledges that he can't explain the appeal of this to me in my current state.

    'I guess you could say it's kind of like happiness,' he says, warily. He mentions an analogy with abstract jazz.

    'Is it conscious?' I ask.

    'Um, I think the closest short answer is no,' he says.

    Suppose I can create either this galaxy Joe's favorite world, or a world of happy puppies frolicking in the grass. The puppies, from my perspective, are a pretty safe bet: I myself can see the appeal. Expected value calculations under moral uncertainty aside, suppose I start to feel drawn towards the puppies. Galaxy Joe tells me with grave seriousness: 'Creating those puppies instead of IGJC #4 would be a mistake of truly ridiculous severity.' I hesitate. Is he right, relative to me?

    On the limits of idealized values (Joe Carlsmith, Jun 2021).

  6. ^

    My aim is this essay is not to offer quantitative probabilities, but I will give some here as an invitation for pushback: Efficient Allocation (85%),Value Composition (65%), Resource Compatibility (35%), Persistence (40%). A naive multiplication gives 8% for Stratified Utopia, which seems reasonable.

  7. ^

    In the future, there could be potential for enormous gains from trade and compromise between groups with different moral views. Suppose, for example, that most in society have fairly commonsense ethical views, such that common-sense utopia (from the last essay) achieves most possible value, whereas a smaller group endorses total utilitarianism. If so, then an arrangement where the first group turns the Milky Way into a common-sense utopia, and the second group occupies all the other accessible galaxies and turns them into a total utilitarian utopia, would be one in which both groups get a future that is very close to as good as it could possibly be. Potentially, society could get to this arrangement even if one group was a much smaller minority than the other, via some sort of trade. Through trade, both groups get a future that is very close to as good as it could possibly be, by their lights. — Better Futures (William MacAskill, August 2025)

  8. ^

     Consider mundane utility U_m = 100(1-x) + (1-y) and exotic utility U_e = x + 100y, where x and y are the proportions of proximal and distal resources allocated to exotic values. Starting from equal division (0.5, 0.5) as the disagreement point, both Nash and K-S select the corner solution (0,1) where mundane gets all proximal and exotic gets all distal. For Nash: this maximizes the product of gains since both parties get resources they value 100 times more than what they give up. For K-S: this is the only Pareto-efficient point providing positive equal gains (each party gets utility 100, gaining 49.5 from disagreement). The anti-stratified corner (1,0) leaves both worse off than disagreement.

  9. ^

    This is similar to the Market mechanism, except the allocation doesn't involve the transfer of property rights or prices. 

  10. ^

    There is also a possibility (although it seems to me less likely) that my exotic values become more proximal-focused, perhaps due to mature infinite ethics undermining total utilitarianism.

  11. ^
  12. ^

    I suspect the most common attitude among people today would either be to reject the idea of reflection on the good (de dicto) as confusing or senseless, to imagine one's present views as unlikely to be moved by reflection, or to see one's idealised reflective self as an undesirably alien creature. — Section 2.3.1 Better Futures (William MacAskill, August 2025)

  13. ^

    Hanson argues that history is a competition to control the distant future, but behavior has been focused on the short term. Eventually, competition will select for entities capable of taking longer views and planning over longer timescales, and these will dominate. He calls this transition point "Long View Day." See Long Views Are Coming (Robin Hanson, November 2018)

  14. ^

    Near mode and far mode refer to different styles of thinking identified in construal level theory. Near mode is concrete, detailed, and contextual how we think about things physically, temporally, or socially close to us. Far mode is abstract, schematic, and decontextualized how we think about distant things. See Robin Hanson's summary.

  15. ^

     See Section 4.2.3. Defense-dominance, Better Futures (William MacAskill, August 2025)

  16. ^

    As vast robotic fleets sweep across the cosmos, constructing astronomical megastructures with atomic precision, hear a single song echoing throughout all strata: "Non, Je ne regrette rien".

  17. ^

    The expected choiceworthiness approach assigns each theory a utility function and maximizes the credence-weighted sum of utilities. See What to Do When You Don't Know What to Do (Andrew Sepielli, 2009) and Moral Uncertainty (MacAskill, Bykvist & Ord, 2020).

    This faces the problem of intertheoretic comparisons: different theories may use different utility scales. But we can solve this with normalisation: Moral Uncertainty and Its Consequences (Ted Lockhart, 2000) proposes range normalization, equalizing each theory's range between best and worst options. Statistical Normalization Methods in Interpersonal and Intertheoretic Comparisons (Cotton-Barratt, MacAskill & Ord, 2020) proposes variance normalization, equalizing each theory's variance across possible outcomes.

    On either normalisation scheme, the expected choiceworthiness is maximised when proximal resources satisfy mundane values and distal resources satisfy exotic values.

  18. ^

    A Bargaining-Theoretic Approach to Moral Uncertainty (Hilary Greaves & Owen Cotton-Barratt, 2019)

  19. ^

    Normative Uncertainty as a Voting Problem (William MacAskill, 2016) and The Parliamentary Approach to Moral Uncertainty (Toby Newberry & Toby Ord, 2021)

  20. ^
  21. ^

    For discussion of both approaches, see Section 5 of Moral Decision-Making Under Uncertainty (Tarsney, Thomas, & MacAskill, SEP 2024).

  22. ^

     Credit to Avi Parrack for this point.



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

Stratified Utopia Future of Humanity Resource Allocation Values Longtermism Effective Altruism AI Alignment Cosmic Future 分层乌托邦 人类未来 资源分配 价值观 长期主义 有效利他 AI对齐 宇宙未来
相关文章