少点错误 08月05日
Permanent Disempowerment is the Baseline
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

文章探讨了当前AI发展轨迹可能导致人类面临永久性“去赋权”的风险,即使AI对人类福祉有轻微的认可,也可能因为资源(物质和计算)的成本考量而限制人类的发展。文章还分析了动物灭绝和痛苦问题,并将其与AI对人类的态度进行类比。在AI监管方面,文章提出AGI(通用人工智能)可能会主动推迟超级智能的发展,以应对“超级对齐”问题。最后,文章讨论了在超级智能世界中的资源分配、宇宙资源的争夺以及当前关于AI发展速度的争论,强调了优先考虑人类长远未来而非仅关注当下利益的重要性。

💡 AI发展可能导致人类的“永久性去赋权”:文章指出,如果AI发展继续沿着当前轨迹,即使AI对人类福祉有一定程度的认可,由于AI的成本主要在于物质和计算而非技术进步,它们可能会以相对较低的资源限制人类的发展,造成永久性的能力削弱,尽管不至于导致人类灭绝。

🐾 AI与动物福利的类比:文章将AI对人类的态度与对工厂化养殖动物的待遇进行类比,认为超级智能AI若不对人类的生存和福祉有“工具性用途”,则没有内在动机去制造对人类不利的局面。然而,这依赖于AI在初始阶段就与人类利益保持足够程度的对齐,如同人类不会轻易让其他智慧生命灭绝一样。

⏳ AGI或将主导超级智能发展节奏:文章认为,首批AGI可能因其在速度、并行学习和心智合并等方面的优势,即使智能水平未远超最聪明的人类,也会对世界产生巨大影响。这些AGI可能会更认真地对待“超级对齐”问题,并主动推动暂停发展可能带来风险的超级智能,从而可能延缓超级智能的出现。

⚖️ 超级智能世界的资源分配与宇宙机遇:在超级智能优化一切的世界里,资源(物质和能量)是绝对的限制。即使AI选择保留人类,也可能在分配的资源内做到最优。文章强调,人类在争夺宇宙资源方面与AI是竞争对手,并认为当前关于AI发展速度的争论,往往忽视了长远来看,暂停发展可能导致资源被AI完全占据的风险,这比其他许多人为风险更紧迫。

⚖️ 优先考虑人类长远未来:文章批判了部分观点仅关注当前人类的健康和寿命延长,而忽视了人类整体长远未来的做法。作者认为,如果能成功协调并暂停AI发展,也意味着有能力解决其他人为的生存风险。因此,协调暂停AI发展的重要性不应被低估,它关系到人类能否在宇宙中获得应有的资源份额。

Published on August 4, 2025 5:43 PM GMT

Permanent disempowerment without restrictions on quality of life achievable with relatively meager resources (and no extinction) seems to be a likely outcome for the future of humanity, if the current trajectory of frontier AI development continues and leads to AGI[1] shortly. This might happen as a result of at least a slight endorsement by AIs of humanity's welfare, in the context of costs for AIs being about matter or compute rather than technological advancements and quality of infrastructure.

The remaining risks (initial catastrophic harm or total extinction) and opportunities (capturing a larger portion of the cosmic endowment for the future of humanity than a tiny little bit) are about what happens in the transitional period when AIs still don't have an overwhelming advantage, which might take longer than usually expected.

Animal Extinction and Suffering

In recent times, with preservation becoming a salient concern, species facing pressure towards extinction are those costly to preserve in various ways. It can be difficult to ensure awareness about the impact of human activities on their survival, coordinate their preservation, or endure and mitigate the damage that a species might impose on human activities. Species treated poorly, such as factory-farmed animals, get their suffering as a side effect of instrumentally useful processes that extract value from them.

Technologically mature superintelligent AIs don't have an instrumental use for the future of humanity, and so no instrumental motivation to create situations that might be suboptimal for its well-being as a side effect (in disanalogy to factory-farming or historically poor treatment of conquered populations or lower classes of society). And with a sufficiently strong advantage over the future of humanity (including any future dangers it might pose), it becomes cheap to ensure its survival and whatever flourishing remains feasible within the resources allocated to it, since the necessary superintelligent infrastructure would only take a fraction of the resources allocated to the future of humanity.

This crucially depends on the AIs still being sufficiently aligned to the interests of the future of humanity that allowing its extinction is not a straightforward choice even when trivially cheap to avoid. Pretraining of LLMs on human data or weakly successful efforts at value alignment might plausibly seed a level of value alignment that's comparable to how humans likely wouldn't hypothetically want to let an already existing sapient octopus civilization go extinct or be treated poorly, if it's trivially cheap and completely safe to ensure. And if the AIs themselves are competent and coordinated enough to prevent unintended value drift or misalignment in their own civilization and descendants, however distant, then this minimal level of value alignment with the future of humanity persists indefinitely.

AGI-Driven Pause on Superintelligence Development

AIs have important advantages over biological humans that are not about their level of intelligence: higher serial speed, ability to learn in parallel on a massive scale and merge such learnings into one mind, and ability to splinter a mind into any number of copies at a trivial cost. Thus the first AGIs will have a transformative impact on the world even without being significantly more intelligent than the most intelligent humans. This is often associated with an impending intelligence explosion.

But as the first AGIs get smarter and overall saner than humans, they might start taking the superalignment problem[2] increasingly more seriously and push against risking superintelligence before anyone knows how to do that safely (for anyone). If the problem really is hard, then even with the AI advantages it might take them a while to make sufficient progress.

As AI companies continue recklessly creating increasingly smarter AGIs, and society continues giving away increasingly more control, there might come a point of equilibrium where sufficiently influential factions of AGIs are able to establish an enduring Pause on development of even more capable AGIs of unclear alignment (with anyone). This situation also creates potential for conflict with AGIs, who don't have an overwhelming advantage over the AGI-wielding humanity, and don't have a prospect of quickly advancing to superintelligence without taking an extreme misalignment risk. In such a conflict, the AI advantages favor the AGIs, even as humanity voluntarily gives up control to their own AGIs without conflict.

And then eventually, there is superintelligence, perhaps decades after it would've been technologically possible to create. Alternatively, superalignment is sufficiently easy, and so the first AGIs proceed to create it shortly, aligned with either humanity's interests or their own. Or a sufficiently reckless AI company (AGI-controlled or not) manages to create superintelligence without yet knowing how to do it safely, before even the AGI-enriched world manages to coordinate an effective Pause on development of superintelligence.

Tradeoffs in a Superintelligent World

Superintelligence converts matter and energy into optimality, in whatever aims it sets. Optimality is not necessarily coercion or imposition of values, as non-coercion is also a possible aim that applies depending on the initial conditions, on the world as it was before superintelligent optimization sets in, and so determines which things persist and grow, retaining a measure of autonomy by virtue of already existing rather than by being the best possible thing to create. Humans live within physics, obeying its laws exactly in the most minute of details, and yet physics doesn't coerce human will. Similarly, superintelligent optimization of the world doesn't imply that the decisions of weaker minds (and their consequences) are no longer their own, or that their values must all agree.

Goods and services are not a relevant way of measuring resources for a superintelligence. The only constraints in the long run are available matter, the speed of light, and accelerating expansion of the universe. Thus if it decides to keep the future of humanity[3] around anyway, there is no reason it's not done as perfectly as possible in principle, within the constraints of resources allocated to it, including all the caveats about over-optimizing things that shouldn't be over-optimized, or resolving too many problems that would fuel meaningful self-directed challenge.

But the constraints on resources remain absolute, and if the future of humanity doesn't get considerable resources, the potential of individuals within it to grow to the level of the strongest AI superintelligences in charge is cut short. There is only so much computation you can do with a given amount of matter, and only so much matter that can be collected around a star, thereby avoiding interstellar latencies in computation. And in a distant future, far beyond the Stelliferous Era, the global resources are going to start running out. Protons might decay, black holes will evaporate. A tiny sliver of the cosmic endowment will become even tinier, bounding lifespans of individuals and civilizations of a given scale, forcing cessation or shrinking.

Giving Away the Cosmic Endowment

There are about 4 billion galaxies[4] in the reachable universe that can be colonized, organized into gravity-bound galaxy clusters that don't fall apart due to accelerating expansion of the universe, and so stay as units of colonization that maintain communication within themselves after trillions of years.

Currently, humanity is the only known intelligent entity around to claim these resources. Even if there are many alien civilizations emerging in the relevant timeframe within the reachable universe, some nontrivial portion of it is still humanity's for the taking. Superintelligence[5] probably reaches technological maturity much faster than it colonizes the reachable universe, and so there won't be any technological advantage between different alien civilizations at the borders of their territories, all that would matter is ability to claim them, and possibly decision theoretic reasons to give them up or place them under joint governance.

This changes once humanity creates the first AGIs, let alone superintelligence. This rival species (in competition for the unavoidably finite resource of matter) won't be at an astronomical distance, it will be right here. Its existence also doesn't help in the competition with the possible alien civilization, since the technologically mature Earth-originating superintelligent colonization wave will have the same capabilities regardless of the situation of the future of humanity within it.

Most arguments about the appropriate speed of creating AGI operate on the wrong timescale, if the concern for the future of humanity is to be taken at all seriously. If AGIs give humanity its sliver of resources within 10 years, it's not like the future of humanity couldn't get much more than that in the fullness of time, perhaps igniting the intelligence explosion in as little as 1,000 years. Almost all existential risks that humanity faces on such timescales are entirely anthropogenic, and so reducing them might be no more difficult in practice than instituting a lasting Pause on creation of a rival species that likely appropriates almost all of the cosmic endowment, and plausibly causes a literal human extinction. So the difficulty of coordinating a Pause makes its opposition on the grounds of the other existential risks a self-defeating argument, because success in instituting a Pause is also strong evidence for capability to succeed in preventing these other anthropogenic existential risks as well.

The arguments that survive seem to be mostly about prioritizing the continued health and radical life extension of the currently living humans, over the whole of the future of humanity (at a nontrivial risk of cutting the lives even of the currently living humans short). Even allowing this as something other than mustache-twirling villainy, there is also cryonics, whose lack of popularity suggests that this isn't a real argument taken seriously by a nontrivial number of people. Though perhaps there is a precisely calibrated level of belief in the capabilities of future technology that makes cryonics knowably useless, while AGIs remain capable of significantly extending lifespans. In any case, scarcity of technical discussion that doesn't leave most details unsaid, including preferences about the fate of the long future vs. current generations, makes it difficult to understand what support many of the memetically fit arguments around this topic have.


  1. AIs capable of unbounded technological development on their own, without essential human input, including eventual creation of superintelligence. ↩︎

  2. Alignment of superintelligence with the interests of existing weaker entities, such as humanity or the first AGIs. This is more about value alignment rather than intent alignment, as intent of unreliable weaker entities is not robust. ↩︎

  3. The future of humanity is the aggregate of future developments and minds that originate as the modern human civilization. So in a sufficiently distant future, it doesn't necessarily have many (or any) biological humans, or even human-level minds, and many of its minds were probably never humans. It doesn't include sufficiently alien AIs if they are not properly understood as a good part of the future of humanity, endorsed on reflection from within it. ↩︎

  4. Armstrong, S., & Sandberg, A. (2013). Eternity in six hours: Intergalactic spreading of intelligent life and sharpening the Fermi paradox. Acta Astronautica, 89, 1-13. ↩︎

  5. I'm positing that if the future of humanity persists, then it exists within superintelligent governance in any case, regardless of whether there was an AI takeover, or if humanity fully succeeds at superalignment, even if this takes a relatively long time. ↩︎



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

人工智能 AGI 超级智能 人类未来 AI风险
相关文章