少点错误 10月28日 03:43
AI 发展与权力集中:风险与应对
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

文章探讨了人工智能(AI)发展,特别是超级智能(ASI)可能带来的进攻性武器化风险。作者认为,随着AI训练成本的快速下降,一旦AI能力被用于制造廉价且毁灭性的武器,将难以有效监管。为应对此威胁,文章提出了一种观点:由少数国家(如中美)垄断AI技术,并以此为基础建立全球性的非扩散机制,类似于核武器的管控模式。这种模式旨在通过先发优势,限制其他行为体获得AI的进攻性能力。然而,这也将导致前所未有的权力集中,可能将全人类置于少数国家意志之下。文章进一步分析了即使在这种权力集中的情况下,人类福祉仍有可能得到保障,因为ASI可能带来的极大富足使得维持普遍高水平的生活成为可能,并且“利他主义者”可以通过掌握进攻性AI技术来制衡“优化者”。最终,作者认为AI的非扩散和权力集中是应对AI武器化风险的可能路径,尽管存在挑战,但值得深入探讨。

🎯 **AI 武器化的双重风险:** 文章指出,AI,尤其是超级智能(ASI),一旦被武器化,可能导致廉价、高破坏性的武器出现。由于AI训练成本的快速下降,早期允许AI能力扩散将导致后期监管的无效性,并可能使少数人掌握毁灭世界的能力。

🛡️ **垄断与非扩散的应对策略:** 为应对AI的进攻性扩散,文章提出一种“垄断”策略,即由少数国家(如中美)率先掌握ASI技术,并利用其战略优势限制其他国家或实体获取AI的进攻性能力。这种模式借鉴了核武器的管控经验,强调“先发制人”的非扩散。

⚖️ **权力集中与潜在福祉:** 尽管AI技术垄断会带来巨大的权力集中,将人类置于少数国家之下,文章认为,ASI带来的极大富足可能使得维持大多数人的高品质生活成为可能。同时,掌握进攻性AI的“利他主义者”可以通过核威慑式的策略来制衡“优化者”,确保人类基本福祉。

🚀 **“去中心化”策略的局限性:** 作者对“去中心化”和“去加速”(d/acc)策略表示怀疑,认为其难以应对广泛分布且廉价的AI进攻性技术。即使在某些领域存在防御性解决方案(如密码学),也无法保证所有未来进攻性技术都能被有效防御,单一的、廉价的、具有生存威胁的技术就足以破坏防御体系。

Published on October 27, 2025 7:38 PM GMT

Title is a question because I'm looking for an answer, not providing it. 

I wanted to stress-test some ideas and reason through some questions I have about concentration of power. These have been bubbling in the back of my head for a while, so I've taken two hours this morning to write some of them up and see how others feel.


The main project I'm currently working involves the mitigation of AI proliferation, particularly for offensive capabilities. The core issue is that a) advanced AI can plausibly allow you to make cheap, ultra-destructive weapons, and b) that the cost to train an advanced AI system declines quickly over time because of algorithmic efficiency improvements. As a result, the more you allow proliferation early on, the more invasive and ineffective enforcement becomes, until we hit a point where a few people are empowered to destroy the world at the expense of everyone else (maybe they built a fail-deadly, they're terrorists, irrational, try to build a superintelligence and end up misaligning it, etc).

In my model of things, these destructive technologies are offense dominant (impluasibly inefficient or impossible to defend against), which means that the only realistic way to control them is to limit their spread preemptively. While it's hard to predict the future, I'd expect that the first ASI labs to end up nationalized by the USG and PRC, meaning that offensive capabilities are initially monopolized. If they recognize the security interest of restricting AI proliferation (as the U.S and Soviet Union mutually agreed to do with nukes), they will use their strategic position to try and disempower other states/actors from getting access to offensive AI capabilities before they can become cheap to access. Perhaps they commandeer the entire AI supply chain, and put restrictions in place such that all future AI systems are aligned to the US-China coalition goals. 

As far as I can see, this would reasonably secure the world against the threat of black ball AI tech. But an important secondary issue it creates is concentration of power, wherein the rest of humanity is subjected to the whims of the USG and PRC, because they have a monopoly on violence through their control of ASI technology and the commandeering they just did, and have to continue to do to in order to make sure an unauthorized ASI is never built.

Here are some of my intuitions about this situation:

    First, there doesn't seem to be a reasonable alternative to this sort of monopolization. Offense-dominant technology is only really curtailable by MAD style policy, which breaks down as the number of actors increases. If you want humanity to enjoy the benefits of powerful AI systems, then you need to make sure that the dual-use systems are preemptively controlled before they're distributed, which means that they're passed through some kind of central authority first. That central authority needs to have a strategic advantage over everyone else, in order to make sure that they're genuinely not building the dangerous technology (think IAEA, sanctions, bombing Iranian enrichment facilities, etc).
      As a simple example, consider allowing other countries/companies/individuals to train their own ASI. If this ASI might end up misaligned, and there are cheap weapons it can develop that can threaten everyone else, each of these actors is a danger to the whole of civilization. The only way to avoid this scenario is to disempower those future actors ahead of time, such that they can never build the misaligned ASI or comparable weapons in the first place.Basically, I think d/acc (at least the defensive and decentralized components) is an unviable strategy, because you won't be able to reliably defend against the widely distributed and cheap offensive technology ASI would allow. While I'm sure there are comprehensive, defense dominant solutions that can be widely distributed in some domains (cryptography being the classic example), I don't think this is true of all future offensive technologies, which it needs to be. If there's even a single cheap, existentially powerful technology that exists (and we already have promising candidates like mirror life or misaligned ASIs), then the plan of allowing AI to proliferate for defense will have the opposite effect of favoring bad actors.[1]
    This doesn't mean that ordinary people can't be allowed to use any AI technology. They can! It just needs to be centrally processed and controlled first to make sure the dual-use stuff is restricted in some way (certain robotics, AI r&d, biology, engineering, and material sciences applications most likely). This looks something like allowing the use of superhuman medical researchers to create wonderdrugs, although the program is backdoored/pre-aligned/monitored to make sure it cannot comply with requests to develop bioweapons.Second, it seems to me that most people might end up ok despite the concentration of power required by having governments monopolize ASI, even in the most extreme dicatorial scenarios.
      The primary reason is that, given the abundance that ASI might enable, both in the efficient production of goods and explosive supply of labor, it would become very cheap (as a fraction of GDP) to have most people live extraordinarily good lives. Eliezer has made the point before that misaligned ASIs would have no reason to spare even a tiny fraction of their resources, because they won't have even a tiny preference to satisfy. While I think this is likely true of misaligned ASIs, it seems wrong to argue that the same is true of most humans, who do have some nonzero preference for being altruistic along with their other goals. If we end up in a situation in which almost all labor ends up controlled by a handful of company executives/politicians, it would only take a few of them committing to a permanent welfare payment for everyone to live great lives (at a small fraction of their actual resources equivalent to their preferences).A common criticism of this altruist/redistributionist perspective is that Moloch devours all, and that those who don't advocate for human welfare will eventually outcompete those that do. However, it seems unclear that there's a realistic mechanism for the non-welfare advocates to do this to the altruist decision makers, given the presence of offense-dominant weapons at the limit of intelligence. Even though the U.S has unbelievable conventional military superiority to North Korea, for instance, the fact that they have nuclear weapons means that we cannot arbitrarily impose our preferences about how North Korea should act onto them. Although the U.S might "win" a nuclear exchange with North Korea, the damage it would be able to cause to the things we value in the process is so large that we don't take the steps that would disempower them in the first place.

      Roughly speaking, you can swap out "U.S" and "North Korea" with "Optimizers" and "Altruists". Although there might be some (or even most) decision makers who would prefer to get rid of all the strategically/politically disempowered humans who are consuming resources to better optimize for more galaxies/inter elite status games/etc, there probably won't be a decisive action they can take against civilians or the altruists that advocate for them that doesn't end in catastrophic retaliation. If an altruist has their hands on offense-dominant ASI technology, then they can just commit to using it in retaliation if they confirm that the optimizers are trying to disempower them. As a somewhat modern parallel, it's reasonable for the U.S to discard plans to assassinate Kim Jong Un on the basis that it wouldn't be able to stop a retaliatory strike when the regime panics. And given the coordination/verification mechanisms of superintelligences for upholding these contracts, rather than relying on human decision making in the moment, you could probably get pretty ironclad guarantees about the behavior of other actors in spite of their preferences.[2]
    This situation would still lead to unimaginable imbalances in material wealth, probably many orders of magnitude more extreme than today. But I don't believe that inequality is intrinsically problematic from a welfare perspective: it's far more important that the people at the bottom meet the absolute threshold for comfort than it is for a society's Gini coeficient to be lower. The fact that Elon Musk can afford to send his kids to university for millions of years doesn't detract from the four years I spent attending on scholarship. The only direct threat the ultra-wealthy pose to my wellbeing is their capture of the state/important institutions that I use, which in the above scenario is compensated for by altruists who can cheaply afford for me to be happy and are able to threaten something their competitors care about over it. 

Given all of this, concentration of strategic power seems both necessary and... probably ok? Even in this worst-case scenario, where power is totally monopolized by a handful of people (say the Oversight Committee from AI 2027), all it takes is a small preference among some of them for civilians to live great lives for them to be able to do so. The state needs to clamp down on offense-dominant offensive capabilities to make sure that non-state actors never get access to them (or else someone will eventually use them), but afterwards, beneficial AI technology can be allowed to mostly flourish.  

Of course, I still think it would be better to have outcomes like value-lock-in of welfare/political rights for all humans at the inception of ASI, or for the U.S and China to somehow avoid intent alignment and instead aim straight at civilizational alignment where the ASI runs everything itself, rather than having to account for the preferences of very few. But I believe the above situation---nationalization of labs, intent aligned AI, realization of the strategic importance of non-proliferation, permanent monopolization of strategic power---represents the default path to dealing with the proliferation problem, so it seems important to reason about what it might look like and whether that's acceptable. 

  1. ^

    As an example of the kinds of defensive plans I'm talking about, I'm referring to Luke Drago and Rudolf L's proposal for defensive augmentation, or Helen Toner's adaptation buffers, where we prepare society ahead of time through the wide-scale distribution of technology.

    I have several thoughts on this, which I'll explore in a future, properly edited post on the strategic implications of proliferating powerful AI systems. My main critique is that I do not share their optimism that you can efficiently respond to all existential threats, given the cost-imbalance of defending against them. As the number of threats and actors increase, spending on defense has to grow exponentially. Offensive investments, on the other hand, are linear, since they can focus all of their effort on targeting the areas that are under-invested or even impossible to defend against. Ie, it's much cheaper to try and overwhelm defensive capacity by just launching two pandemics at the same time than it is to make sure you are prepared for multiple pandemics. 

    Analogously, we can think about how ridiculously inefficient trying to intercept nuclear ICMBs is, and how you could obsolete the hundreds of billions of dollars worth of interceptor research you'd need by a) building faster missiles, b) using more missles, or c), just putting your bomb in a shipping container and docking it next to Washington.

    And even if you secure one avenue (wastewater monitoring, DNA synthesis screening, and KYC ends up being enough defense-in-depth for bioweapons that it's basically solved), there's no guarantee you can solve them all. If you upgrade the problem from a pandemic to mirror life, for example, outbreak monitoring and antibiotic response plans start to break down completely. And so on for swarms of insect-sized LAWs, nanotechnology, misaligned ASIs, and the unknown-unknowns of weapons development that lurk at the top of the tech tree. The only reasonable solution to this problem is just to make sure other people don't have weapons in the first place, or preemptive non-proliferation.

    I do not have the confidence to distribute general ASI without a guarantee that people will not have access to these offense-dominant technologies, and the only way to make sure that their AI cannot do those things would be for a more powerful central authority to have achieved ASI first and used their first-mover advantage to strip all future offensive capabilities away from other actors. 

  2. ^

    Similarly, even if the Chinese government decided to try to be more competitive by ignoring the welfare of its own citizens, it wouldn't necessarily be able to impose its will on the U.S given the disproportionate leverage of offense-dominant weapons.



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI 人工智能 超级智能 ASI AI安全 权力集中 非扩散 AI伦理 AI风险 AI Governance Artificial Intelligence Superintelligence AI Safety Power Concentration Non-Proliferation AI Ethics AI Risks AI Governance
相关文章