少点错误 09月25日
AI发展路径:全球停滞与可控加速的权衡
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

文章探讨了两种应对AI失控风险的策略:“全球停滞”(Global Shutdown)和“全球可控加速”(Global Controlled Takeoff)。“全球停滞”主张集中GPU资源并限制前沿研发;“全球可控加速”则允许在严格监管下,利用AI进行更多对齐研究。作者认为,两种方案在政治执行层面都面临巨大挑战,且“全球停滞”在协调上可能更简单。此外,文章强调了消除加速研发的经济压力、重视精确概念推理能力,以及利用未来可能出现的“足够吓人的AI演示”(Scary Demos)来推动政治行动的重要性。

💡 **两种AI发展策略的对比:** 文章提出了“全球停滞”和“全球可控加速”两种AI发展路径。前者旨在通过集中计算资源并限制前沿研发来阻止AI失控,后者则主张在国际监管下,利用AI加速安全对齐研究。作者认为,这两种方案都需要巨大的政治推动力,且“全球停滞”在协调上可能更易实现。

⚖️ **政治可行性与协调难度:** 文章深入分析了两种策略的政治可行性。作者指出,实现任何一种策略都需要全球主要国家和企业在GPU等关键资源上进行大规模协调,这本身就标志着政治格局的重大变化。同时,作者认为“全球停滞”比“全球可控加速”更简单,后者需要复杂的国际协议来区分安全与否的训练运行,并避免监管被俘获。

⏳ **移除加速压力与重视理性推理:** 作者强调,无论是哪种策略,消除经济上急于求成的压力都至关重要。文章批评了那些在追求AI发展的同时,忽视了精确概念推理能力的策略,认为这种能力对于应对失控的超级智能至关重要。作者暗示,一些公司文化可能过于依赖经验主义,而低估了理论推理的重要性。

🚀 **“足够吓人的AI演示”与政治推动:** 文章预测,随着AGI的临近,会出现“足够吓人的AI演示”,这将显著提高公众和政治家对AI风险的认识。作者建议,应提前为利用这些演示来推动全球性的AI监管协议做好准备,以便在关键时刻能够迅速采取行动,而不是被动应对。

⚙️ **关键考量因素而非最终结论:** 作者明确表示,以上分析是用于辅助决策的“齿轮”或考量因素,而非最终结论。他承认自己仍在处理信息,并可能遗漏其他人的视角。但总体而言,他认为这些因素对于理解和规划AI的未来发展至关重要。

Published on September 24, 2025 5:21 PM GMT

Two somewhat different plans for buying time and improving AI outcomes are: "Global Shutdown" and "Global Controlled Takeoff."

(Some other plans some people believe in include "ad hoc semi-controlled semi-slowed takeoff" and "race, then burn the lead on either superalignment or scary demos" and "decentralized differential defensive tech world.". I mostly don't expect those to work, but am mostly not talking about them in this post.)

"Global Shutdown" and "Global Controlled Takeoff" both include an early step of "consolidate all GPUs and similar chips into locations that can be easily monitored." 

The Shut Down plan then says things like "you cannot do any frontier development with the consolidated GPUs" (maybe you can use GPUs to run existing models that seem pretty safe, depends on implimentation details). Also, maybe, any research into new algorithms needs to be approved by an international org, and frontier algorithm development is illegal. (This is maybe hard to enforce, but, it is might dramatically reduce the amount of R&D that goes into it, since you can't be a billion dollar company who straightforwardly pours tons of resources into it without going to jail)

Controlled Takeoff says instead (as I currently understand advocates to advocate) something like "Frontier research continues, slowly, carefully, leveraging frontier controlled AI to do a ton of alignment research."

I'm generally pro "Shut It Down", but I also think Global Controlled Takeoff is much better than the status quo (both because it seems better in isolation, and because achieving it makes Shut Down easier), and I see some of the appeal depending on your exact beliefs.

But, some notes on strategy here.


"What's more impossible?"

A lot of AI safety arguments boil down to "what seems least impossible?". Is it more impossible to get a Global Shutdown, or to solve safe superintelligence with anything remotely like our current understanding or the understanding we're likely to get over the next 5-10 years?

I've heard a number of people say flatly "you're not going to get a global shut down", with a tone of finality that sounds like they think this is basically impossible. 

I'm not entirely sure I've correctly tracked which people are saying which things and whether I'm accidentally conflating statements from different people. But I think I've heard at least some people say "you're not getting a shut down" with that tone, who nonetheless advocate for controlled takeoff.

I certainly agree getting a global shutdown is very hard. But it's not obvious to me that getting a global controlled takeoff is much easier.

Two gears I want to make sure people are tracking:

Gear 1: "Actually consolidate the GPUs" is a huge political lift, regardless. 

By the time you've gotten various world powers and corporations to do this extremely major, expensive action, I think something has significantly changed about the political landscape. I don't see how you'd get it without world leaders taking AI more "fundamentally seriously", in a way that would make other expensive plans a lot more tractable.

Gear 2: "You need to compare the tractability of Global Shut Down vs Global Controlled Takeoff That Actually Works, as opposed to Something That Looks Close To But Not Actually A Controlled Takeoff."

Along with Gear 3: "Shut it down" is much simpler than "Controlled Takeoff."

A Global Controlled Takeoff That Works has a lot of moving parts. 

You need the international agreement to be capable of making any kind of sensible distinctions between safe and unsafe training runs, or even "marginally safer" vs "marginally less safe" training runs. 

You need the international agreement to not turn into molochian regulatory-captured horror that perversely reverses the intent of the agreement and creates a class of bureaucrats who don't know anything about AI and use the agreement to dole out favors.

These problems still exist in some versions of Shut It Down too, to be clear (if you're trying to also ban algorithmic research – a lot of versions of that seem like they leave room to argue about whether agent foundations or interpretability count). But, they at least get coupled with "no large training runs, period."

I think "guys, everyone just stop" is a way easier schelling point to coordinate around, than "everyone, we're going to slow down and try to figure out alignment as best we can using current techniques."

So, I am not currently convinced that Global Controlled Takeoff That Actually Works is any more politically tractable than Global Shut Down.

(Caveat: Insofar as your plan is "well, we will totally get a molochian moral maze horror, but, it'll generally move slower and that buys time", eh, okay, seems reasonable. But, at least be clear to yourself about what you're aiming for)


Gear 4: Removing pressure to accelerate is valuable for the epistemics of the people doing the AI-assisted alignment (if you're trying that).

One reason I think the Anthropic plan is actively bad, instead of "at-least-okay-ish," is that (given how hard they seem to actively oppose any kind of serious regulation that would slow them down), they seem intent on remaining in a world where, while they are supposedly working on aligning the next generation of AI, they have constant economic pressure to ship the next thing soon.

I believe, maybe, you can leverage AI to help you align AI.

I am pretty confident that at least some of the tools you need to navigate aligning unbounded superintelligence (or confidently avoiding creating unbounded superintelligence,) involves "precise conceptual reasoning" of a kind Anthropic-et-all seem actively allergic to. (see also behaviorism vs cognitivism. Anthropic culture seems to actively pride itself on empirics and be actively suspicious of attempts to reason ahead without empirics)

I'm not confident that you need that much precise conceptual reasoning / reasoning ahead. (MIRI has an inside view that says this is... not like impossible hard, but, is hard in a fairly deep way that nobody is showing respect for. I don't have a clear inside view about "how hard is it", but I have an inside view that it's harder than Anthropic's revealed actions thinks it is)

I think thinking through this and figuring out whether you need conceptual tools that you aren't currently good at in order to succeed, is very hard, and people are extremely biased about it. 

I think the difficult is exacerbated further if your competitor is shipping the next generation of product, and know-in-your-heart that you're reaching ASL danger levels that at least should give you some pause to think about it, but, the evidence isn't clear, and it would be extremely convenient for you and your org if your current level of control/alignment was sufficient to run the next training run.

So a lot of what I care most about with Shutdown/Controlled-Takeoff is making it no longer true that there is an economic incentive to rush ahead. (I think either Shutdown-y or Controlled Takeoff-y can both potentially work for this, if there's actually a trusted third party who is the one that makes calls about whether the next training run is allowed, who has the guns and compute).

 

Gear 5: Political tractability will change as demos get scarier. 

I'm not super thrilled with the "race to the edge, then burn the lead on scary demos" plan (specifically the "racing" part). But, I do think we will get much scarier demos as we approach AGI. 

Politicians maybe don't understand abstract arguments (although I think responses to If Anyone Builds It suggests they at least sometimes do). But I think there are various flavors of Sufficiently Scary Demos that will make the threat much more salient without needing to route through abstract arguments.

I think one of the most important things to be preparing for is leveraging Sufficiently Scary Demos when they arrive. I think this includes beginning to argue seriously now for global treaty shaped things and have them on hand so people go "oh, okay, I guess we do need that thing Those Guys Were Talking About After All" instead of just being bewildered.

 

Gears rather than Bottom Lines

I won't make a decisive claim that any of the above should be decisive in anyone's decisionmaking. I'm still processing the update that I can't really simulate the entire political disagreement yet and I'm not sure what other gears I'm missing from other people's perspective.

But, these are all individual gears that seem pretty important to me, which I think should be part of other people's overall strategizing.

I have a similar point comparing the feasibility of "Global Shut Down" vs "Decentralized Differentially Defensive Tech world that Actually Works", but, that's a fairly complex and different argument.

  1. ^

    to be clear this bias also applies in the MIRI-esque direction. But, they're not the one rushing ahead inventing AGI.



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI安全 人工智能 全球停滞 可控加速 AI治理 AI risk AI safety AI governance Global Shutdown Controlled Takeoff
相关文章