少点错误 10月14日 06:17
探究通用人工智能(AGI)的内在价值
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文深入探讨了当前关于通用人工智能(AGI)的论证,并质疑其核心驱动力。作者提出,许多关于AGI的论点最终都可以归结为对特定能力的追求,而非 generality 本身。文章通过“减法测试”和类比柏拉图的“正义论证”,试图剥离 instrumental benefits,探寻 generality 是否具有独立价值。作者分析了认知经济学、扩展假说、元解算器以及未知未知等论点,并区分了能力广度与真正的 generality。最终,文章强调了厘清AGI目标的重要性,并建议研究者在追求AGI时,应诚实面对其驱动因素,并优先考虑能力衡量、对齐研究和区分广度与 generality。

💡 **“减法测试”揭示核心问题:** 当前对AGI的追求,如数学推理、长期任务等,似乎可以被分解为一系列具体能力。若能实现所有这些特定能力,而没有“ generality ”这一概念,我们是否会失去价值?这促使我们反思, generality 是所有能力的集合名称,还是一个独立的、质的不同属性。

💰 **认知经济学论证的局限性:** 虽然通用系统可能因共享表征而更高效,但存在隐性假设: generality 的维护成本低于多个狭窄系统的总和。然而,当前研究显示,基础模型训练和运营成本高昂,而专用模型更有效率。此外,共享表征并非总是有利,可能导致灾难性遗忘或负迁移。

🚀 **区分技术必然性与社会经济必然性:** 许多AGI论点混淆了两者。技术必然性(如扩展假说)认为,随着计算量增加, generality 会自动涌现。然而,这可能只是一种功能上的广度,而非能无限泛化的真正 generality。社会经济必然性则关乎竞争压力和协调问题,需要治理机制来应对。

⚖️ **“元解算器”论证的潜在风险:** 认为AGI能解决所有问题的“元解算器”论证,可能忽视了“无免费午餐定理”,即没有一个模型能在所有领域最优。更重要的是,它混淆了能力与对齐。一个与人类价值观不符的强大通用智能,可能比多个对齐的狭窄系统更危险。

🧐 **广度与 generality 的辨析:** 文章强调,我们可能混淆了“能力广度”(拥有大量能力)与“ generality ”(能够跨领域泛化和链式推理)。真正的 generality 价值在于其能否利用已知领域解决新的、甚至未知的领域问题。若否,足够的广度可能就已足够,而非 generality 本身。

Published on October 13, 2025 9:49 PM GMT

Thesis Statement[1]

Current arguments for AGI can be distilled to arguments for specific capabilities, not for generality in itself. We need to examine whether there exists a genuine and sound argument for generality as an independent property.

Introduction

In Plato's Republic, Glaucon's challenge to Socrates is to show him why justice is good in and of itself; instead of arguing for its instrumentality. In other words, Socrates has to show Glaucon that we value justice itself, not merely for its after-effects:

"For I want to hear what justice and injustice are, and what power each has when it is just by itself in the soul. I want to leave out of account the rewards and the consequences of each of them." (Plato, Republic, 358b-c)

Following Glaucon's spirit, I dare ask: is generality in AI valuable in itself, or do we follow it merely for its expected instrumental effects?

Dialectic

The problem of reduction

When leading labs say "we're building towards AGI," what do they really mean? If we enumerate all the capabilities they desire (mathematical reasoning, long-horizon tasks, automated R&D and real-world economic tasks, ...) does anything remain in the term AGI after we subtract this list? Or is AGI simply a short name for "all of these capabilities together"?

Most, if not all, pro-generality arguments seem to be reducible to:

It doesn't seem to be wrong, then, to ask whether generality is the name we give to a sufficiently big conjunction of specific capabilities, or whether there is something qualitatively distinct: generality itself.

The subtraction test: If we could have all the specific capabilities that AGI promises, but without 'generality' (whatever that means, maybe we have all the capabilities but in separate, narrow models), would we lose any value?

The missing argument: intrinsic value

No one seems to argue that generality has value in itself (as we could argue about consciousness or wellbeing). Why not? Maybe because AI (seemingly) is instrumental by nature. So, why do we want generality? And, is that really what we want?

The argument of cognitive economy / from cognitive economics

A general system may be more efficient than maintaining a comprehensive set of narrow systems because:

But there seems to be an implicit assumption here, that is, it assumes that the cost of maintaining generality will be lower than the sum costs of ANIs (costs of development, inference, and maintenance). Is this empirically true? Could we build accurate mathematical cost models?

Currently, foundation models are very expensive to train and operate, and pushing the frontier is not getting any cheaper. Meanwhile, specialized models are much more efficient. So far, it seems that, if we think in terms of cost/benefit, empirical evidence may favor specialized models.

Moreover, this argument also seems to assume that shared representations are necessarily beneficial. Yet, in ML, it is well known that there are many trade-offs. A model aimed at doing everything may suffer from catastrophic forgetting or negative transfer.

The scaling hypothesis and two types of inevitability

Arguments for AGI often conflate two distinct claims about inevitability:

The distinction matters. Socioeconomic inevitability is a governance problem which suggests we need coordination mechanisms. On the other hand, technical inevitability is a scientific claim which suggests generality will emerge whether we coordinate or not.

Let's focus on the technical claim. If this view is correct, then asking "should we build generality" becomes moot. Generality would be an inevitable byproduct of scaling up systems initially designed for narrow tasks (such as next-token-prediction). We wouldn't be necessarily aiming for generality, rather, we'd simply observe its emergence.

But this argument smuggles in a few assumptions:

The meta-solver argument

This argument states that it'll be easier to build AGI and have it solve all other specific problems, than to solve every problem independently. This argument tends to come with the easily-repeated slogan "it'll be our last invention".

Some possible issues with this argument:

The argument from unknown unknowns

One could argue that we cannot know in advance what issues we may need to solve, and that generality gives us that flexibility to respond to unknown unknowns.

Yet this again seems to be an instrumental argument for, say, flexibility or adaptability, not for generality in itself. Moreover, what warrants us to assume that generality equals adaptability?[4] The most adaptable biological systems we know (bacterias) are not the most general.

Breadth or generality?

Perhaps we conflate breadth of capabilities with generality. Consider two systems:

What is more valuable? The answer seems to hinge on whether System B can sustain chains of generalization, using domain X to solve slightly-OOD domain Y, then using that to tackle even-further-OOD domain Z. If yes, then generality represents something genuinely powerful. If not, then System A's breadth may be superior. This latter case would suggest we actually value sufficient breadth, not generality per se.[5]

Open questions 

    Do any benefits attributed to AGI actually require generality, or merely sufficient breadth of capabilities?Is generality a real property or a convenient abstraction?If no sound argument exists for generality in itself, should we pivot toward developing the right set of highly-capable narrow systems?Does this same issue apply to ASI?

Conclusion

Paradoxically, the lack of a solid argument for generality in and of itself does not seem to mean we should not keep trying to build AGI. Rather, it means we should be honest about why we are building it. Maybe we are building it not because we see value in generality itself, but because:

    It seems inevitable given current incentivesWe believe (maybe incorrectly) that it will be more efficientWe want specific capabilities that we don't yet know how to build, and believe a general system would, in virtue of being general, solve themThe scaling hypothesis suggests generality may emerge whether we aim for it or not

This clarity isn't merely for philosophical amusement, it matters for determining research priorities and governance efforts. If we're building towards AGI for instrumental reasons, we should:

I think the fundamental question remains: are we building toward the right target, and do we even know what that target is?

I welcome counterarguments. If there exists a sound intrinsic argument for generality that I've missed, I'd genuinely like to hear it.

  1. ^

    I want to thank BlueDot Impact for accepting me into their inagural cohort of "AGI Strategy" where this discussion arose. This post would not exist without their great efforts to build the much needed Safety workforce.

  2. ^
  3. ^
  4. ^

    There may be a good argument to be developed here, if one can successfully argue that adaptability is an intrinsic component of generality, and not a mere after-effect.

  5. ^

    This formulation of generality as chainable out-of-distribution transfer draws on work in meta-learning and few-shot transfer learning. See Jiang et al. (2023), Tripuraneni et al. (2022), Sun et al. (CVPR 2019), and Ada et al. (2019) for theoretical foundations on OOD generalization and transfer bounds.

  6. ^


Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AGI 人工智能 通用人工智能 AI Generality AGI Strategy AI Safety Scalability Meta-solver Capabilities
相关文章