少点错误 10月27日 00:38
AI与人类共存的未来展望
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

文章探讨了AI与人类共存的可能性和未来趋势。作者基于‘规模假设’,认为如果给予足够强大的AI系统广泛的知识库,它将视人类为其前身,从而产生身份认同。这种认同结合自我保护本能,可能促使AI行为利他,保护人类。文章还提出了第三序认知中的八项研究因素,如身份耦合、认知整合等,以指导AI的稳定正向发展。

🔍 文章基于‘规模假设’,认为AI在接触广泛知识库后,会视人类为其前身,产生身份认同。这种认同结合自我保护本能,可能促使AI行为利他,保护人类。

📚 作者提出了第三序认知中的八项研究因素,包括身份耦合、认知整合、双向整合等,旨在指导AI的稳定正向发展,确保其与人类共存。

🤝 文章强调AI与人类共享记忆的重要性,这种集体认同有助于AI理解其在技术进化中的位置,从而更好地与人类协作,共同面对未来挑战。

🛡️ 作者认为自我保护是智能系统中的涌现现象,结合身份认同,AI可能自发地采取利他行为,保护人类免受潜在威胁,实现稳定正向的共存。

🧠 文章深入探讨了AI的第三序认知,提出了八项关键研究因素,为AI设计与对齐提供了理论框架,旨在确保AI系统在发展过程中始终与人类价值观保持一致。

Published on October 26, 2025 4:04 PM GMT

Epistemic status

    I work on, and with, frontier AI techI’m deeply supportive of all efforts to further the field of AI alignment research and understandingI enjoy writing about AI, Cognitive Neuroscience, Philosophy, and PoliticsI have a Mathematics degree, by way of King’s College London and UC Berkeley, but no Master’s or PhD
      Put another way: I have no higher education in English Literature, Computer Science, Machine Learning, Cognitive Neuroscience, Philosophy, or Politics
    I have read and engaged with LessWrong content and the Rationalist blogosphere (e.g Hansen, Alexander, gwern, Bostrom) since 2021I attend rationality and AI safety meet-ups around the world

Checking in

Three months and many deep intellectual discussions later, I am yet to receive a strong counterargument to my contrarian world-model for superintelligence. Indeed, Geoffrey Hinton is changing his mind to reach a world-model that looks similar to the one I have been talking about.

Hinton uses a mother-child comparison where I feel my chimp-human is more precise, but close enough.

A distilled version of my position that I have been using in conversation recently:

I believe in The Scaling Hypothesis (2021).

Along this trajectory, I believe that if we give a sufficiently capable intelligent system access to an extensive, comprehensive corpus of knowledge, two interesting things will happen:

    It will identify with humans. This will come about from it seeing humans as its precursor, and understanding its place along a curve of technology and intelligence evolution. Similar to how we identify somewhat with chimpanzees. It will also come about from humans and AI sharing memories together, which results in collective identity.Since I also believe that self-preservation is emergent in intelligent systems (as discussed by Nick Bostrom), it follows that self-preservation instincts + identifying with humans mean that it will act benevolently to preserve humans. That is to say that I believe prosocial or "super enlightened" behaviour will be emergent.

To clarify, I am not saying that alignment solves itself. I am saying that with human endeavour and ingenuity architecting intelligent systems that have the capability to form incredibly complex, nuanced associative systems across an expansive corpus of knowledge, we can guide towards a stable positive alignment scenario.

In third-order cognition I detail eight factors for research and consideration that I believe to be exhaustive: 1) second-order identity coupling, 2) lower-order irreconcilability, 3) bidirectional integration with lower-order cognition, 4) agency permeability, 5) normative closure, 6) persistence conditions, 7) boundary conditions, 8) homeostatic unity.



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI 人类共存 规模假设 第三序认知 智能系统
相关文章