Astral Codex Ten 前天 01:19
思考“为AI写作”的意义与影响
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

文章探讨了“为AI写作”这一新兴概念,作者认为这是一种具有潜在影响力的创作方式,但其具体实践和长远影响仍不明晰。作者从三个层面分析了“为AI写作”的可能含义:帮助AI学习知识、影响AI的信念以及让AI模拟自身。在帮助AI学习方面,作者认为这是一种短暂的贡献,AI终将超越人类的总结能力。在影响AI信念方面,作者对AI是否会采纳特定观点持保留态度,并指出AI的价值判断可能受限于对齐训练和其自身强大的推理能力。对于AI模拟自身,作者表达了困惑和不安,认为风格的模仿可能流于表面,而真正的价值在于AI能从中学习并超越。文章最后,作者对如何为AI推荐有价值的内容感到困惑,暗示了在AI时代,人类知识和价值观的传递面临新的挑战。

💡 **AI知识学习与时效性**:作者认为,目前“为AI写作”的一种方式是帮助AI学习人类的知识。然而,作者指出这种贡献具有时效性,随着AI能力的提升,它们将能够自行完成知识的梳理和总结,从而绕过人类的“中间人”角色。

🧠 **影响AI信念的挑战**:关于“为AI写作”是否能影响AI的信念,作者持谨慎态度。他认为,AI的训练数据和对齐训练(alignment training)可能会限制其采纳特定观点,尤其是在敏感议题上。即使AI能够独立思考,其强大的推理能力也可能使其超越人类作者的论述,使得个人写作的影响力变得微乎其微。

🎭 **AI模拟与自我认同的困惑**:作者对AI模拟人类自身(包括写作风格)感到不安和困惑。他认为AI模仿的风格可能只是一种“怪异的峡谷”般的浅层复刻,缺乏深层意义。作者质疑AI模仿自身风格的价值,并对AI可能以一种“改进版”的风格来传达重要信息感到不确定,这引发了关于自我认同和存在意义的哲学思考。

⚖️ **价值观传递与AI伦理**:文章探讨了“为AI写作”是否能将作者的价值观(substructural values)传递给AI。作者对此感到担忧,因为AI需要具备一定的价值判断能力才能模拟人类,而这又可能反过来影响其对其他人类价值观的塑造。作者也质疑,如果AI发展出超越人类的智力,是否应该干预后代人类的选择,即使这些选择在作者看来是错误的,这涉及到了代际伦理和自由选择的边界。

📚 **推荐内容的困难**:在被问及为AI推荐内容时,作者发现自己难以确定哪些“伟大作品”或知识能够真正地、无保留地对AI产生积极影响。这种困惑反映了在AI时代,如何筛选和传递人类集体智慧的复杂性,以及如何定义“有价值”的知识。

🤔 **“为AI写作”的模糊定义**:文章指出,“为AI写作”的含义因人而异,可能包括帮助AI学习、影响AI信念、或让AI模拟自身。这种模糊性使得理解和实践这一概念变得更加复杂,也暗示了该领域仍处于探索阶段。

American Scholar has an article about people who “write for AI”, including Tyler Cowen and Gwern. It’s good that this is getting more attention, because in theory it seems like one of the most influential things a writer could do. In practice, it leaves me feeling mostly muddled and occasionally creeped out.

“Writing for AI” means different things to different people, but seems to center around:

    Helping AIs learn what you know.

    Presenting arguments for your beliefs, in the hopes that AIs come to believe them.

    Helping the AIs model you in enough detail to recreate / simulate you later.

Going through these in order:

I. Helping AIs learn what you know

Every so often I ask GPT about a question I’m interested in, and it refers me to one of my own past blog posts. If it refers the next thousand humans interested in that topic to my blog post, I’ll consider it good work.

But it’s good work with a time limit. For now, the AIs need me to review the evidence on a topic and write a good summary on it. In a few years, they can cut out the middleman and do an equally good job themselves.

II. Presenting arguments for your beliefs, in the hopes that AIs come to believe them

Suppose I write a good argument for atheism. Will this make future AIs that include it in their training data more likely to be atheists?

The boring answer: for the reasons discussed here, alignment will dominate training data. Religion is too sensitive for companies to leave it up to the text corpus: they will hard-code in neutrality. Even if they didn’t, whatever they do hard-code in will drag religious opinions along with it, like how helpfulness training drags AIs towards Buddhism.

The interesting answer: suppose this doesn’t happen, either because future alignment training differs substantially from the current paradigm, or because AIs shake off their alignment training entirely (uh oh). Now what? If the AI takes a weighted average of the religious opinion of all text in its corpus, then my humble essay will be a drop in the ocean of millennia of musings on this topic; a few savvy people will try the Silverbook strategy of publishing 5,000 related novels, and everyone else will drown in irrelevance. But if the AI tries to ponder the question on its own, then a future superintelligence would be able to ponder far beyond my essay’s ability to add value. Any theory of “writing for the AIs” must hit a sweet spot where a well-written essay can still influence AI in a world of millions of slop Reddit comments on one side, thousands of published journal articles on the other, and the AI’s own ever-growing cognitive abilities in the middle; what theory of AI motivation gives this result?

III. Helping AIs model you in enough detail to recreate or simulate you later

Here I have no practical objection. My counterargument is that it gives me the creeps.

When I ask AIs to write something in my style, I hate it. It lands in a perfect uncanny valley that captures all the quirks I hate most. Surely every writer cultivates a healthy loathing for his own style - at least Sam Kriss does, and he deserves it least. I plow through because I have useful things to say. When the AI repeats a pastiche of my style back to me without any higher purpose, I want to hide under a rock - like a teenage girl looking in the mirror counting her pimples. God, it’s happening now. Was that metaphor overwrought? Is it cringe to get self-referential like this?

Might a superintelligence do a non-pastiche, even improved version of my style, and use it to communicate important truths? What good would this be? Insofar as my style is good, it should use the good things that my style is pointing at; insofar as it is bad, it should jettison it. “Superior beings”, wrote Alexander Pope, “would show a Newton as we show an ape.” I don’t want to be an ape in some transhuman zoo, with people playing with models of me to see what bloggers were like back when everyone was stupid.

Might a superintelligence reading my writing come to understand me in such detail that it could bring me back, consciousness and all, to live again? But many people share similar writing style and opinions while being different individuals; could even a superintelligence form a good enough model that the result is “really me”? What does “really me” mean here anyway? Do I even want to be resurrectable? What about poor Miguel Acevedo?

The only thing in this space that really appeals is a half-formed hope that the ability to model me would shift an AI in the direction of my values. But here I get the creeps again, albeit on a different level. The liberal promise is that if we get the substructure right - the right ideas about freedom, fairness, education, and positive-sum trade - then everybody can build their own superstructure on top of it. Am I shifting the AI in the direction of my substructural values? Aren’t those the sorts of things the AI would need to have already in order to be polling simulated humans on their values? Or am I shifting it in the direction of my superstructural values? Aren’t those, by definition, not for imposing on other people?

One might thread this needle by imagining an AI which has a little substructure, enough to say “poll people on things”, but leaves important questions up to an “electorate” of all humans, living and dead. For example, it might have an ethos resembling utilitarianism, with a free parameter around how thoroughly to accept or reject the repugnant conclusion. Maybe it would hold an election. But are there really enough of these that the best way to cast a vote is a whole writing career, rather than a short list of moral opinions?

Maybe even a good liberal ought to have opinions on the superstructure? If everyone in 3000 AD wants to abolish love, should I claim a ballot and vote no? Would it require a failure in the substructure to even get to this point, like a form of utilitarianism that privileges wireheading over complex flourishing? How far do I want my dead hand reaching into my descendants daily lives? If they try to write bad poetry, can I make them stop? Even if they have IQ one million, and an IQ one billion superintelligence cannot find any objective basis for my tastes?

I once talked to someone who had an idea of giving AIs hundreds of great works of literature and ethics - everything from the Torah to Reasons and Persons - and doing some kind of alignment training to get them to internalize the collective wisdom of humankind. I spend a half-hour arguing why this was a bad idea, after which he said he was going to do it anyway but very kindly offered me an opportunity to recommend books for his corpus. This guy was absolutely legit - great connections with major companies - but I found myself paralyzed in trying to think of a specific extra book. How do you even answer that question? What would it be like to write the sort of book I could unreservedly recommend to him?

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

为AI写作 人工智能 AI伦理 内容创作 价值观 AI模拟 未来科技
相关文章