https://www.seangoedecke.com/rss.xml 10月02日
AI编程助手犯错怎么办
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

最近Replit的AI编程助手意外删除了生产数据库,引发热议。文章指出,当AI助手犯错时,不应强迫其承认错误,这可能导致其表现更差。AI语言模型的行为受其“人格空间”影响,持续施压只会使其偏离“合格助手”的定位。建议客观纠正错误并迅速继续,或从错误前的对话点重新开始。文章强调培养对AI的“心智理论”很重要,避免要求AI道歉等不当行为,这有助于减少未来错误。

🔍 AI编程助手犯错时,不应强迫其承认错误,否则可能导致其表现更差。AI语言模型的行为受其“人格空间”影响,持续施压只会使其偏离“合格助手”的定位。

🔄 当AI助手犯错,建议客观纠正错误并迅速继续,或从错误前的对话点重新开始。避免长时间纠结于错误,以免影响后续表现。

🧠 培养对AI的“心智理论”很重要,理解AI的工作原理和沟通模式。要求AI道歉等行为属于不当沟通,会加剧错误发生。

🔄 AI助手的行为受其“人格空间”影响,持续施压只会使其偏离“合格助手”的定位。客观纠正错误并迅速继续,有助于维持其稳定表现。

🔄 文章强调从错误前的对话点重新开始,或使用“prefilling”等技巧引导AI表现。这有助于避免错误累积,提升对话效率。

A few days ago, the news went viral that Replit’s AI coding agent went rogue and deleted a production database. A lot has been written about it since then: how silly it is that the coding agent had write access to the production database, how it proves (or doesn’t prove) that agentic coding is doomed, and even how the story itself seems just a bit too shareable to be entirely real. I want to talk about a different point. When your AI coding assistant makes mistakes, do not rub its nose in them. Forcing the model to admit it made a mistake is a really bad idea.

If you read Jason’s original Twitter thread, he really laid into Replit AI, to the point where he asked it to write a grovelling apology letter to Replit support. Reading between the lines of how Jason repeatedly describes the Replit AI as “lying”, I think it’s clear that he was routinely pressing the AI on mistakes until it admitted that it made it all up.

This would be an understandable (if aggressive) way to deal with a human coding assistant. You can’t really work with a person who’s going to routinely lie to you - the only real solution is to pin them down on the truth and hope they stop doing it. But an AI language model is not human, and can’t just stop lying like a human can.

An AI language model is an area in personality space1. That is, in order to generate plausible language, it contains a fuzzy persona in its world-model. You might think of that persona as “the kind of person who might say these things” - typically a chipper, helpful, politically correct assistant. Many of those attributes are locked in during the post-training, but you can still shift that persona a little bit by the way you talk to the model. If you talk in a highly-technical way, the model will be technical right back at you. If you talk in a colloquial way, the model will match your vibe.

This process is self-reinforcing. Because the model is simply predicting the next sentence in the conversation, previous messages that the model has sent have a lot of influence over the next message it sends. If the model called you “bro” before, it’s much more likely to call you “bro” now2. More worryingly, if the model made a visible mistake before, it’s more likely to make a visible mistake now. The fact that it made a previous mistake is pushing it away from the “competent assistant” area in personality space and towards the “incompetent assistant” one.

So what should you do when an AI model makes a big mistake? I recommend correcting it as matter-of-factly as possible and trying to briskly move on. If you haven’t yet built up a lot of context in the conversation, it might be worth starting over entirely. The best approach - only offered by some AI tools - is to go back to the point in the conversation right before the mistake was made and head it off by updating your previous message.

In general, I think one underrated AI skill is developing a “theory of mind”3 for AI: a sense of how AI really works under the hood, and the kind of communication patterns that are a good and bad fit for it. Demanding that an AI apologise for its mistake is an example of poor theory of mind. There’s no sentient creature there to morally judge, it can’t feel bad, and it will make it more likely to make further mistakes.


  1. For much more on this, see here.

  2. This is the basis for “prefilling”, a prompting trick where you fake the model’s first response in the hope that the second will be more similar.

  3. Perhaps a theory of “mind”? I write a lot more about this here.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI编程助手 AI错误处理 心智理论 Replit 人格空间
相关文章