https://simonwillison.net/atom/everything 09月30日
LLM agent定义明确化
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

近年来,'agent'一词在AI领域逐渐普及,但其定义多样导致沟通困难。本文提出'LLM agent'的新定义:通过在循环中运行工具以实现目标。这一定义区分了技术讨论中的agent与比喻性用法(如旅行代理或人类替代品),强调工具调用和目标导向的重要性。同时指出,长期记忆可通过额外工具实现,而人类特有的责任感和自主性仍是AI agent无法复制的。该定义有助于技术社群更清晰地进行讨论,避免混淆。

🔍 在技术讨论中,'LLM agent'被明确定义为通过在循环中运行工具以实现目标,这一概念已逐渐被AI工程领域广泛接受,有助于减少沟通障碍。

🛠️ 该定义强调LLM agent的核心机制是调用工具并利用工具的输出进行迭代推理,形成闭环以解决问题,这与许多LLM API的设计模式一致。

🎯 定义中的'实现目标'表明agent操作具有终止条件,区别于无限循环,且目标可由用户设定或由其他LLM生成,支持复杂的子代理模式。

🧠 尽管未明确要求记忆,但工具调用本身构成了与模型的对话历史,提供了必要的短期记忆支持;长期记忆可通过扩展工具集实现,而非固有属性。

🤖 将agent等同于人类替代品(如客服或营销代理)是一种常见的误解,因AI缺乏人类的责任感和自主性,导致此类应用前景有限且效果不彰。

I think “agent” may finally have a widely enough agreed upon definition to be useful jargon now

18th September 2025

I’ve noticed something interesting over the past few weeks: I’ve started using the term “agent” in conversations where I don’t feel the need to then define it, roll my eyes or wrap it in scare quotes.

This is a big piece of personal character development for me!

Moving forward, when I talk about agents I’m going to use this:

An LLM agent runs tools in a loop to achieve a goal.

I’ve been very hesitant to use the term “agent” for meaningful communication over the last couple of years. It felt to me like the ultimate in buzzword bingo—everyone was talking about agents, but if you quizzed them everyone seemed to hold a different mental model of what they actually were.

I even started collecting definitions in my agent-definitions tag, including crowdsourcing 211 definitions on Twitter and attempting to summarize and group them with Gemini (I got 13 groups, here’s the tool-using LLMS one.)

Jargon terms are only useful if you can be confident that the people you are talking to share the same definition! If they don’t then communication becomes less effective—you can waste time passionately discussing entirely different concepts.

It turns out this is not a new problem. In 1994’s Intelligent Agents: Theory and Practice Michael Wooldridge wrote:

Carl Hewitt recently remarked that the question what is an agent? is embarrassing for the agent-based computing community in just the same way that the question what is intelligence? is embarrassing for the mainstream AI community. The problem is that although the term is widely used, by many people working in closely related areas, it defies attempts to produce a single universally accepted definition.

So long as agents lack a commonly shared definition, using the term reduces rather than increases the clarity of a conversation.

In the AI engineering space I think we may finally have settled on a widely enough accepted definition that we can now have productive conversations about them.

An LLM agent runs tools in a loop to achieve a goal. Let’s break that down.

The “tools in a loop” definition has been popular for a while—Anthropic in particular have settled on that one. This is the pattern baked into many LLM APIs as tools or function calls—the LLM is given the ability to request actions to be executed by its harness, and the outcome of those tools is fed back into the model so it can continue to reason through and solve the given problem.

“To achieve a goal” reflects that these are not infinite loops—there is a stopping condition.

I debated whether to specify “... a goal set by a user”. I decided that’s not a necessary part of this definition: we already have sub-agent patterns where another LLM sets the goal (see Claude Code and Claude Research).

There remains an almost unlimited set of alternative definitions: if you talk to people outside of the technical field of building with LLMs you’re still likely to encounter travel agent analogies or employee replacements or excitable use of the word “autonomous”. In those contexts it’s important to clarify the definition they are using in order to have a productive conversation.

But from now on, if a technical implementer tells me they are building an “agent” I’m going to assume they mean they are wiring up tools to an LLM in order to achieve goals using those tools in a bounded loop.

Some people might insist that agents have a memory. The “tools in a loop” model has a fundamental form of memory baked in: those tool calls are constructed as part of a conversation with the model, and the previous steps in that conversation provide short-term memory that’s essential for achieving the current specified goal.

If you want long-term memory the most promising way to implement it is with an extra set of tools!

Agents as human replacements is my least favorite definition

If you talk to non-technical business folk you may encounter a depressingly common alternative definition: agents as replacements for human staff. This often takes the form of “customer support agents”, but you’ll also see cases where people assume that there should be marketing agents, sales agents, accounting agents and more.

If someone surveys Fortune 500s about their “agent strategy” there’s a good chance that’s what is being implied. Good luck getting a clear, distinct answer from them to the question “what is an agent?” though!

This category of agent remains science fiction. If your agent strategy is to replace your human staff with some fuzzily defined AI system (most likely a system prompt and a collection of tools under the hood) you’re going to end up sorely disappointed.

That’s because there’s one key feature that remains unique to human staff: accountability. A human can take responsibility for their actions and learn from their mistakes. Putting an AI agent on a performance improvement plan makes no sense at all!

Amusingly enough, humans also have agency. They can form their own goals and intentions and act autonomously to achieve them—while taking accountability for those decisions. Despite the name, AI agents can do nothing of the sort.

This legendary 1979 IBM training slide says everything we need to know:

OpenAI need to get their story straight

The single biggest source of agent definition confusion I’m aware of is OpenAI themselves.

OpenAI CEO Sam Altman is fond of calling agents “AI systems that can do work for you independently”.

Back in July OpenAI launched a product feature called “ChatGPT agent” which is actually a browser automation system—toggle that option on in ChatGPT and it can launch a real web browser and use it to interact with web pages directly.

And in March OpenAI launched an Agents SDK with libraries in Python (openai-agents) and JavaScript (@openai/agents). This one is a much closer fit to the “tools in a loop” idea.

It may be too late for OpenAI to unify their definitions at this point. I’m going to ignore their various other definitions and stick with tools in a loop!

There’s already a meme for this

Josh Bickett tweeted this in November 2023:

What is an AI agent?

I guess I’ve climbed my way from the left side of that curve to the right.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

LLM agent AI定义 工具调用 目标导向 责任与自主性
相关文章