少点错误 10月10日 00:37
将模糊的道德情感转化为具体政策的四个关键问题
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文探讨了如何将抽象的道德情感转化为可执行的政策提案。作者在LessOnline的一次演讲中,通过提出四个核心问题,帮助听众细化他们的政策想法,并获得了积极的反馈。这四个关键问题是:谁被要求采取行动?他们被要求做什么?谁来评估是否已完成?如果不完成会有什么后果?文章详细阐述了每个问题的重要性,并强调了具体性的价值,包括明确主体、行动内容、评估机制和违规后果。作者认为,政策的细节是改变行为的关键,越具体越能有效推动政策的落地与执行,并以不同领域的例子进行了说明。

🎯 **明确行动主体:** 政策的成功执行首先在于清晰界定谁是必须遵守规则的个体或组织。例如,笼养鸡蛋政策应具体到“年销售超过100个鸡蛋的任何人”,而非笼统的“肉类行业”,以避免不必要的争议,将精力集中在核心理念的推广上。在AI领域,明确“前沿AI开发者”的标准,如通过训练FLOPs阈值或性能基准,能有效避免政策适用于不相关对象,或被应覆盖对象规避。

📝 **界定具体行动:** 政策要求必须包含一个清晰的、可执行的动词,并辅以具体的直接宾语。例如,“超市必须储备至少两种无笼鸡蛋”比“超市应优先考虑低残忍品牌”更具操作性。“父母必须在孩子2岁前接种麻疹疫苗”是明确的政策,而“儿童应免于可预防疾病”仅是目标。若直接宾语含义不清(如AI“模型卡”),则需进一步明确其内容、来源、准确性要求及更新频率。

🧐 **设立评估机制:** 政策的有效性依赖于一个明确的评估者和评估标准。这可以是一个司法或行政机构,也可以是行业协会、贸易团体或公众利益组织。例如,海洋管理委员会(MSC)负责评估其可持续海产品政策的遵守情况。一个有效的评估机制能确保政策被认真对待,并为后续的后果追究提供依据。AI安全政策领域目前在这方面存在显著不足。

⚖️ **明确违规后果:** 政策的约束力在于其明确的、可预期的后果。政策提供的是一种条件性承诺:若采取行动X,则将产生后果Y。例如,未接种疫苗的儿童可能被学校拒之门外,违反MSC规定的渔业公司将失去使用其蓝色标签的资格。后果的明确性和可信度是促使政策被遵守的关键,需要考虑实际执行中的执法情况,而非仅停留在纸面规定。

Published on October 9, 2025 4:30 PM GMT

I gave a talk this year at LessOnline about how to make AI governance research more relevant to policymakers (videos available here), and the subtopic from that talk that got the most audience engagement was how to turn a vague moral sentiment into a concrete policy proposal. During the talk, I helped people refine their proposals by asking them a series of questions about what they wanted to do and why. This works better in a conversational setting, but for those who couldn’t attend, I wanted to write up a list of some of the most common and/or important questions, so that everyone can try to make their policy proposals more concrete. 

This is a skill that can (and should) be used in all policy areas, not just in AI, so after I explain the questions, I’ll include a variety of examples from different fields.

The four questions you need to answer to make most policy proposals reasonably concrete are:

    Who is required to act?What are they required to do?Who will evaluate whether it’s been done?What is the consequence if they haven’t done it?

QUESTION 1: Who is required to act?

The answer to the first question needs to be framed in such a way that it clears up most of the edge cases and will create common knowledge about who is and isn’t subject to your policy. For example, if you want to make eggs cage-free, a really crisp answer to “who is required to act?” is something like “anyone who sells more than 100 chicken eggs per year.” An answer that’s somewhat looser but still acceptable is “large commercial chicken farmers.” If your answer is instead a very broad concept like “the meat industry,” then that’s rhetorically satisfying, but you can and will spark an unwanted debate about who qualifies as “small family farmers” and whether they should have to follow your cage-free policy. The more time you spend arguing about the exact borders of your policy, the less time you can spend explaining why your core idea is a good idea.

Being specific here is surprisingly helpful, especially if you’re willing to give some ground. If you just say you want to regulate “the meat industry” and refuse to clarify any further, then small family farmers will never trust you. On the other hand, if you start with “100 chicken eggs per year,” and the local farmers mock you for setting such a low threshold, then you can offer to add a few zeroes, or ask the farmers what they think the threshold should be. This makes the discussion usefully concrete. If the family farmers would stop opposing you in exchange for raising the threshold to 100,000 chicken eggs per year, then you’ve found a useful political compromise. If there’s no even halfway-sane number they would agree to, then their true concerns probably aren’t really about family farms – at least you know you’re not missing out on a potential deal, and if you can find even a couple of local farmers to endorse your higher threshold, then you can undermine your opponents’ credibility.

Similarly, if you want frontier AI developers to publish model cards, it’s really helpful if you can crisply define who counts as a frontier AI developer – perhaps by using a FLOP threshold (e.g. more than 10^26 FLOPs used in training), or with reference to a performance benchmark, or just by trying to count and name the top 5 or top 10 developers. Otherwise, you can and will waste a ton of energy on type I and type II errors – you’ll get people who you never wanted to be subject to your policy coming out and complaining that model cards don’t make sense for them, and then you’ll get people who the policy should have covered who quietly ignore the policy and then later claim they were confused and didn’t realize the policy applied to them. 

Defining exactly who has to follow your rules is challenging and you’re likely to mis-classify at least a few people, but failing to offer a definition is worse, because then your proposal isn’t clear enough for people to understand what they’d be agreeing to, and you will lose political momentum while people argue and complain about that.

QUESTION 2: What are they required to do?

The answer to the second question needs to include at least one active verb. When you say what people are required to do, you can require people to “test,” “publish,” “vaccinate,” “buy,” or “install” something. You can’t require people to “be safe” or “be kind” or “avoid suffering” – or, at least, you can’t do that using any of the legal and policy tools common in English-speaking countries.

After the active verb, you usually need a direct object, ideally one that comes with numbers or at least a very concrete description of who or what will be acted upon. “Grocery stores must stock at least two brands of cage-free eggs” is a concrete policy; “Grocery stores should prioritize lower-cruelty brands” is merely a vague expression of desire. “Parents must vaccinate their children for measles by age 2” is a firm policy; “children should be free from preventable disease” is merely a goal.

If your direct object isn’t something with a meaning that everyone will agree on, then you need a crisp definition for your direct object, too. For example, what exactly is an AI “model card?” If you want to require AI developers to publish model cards, you might need to say what information is required to be in a model card, where this information will come from, what level of accuracy is required, or how often the model card needs to be updated. Differing levels of detail are appropriate for different contexts; you would not, for example, put all of that information on a poster that you’re waving at a protest. However, if you yourself don’t have a satisfying answer to those questions that you can share on request, then you don’t yet have a concrete policy.

QUESTION 3: Who will evaluate whether it’s been done?

The answer to the third question typically needs to specify some kind of institution that will make the decision about whether a policy requirement has been satisfied, as well as what criteria they’ll use to make that decision. The institution could be a judicial or administrative court, especially if the policy is a legal requirement, but it doesn’t have to be – often the institution making the decision is some kind of industry association, trade group, or public interest group. For example, the Marine Stewardship Council makes decisions about whether its policies on responsible seafood are being followed by each supplier. Your local school or university might make decisions about whether a child has received enough vaccinations to attend classes. The Open Source Initiative claims the authority to decide whether a given piece of software has complied with its policy to allow everyone to freely use and modify that software, although my understanding is that this authority is somewhat contested.

Notably, I can’t think of an example of a decision-making body that rules on AI safety policy requirements, because there is a huge deficit in the AI safety field’s ability to answer this third question. Zach Stein-Perlman over at AI Lab Watch has done heroic work in collecting data about what AI developers are and aren’t doing, but it doesn’t seem to be tied to anything as crisp as a decision-making process; there’s no single “yes or no” question that gets answered by his evaluations, nor would AI developers expect to suffer any particular consequences if AI Lab Watch decided that the answer to one of their questions was “no.”

QUESTION 4: What is the consequence if they haven’t done it?

This brings me to the fourth question: who cares? Or, more literally, what is the consequence that the people regulated by your policy can expect to suffer if they fail to comply with your required actions? Policies can’t and don’t magically compel people to follow them; instead, they offer a conditional consequence: if you take action X, then we will ensure that consequence Y follows. Depending on how tempting action X is and how credible the thread of consequence Y is, some people might change their behavior in response to this threat.

In the examples we’ve discussed above, a school’s consequence for not vaccinating your children might be that those children are turned away at the door and not allowed to attend classes. The Marine Stewardship Council’s consequence for fishing endangered species with drag nets is that they won’t let you put their fancy blue label on your product, and if you try to put the label there anyway, they’ll sue you for fraud. If your policy involves some kind of grant funding or subsidy, then the consequence of not following the policy could be as simple as not receiving the money.

The state government of California’s consequence if you put chickens in cages and then sell their eggs to grocery stores is that they will (theoretically) send government inspectors to your farm, notice that your farm has cages, order you to remove those cages, and then fine you, shut down your business, or put you in jail if you disobey that order. 

It’s worth asking yourself how many of the steps in this process you really expect to occur – what matters is not just what’s written down on paper about policies, but also how those policies are likely to be enforced (or not) in practice. In America, driving a car over the speed limit is technically a criminal infraction that can carry jail time, but in practice the number of people who go to jail solely for speeding is so low as to be statistically zero; if you’re not drunk and you don’t injure others, then it’s incredibly unlikely that an American judge will sentence you to jail time purely for speeding.

If you want to make sure that violating your policy will come with meaningful consequences that could actually motivate people to make large or unpleasant changes to their behavior, then you may need to go back to step 3 and define the evaluation process in more detail. It’s usually not enough to say “the government will decide whether you’ve published a model card” – you may need, instead, to say something like “the office of AI Model Review, which reports directly to the Secretary of Commerce, will evaluate each AI company that has at least $100 million in assets at least once each calendar year and fill out this checklist of criteria about whether their model cards are adequate.”

Specifics are the lifeblood of policy: the more you can be concrete about who will do what to whom, and about when, where, and how they will do it, and about what will happen if they don’t, then the more likely you are to change other people’s behavior.



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

政策制定 AI治理 政策工具 可执行性 Policy Making AI Governance Policy Tools Actionability
相关文章