Communications of the ACM - Artificial Intelligence 08月14日
Concerning the Responsible Use of AI in the U.S. Criminal Justice System
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

文章探讨了人工智能(AI)在刑事司法系统中的应用,强调了AI在提高效率、分析证据和潜在减少偏见方面的作用。然而,文章也深入分析了AI在司法领域面临的严峻挑战,特别是“黑箱”问题导致的透明度不足,这可能侵犯被告的权利。文章呼吁AI系统在司法决策中应具备高度透明性,清晰解释其运作逻辑和数据来源,并强调了对AI进行独立验证、定期审计以及对使用者进行培训的重要性,以确保AI的应用符合公平、准确和宪法的原则,不应牺牲司法体系的核心价值。

⚖️ AI在刑事司法中的潜在益处:AI能够极大地提高司法系统的效率,通过快速分析海量数据(如犯罪记录、证据和法律文件)来识别模式和联系,从而解放执法人员和司法人员,使其能专注于更具战略性的工作。此外,在理想情况下,AI可以通过排除与种族或收入相关的输入因素,有可能减少人为偏见,从而促进更公平的保释和量刑决策。

👁️ 透明度是AI在司法领域的核心挑战:文章指出,AI决策过程的不透明性是其在司法系统中最令人担忧的问题。当AI系统做出关乎个人自由的决策时(如保释、量刑或假释),被告、辩护律师和法官都应了解其决策逻辑、所用数据来源及其处理过程。一个不透明的AI系统相当于一个被告无法面对、无法交叉质证的“证人”,这与以个体为中心的司法体系相悖。

💡 推动AI透明化的具体措施:为解决不透明性问题,文章提出,AI系统应清晰描述其内部运作机制、使用的数学和逻辑方法,确保其逻辑可被决策者和受影响者理解,并能进行独立测试。这并非要求AI开源,而是要求其能够解释其决策依据,包括特征选择和训练过程。政府在采购AI系统时,应强制要求透明度。

📚 提升司法人员对AI的理解至关重要:文章强调,辩护律师、检察官和法官需要对AI系统的运作方式有基本的了解。这种知识有助于他们识别AI决策过程中的潜在偏见或错误,并形成有力的论证。为此,建议建立一个教育资源库,帮助司法人员学习AI训练数据、数据收集方式以及系统得出建议的路径,从而能够恰当使用、质疑或信任AI。

🔄 AI应用的持续监督与审计:AI系统的性能会随时间变化,且可能因地域差异而产生不同的偏见。因此,对AI风险评估算法进行定期审计至关重要,尤其是在新的司法项目实施后。文章以加州为例,要求对审前风险评估算法进行三年一次的审计,并建议将此要求推广到司法系统的所有层面,以确保其准确性和公平性。

📈 AI输出应量化且明确风险程度:文章批评了诸如“高风险”等抽象标签,认为它们容易引发刻板印象和误判。AI的输出应尽可能具体和量化,明确风险发生的概率以及风险的具体内容。这种量化有助于法官更准确地评估被告释放的风险,并使AI的决策更具可用性和问责性。

Artificial intelligence (AI) is advancing quickly and is being adopted in most industries. Using AI to draft an email message or check your grammar is typically not a cause for concern, but using it to make decisions that affect people’s lives is another matter. When constitutional rights are involved, as in the justice system, transparency is paramount.

During the Biden-Harris administration, Executive Order 14110 directed agencies to develop guidelines for acceptable uses and regulation of AI. Some of these uses, like summarizing and notetaking, will occur across the government. However, the nature and mission of each agency will create specific use cases for AI, such as monitoring threats at the DoD or simulating pollution at the EPA. The ways that agencies implement AI will change how they operate and may significantly impact the American people, and in some cases, people beyond our borders. One such area of application is the American criminal justice system.

The National Institute of Justice aims to “improve knowledge and understanding of crime and justice issues through science.” They conduct research on controversial topics, such as predictive policing and risk assessment, to improve the justice system at all levels: from surveillance to forensic evidence to bail reform. In response to this Executive Order, the NIJ requested input on how AI will affect each layer of the justice system, and how to develop a plan to implement AI in ways that are accurate, fair, and constitutional. This Opinion column is a summary of our response.

That Executive Order has since been rescinded. Nevertheless, Executive Order 13859 calls for safe testing of AI and to “foster public trust and confidence in AI technologies and protect civil liberties, privacy, and American values in their application.” How to protect these principles in the justice system, while using AI where appropriate, remains a vital question for us as computer scientists, for judges and policymakers, and for society as a whole.

Let’s start by defining AI. We have yet to find a definition that does not cause many computer scientists to disagree; here we choose to define it broadly. “AI,” as it is commonly understood, spans a wide range of technologies, from relatively simple algorithms that use statistical and machine learning techniques, to more advanced systems like deep neural networks and large language models. Some states have adopted the term “automated decision system” to include both current AI and simpler technologies, but we view these as belonging to one continuum.

AI has many obvious applications in the criminal justice system that can improve efficiency and effectiveness for members of law enforcement and the judiciary. AI can analyze criminal records, evidence, and legal documents much faster than humans and can identify patterns and connections that might otherwise be missed. Using AI to analyze legal documents and search for precedents can free up law enforcement officers and attorneys to focus on more strategic work. The danger that AI can perpetuate bias by being trained on biased data is well known; but at its best, AI may reduce bias by excluding factors correlated with race or income from its inputs, resulting in more equitable bail and sentencing decisions.

However, even when it satisfies statistical measures of accuracy and fairness, AI can pose significant concerns regarding procedural fairness, a vital aspect of the justice system. For one, AI is often not transparent. When an AI system makes a decision, it is not always clear—to defendants, judges, or other stakeholders—how the system came to its conclusions. This is especially concerning in bail, sentencing, and parole. Denying a citizen their liberty is one of the most fundamental and momentous actions a government can take. If this decision is supported in part by an AI system, then that citizen (and their defense counsel) needs to know what data were used by the AI, where the data came from, and the logic by which its recommendation was produced from this data.

The possibility that an opaque AI—which neither defendants, nor their attorneys, nor their judges understand—could play a role in major decisions about a person’s liberty is repugnant to our individualized justice system. An opaque system is an accuser the defendant cannot face; a witness they cannot cross-examine, presenting evidence they cannot contest.9 In our view, any AI system used for criminal justice must be transparent, as opposed to being a “black box” that produces outputs using a hidden process.

By “transparent” we do not necessarily mean making an AI system “open source,” that is, publishing the source code of its program. While this may be necessary in some domains, in general it is neither sufficient nor necessary. Transparency means a clear description of the internal workings of an AI system: the mathematical and logical methods it uses to produce its output. A system should be transparent enough so that its logic is intelligible to the decision makers it advises and the people it affects, and so that it can be independently tested to see if it performs with the accuracy and characteristics claimed by its developers.

Some tech companies complain that requiring this kind of transparency would violate their intellectual property rights, reveal trade secrets, or discourage innovation. We reply that opaque, proprietary AI systems might be acceptable in certain domains—recommending movies, translating speech, and so on—but they should not play a role in the justice system of a society that values individual rights and government accountability. If the government is considering the use of an AI system, the public has every right to require transparency, and this requirement should be implemented in procurement policies. The public might also demand an explanation, not just of what input factors and weights the AI system uses, but why it uses those factors and those weights: in technical terms, the feature selection and training process.

Beyond providing an explanation of an AI’s decision, attorneys for the defense and prosecution, as well as judges, need to possess a fundamental understanding of how these systems operate. This knowledge is crucial because it enables attorneys to identify potential biases or errors in the AI’s decision-making process and formulate cogent arguments for or against its findings. Similarly, judges can’t hope to fairly and accurately weigh an AI’s recommendations if they have no understanding of how it generates them. We recommend a repository of educational resources so that attorneys and judges can learn what data AIs are being trained on, how these data are being collected, and how these systems arrive at their recommendations. This understanding is vital for all parties to properly use AI, contest it, and trust it when trust is warranted.

When relying on AI systems to make judicial recommendations, such as in sentencing or probation decisions, judges should always view these systems’ results through a critical lens. AI systems are often subject to biases and operate based on limited and/or partial data. An AI system typically uses prior convictions and prior failures to appear when recommending a bail decision. It treats each defendant as a member of a group, namely those with similar criminal records. But it has no access to individualized facts about a defendant that might make them more, or less, dangerous than others in this group. The judge should be open to additional arguments made by the defense and prosecution, whose job it is to present information not considered by the AI.

Transparency also applies to the decision-making process: not just the AI itself, but how humans use it. How and when AI systems are used in courtrooms must be standardized and fully explained, and all parties involved in a particular case must be informed when the results of an AI contribute to a decision. Standardizing procedures regarding when AI can and should be used will also improve the processes of auditing and evaluating the use of AI in the criminal justice system. Demanding transparency from AI can make the justice system as a whole more transparent.

AI systems—and the social science and data science behind them—can also help advance policy discussions about crime and justice. But to do this, risk assessment systems should not lump all types and levels of crime together.6 Unfortunately, while they typically distinguish violent from non-violent charges, many risk assessments in use today do not make a distinction between felony and misdemeanor arrests, even though many legal scholars have pointed out the need to do so.3,9,11,12 Of course, a judge might feel that the risk of a misdemeanor is enough to justify pretrial detention. But if an AI lumps all kinds of risk together, it does not help the judge consider the “nature and seriousness of the danger to any person or the community that would be posed by the person’s release” as the law requires.10

Worse, many risk assessments provide the judge with an abstract label like “high risk,” as opposed to a quantitative estimate of the probability of rearrest. This points out the need for another kind of transparency: judges should know what an AI’s output actually means. Phrases like “high risk” allow our preconceptions and stereotypes to run wild. Psychologists have found that human decision-makers often overestimate the probability of bad events: mock jurors given categorical labels like “high risk” greatly overestimate the corresponding probabilities.5 To be useful and accountable, an AI system should make predictions that are as specific and quantitative as possible: it should say how much risk, and risk of what.

Moreover, the meaning of an AI’s output might not translate from one jurisdiction to another.7 Due to differences in demographics, policing, and many other factors, an AI might display racial bias in one state or city even though it is unbiased in another. The performance of AI systems can also change over time: for instance, new diversion programs might reduce the level of risk some defendants pose. In that case, using a risk assessment trained on data before those programs were implemented will overestimate risk.4 For this reason, it is vital to perform periodic audits of risk assessment algorithms in each jurisdiction they are used, whenever there is enough data to obtain good statistics on their accuracy and various measures of bias.8 California currently requires that pretrial risk assessment algorithms be audited every three years.1 Other states should follow suit and extend similar requirements to AIs used at every layer of the justice system including sentencing, parole, prison classification, and forensic evidence.

Finally, an AI system should provide information about its uncertainty and the confidence level of its predictions. If a defendant has an unusual criminal record, with very few similar defendants in the training data, an AI should report that its output is less certain. If an AI model generates a prediction or determination that is highly unusual or that humans struggle to justify, those results should not be considered by the court as evidence in support of or in opposition to a defendant. Unusual edge cases where the AI model makes confusing predictions should also be reported to the developers to improve the model in the future. Developing principled techniques for calculating confidence levels is an active area of research, as is developing effective ways to communicate uncertainty to human decision makers.

While we focused here on risk assessments, many of these principles apply just as well to the use of AI to generate forensic evidence, such as facial recognition and probabilistic genotyping of DNA samples. Despite the weight given to this kind of evidence by judges and juries, many of these software products have never been independently validated. Developers typically oppose revealing anything about their inner workings unless forced to do so by a judge, again threatening the right of defendants to confront evidence against them and cross-examine witnesses.14 Indeed, systems that have been used for DNA evidence in thousands of cases have turned out to have important flaws when independent studies are finally carried out.6 We agree that judges and legislatures should require independent verification and validation of these systems, just as we do for software in other high-risk areas like medicine or power plant management.2

Where does this leave us? AI can help us make consequential decisions, and can make human decision-making more accountable and fair. But only if it is transparent, so that those affected by it—and the decision makers advised by it—understand what data the AI uses, what it does with this data, and what mistakes it can make.

The law is a dynamic system. It has been in development for thousands of years, since the first humans disagreed and found resolution. A key pillar of a just legal system is a human-centered design where evidence and arguments can be contested and disputed, and where the system itself can be revised to make its outcomes and processes more just. In our society we believe this system should be accountable, fair, and transparent, and open to the unique characteristics of individuals. We should not throw these values away in pursuit of technological improvements.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

人工智能 刑事司法 透明度 AI伦理 风险评估
相关文章