OpenMined Blog 09月12日
利用隐私增强技术实现AI系统外部审查
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文介绍了隐私增强技术(PETs)如何助力对AI系统进行外部审查,同时保障安全、隐私和知识产权。文章指出,AI系统的广泛影响引发了对个性化误导信息、算法偏见和网络攻击等风险的担忧,而AI公司对外部访问的顾虑阻碍了有效的独立审查。OpenMined开发的端到端技术基础设施,利用PySyft等工具,通过安全计算技术,使得研究人员可以在不暴露敏感信息的情况下提出问题并获得模型答案。文章还列举了Christchurch Call Initiative和英国AI安全研究所的成功案例,证明了该技术在实际应用中的有效性,并展望了未来在分析推荐系统、检测有害内容等方面的应用前景,强调PETs为AI治理提供了关键的透明度和问责机制。

🔒 **AI系统的广泛影响与审查挑战**:随着Facebook新闻推送算法和ChatGPT等大型语言模型在全球范围内的应用,其对数十亿用户的影响带来了个性化误导信息、算法偏见和针对关键基础设施的新型网络攻击等潜在风险。尽管专家普遍认同对关键AI系统进行独立外部审查的必要性,但AI公司对用户数据隐私、系统安全漏洞以及知识产权保护的担忧,使得实际实施面临巨大障碍。

💡 **OpenMined的隐私保护技术基础设施**:OpenMined开发了一套端到端的技术基础设施,通过其核心软件库PySyft,支持一种基本的工作流程:研究人员远程向模型所有者提出问题,模型所有者批准后,研究人员在不了解专有系统其他信息的情况下获得答案。该基础设施集成了安全计算(如安全多方计算、零知识证明)、联邦学习和差分隐私等成熟技术,确保了审查过程的隐私性。

🚀 **实际应用案例与未来展望**:文章列举了两个成功案例:Christchurch Call Initiative on Algorithmic Outcomes,首次成功利用集成隐私保护工具对社交媒体平台进行外部研究,无需查看原始数据;以及与英国AI安全研究所和Anthropic合作,在保护生物数据集和AI模型权重的“互不透露”前提下,对前沿AI模型进行安全评估。未来,该技术还可用于分析推荐系统的党派倾向、检测聊天机器人的有害回应,并保护研究人员的代码和数据隐私。

Researchers from Georgetown University’s Center for Security and Emerging Technology (CSET) have detailed how new privacy-enhancing technologies (PETs) can facilitate external scrutiny of AI systems without compromising security, privacy, or intellectual property.

Read the full paper here. The paper is summarized below.  

The Challenge of AI Scrutiny

AI systems like Facebook’s newsfeed recommendation algorithm and large language models such as ChatGPT now operate globally, affecting billions of users. This widespread impact has raised legitimate concerns about potential risks including:

While experts across the field agree on the necessity of independent external scrutiny for consequential AI systems, practical implementation has faced significant hurdles. AI companies have been reluctant to grant access to external researchers due to concerns about:

Technical AI Governance Infrastructure

OpenMined has developed end-to-end technical infrastructure that enables privacy-preserving audits of AI systems. The core software library, PySyft, works to support a fundamental workflow:

    A researcher remotely proposes questions to a model ownerThe model owner approves the researchers’ questionsThe researcher receives answers without learning anything else about the proprietary systems

This infrastructure leverages well-established technologies, including secure enclaves, secure multi-party computation, zero-knowledge proofs, federated learning, and differential privacy.

Real-World Success Stories

Case Study 1: The Christchurch Call Initiative on Algorithmic Outcomes

Following the 2019 terrorist shootings in New Zealand, the Christchurch Call coalition launched the Initiative on Algorithmic Outcomes (CCIAO). In 2023, OpenMined collaborated with the initiative in what became the first-ever use of an integrated privacy-preserving access tool for external research on social media platforms.

This pilot program demonstrated that external researchers could leverage private assets (video impression data) to investigate algorithmic impacts without seeing the raw data, precluding the need for extensive legal review and making the process significantly less burdensome.

Case Study 2: UK AI Safety Institute

More recently, OpenMined partnered with the UK AI Safety Institute and Anthropic to trial safety evaluations of frontier AI models while maintaining privacy. The setup enabled “mutual secrecy,” with the contents of a biology dataset remaining private to UK AISI while the AI model weights remained private to Anthropic.

This successful demonstration proved that government entities and AI companies can negotiate and enforce shared governance over model evaluations.

Future Directions

OpenMined’s technical infrastructure can apply to various scrutiny paradigms, including:

Currently, we are developing features to allow researchers to keep their code private in addition to their data, further empowering sophisticated research with less oversight from model owners.

Conclusion

External scrutiny of AI systems provides crucial transparency into AI development and should be an integral component of AI governance. With OpenMined’s privacy-preserving technical solutions now successfully deployed in real-world governance scenarios, AI companies can no longer use privacy, security, and IP as conclusive excuses for refusing access to external researchers.

These innovative approaches deserve further exploration and support from the AI governance community as we work toward more transparent and accountable AI systems.

The post Enabling External Scrutiny of AI Systems with PETs appeared first on OpenMined.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI Privacy-Enhancing Technologies PETs AI Governance External Scrutiny OpenMined PySyft AI Security AI Ethics 人工智能 隐私增强技术 AI治理 外部审查 AI安全 AI伦理
相关文章