Communications of the ACM - Artificial Intelligence 4小时前
AI安全会议呼吁全球合作,共同应对先进AI风险
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

联合国大会期间举行的AI安全连接(AISC)会议汇聚了政策制定者、科学家和非政府组织领袖,共同探讨先进人工智能对全球安全与稳定构成的挑战。会议强调了跨部门和国际合作的重要性,以有效缓解AI带来的潜在风险。专家们就“红线”设定、私营部门的风险管理策略以及非营利AI研究机构的作用等议题展开深入讨论,旨在为快速发展的AI技术建立有效的治理框架。会议还特别关注如何赢得公众信任,确保AI技术的发展符合人权和人道主义法,并促进全球范围内的公平与民主。

🌐 **全球合作与治理框架**:AI安全连接(AISC)会议在联合国大会期间举行,旨在促进政策制定者、科学家、非政府组织、学术界和私营部门之间的对话,共同应对先进AI系统带来的全球安全与稳定挑战。会议强调建立国际性和跨部门的合作机制,以有效管理AI风险,并讨论了AI治理的必要性,包括设定“红线”、私营部门的风险管理以及非营利AI研究机构的作用。

🤝 **赢得公众信任与伦理规范**:多国代表强调,AI的部署必须始终遵守国际人权和人道主义法。发展AI不应以牺牲关键的监督措施为代价,而应致力于确保AI技术为全人类带来繁荣。公众信任是AI应用的关键,必须通过在道德和负责任的方式下开发和部署AI来赢得,并倡导AI成为促进民主化进程的工具,包括透明和广泛可及的资源。

⚠️ **AI风险与“红线”设定**:专家指出了AI发展的潜在灾难性后果,包括市场和权力过度集中、恶意行为者的滥用(如网络攻击、虚假信息传播、生化武器制造)以及AI超越人类控制的可能性。会议深入探讨了设定AI“红线”的必要性,即对AI能力施加硬性限制,并呼吁科学家和政治家共同参与,通过国际合作和协议来界定可接受的边界,借鉴核技术发展的成功经验。

🔬 **技术框架与非营利机构的作用**:会议还关注了通用模型(如LLMs)的安全技术和制度框架。私营部门的代表介绍了其在识别和管理生物及网络安全风险方面的努力。同时,非营利AI安全研究所(AISIs)的兴起及其全球网络被视为应对AI安全盲点、促进外部研究审查的重要力量,这些机构正通过合作网络共同提升AI安全研究的整体水平。

In late September, policymakers, scientists, and leaders of non-governmental organizations met during the 80th session of the United Nations General Assembly to discuss a pressing contemporary issue: the implications of artificial intelligence (AI) for global safety and security.

This second meeting of AI Safety Connect (AISC) addressed the growing need for international and cross-sector collaboration to mitigate the risks associated with advanced AI systems. Keynote speeches from ACM Turing laureate Yoshua Bengio and University of California, Santa Cruz Professor of Physics Anthony Aguirre outlined the hazards of agentic AI, while panels on “red lines,” private sector approaches to risk, and the rise of nonprofit AI research institutes shone a light on the challenges of AI governance in an age of rapid innovation.

AISC was started by Nicolas Miailhe and Cyrus Hodes, co-founders of The Future Society. Miailhe and Hodes envision AISC as an event series to facilitate dialogue between nonprofit organizations, academic and industry laboratories, governments, and investors. AISC’s first meeting took place as the Artificial Intelligence Action Summit in Paris in early February 2025 and included speakers from a global array of universities, nonprofits, and tech companies.

The second meeting welcomed NGOs and governments in addition to academia, nonprofits, and the private sector, with panels featuring representatives from the United Nations Development Programme (UNDP) and the governments of Singapore, Canada, and Brazil. As Hodes told participants, AISC aims to foster “shared understanding” and establish “responsibility mechanisms” across a broad range of sectors.

Earning Trust Across the Globe

The event opened with remarks from political leaders and representatives from the UNDP and International Telecommunication Union. Speakers agreed that AI innovation need not come at the cost of critical oversight measures.

Elina Valtonen, Finland’s Minister of Foreign Affairs, said AI “must always be deployed in accordance with international human rights and humanitarian law.”

Minyoung Han, Korea’s Director General for Climate Change, Energy, Environment and Scientific Affairs, described Korea’s efforts to “ensure that AI drives prosperity for all,” primarily by emphasizing safety as a cornerstone of research and development.

Building on the theme of civic innovation, Josephine Teo, Singapore’s Minister for Digital Development and Information, declared that public trust “has to be earned” rather than assumed. “People need to be assured that the AI we’re letting them use is developed and deployed in an ethical manner,” Teo said.

S. Krishnan, India’s Secretary for the Ministry of Electronics and Information Technology, echoed this sentiment, asserting that “AI should serve as a force for democratization” within and across national borders. Krishnan’s vision for democratic AI emphasizes transparent and widely-accessible resources, including databases, algorithms, and public education.

The Stakes of Unchecked Development

Following opening remarks, Bengio and Aguirre took to the stage to share their perspectives on the stakes of an AI-driven future.

Bengio outlined three major targets for safety research: market and power concentration; abuse by bad actors; and the potential for AI to advance beyond the threshold of human control. Each of these areas is associated with potentially catastrophic outcomes, he said. The first seeks to prevent “radically disproportionate” power discrepancies that could ensue from unregulated AI. The second addresses AI’s potential to be used in cybersecurity attacks, disinformation campaigns, and the construction of chemical and biological weapons. The third considers the current limitations of scientific research; as Bengio pointed out, today’s researchers “don’t know how to design AI that won’t harm people.”

Aguirre’s keynote took a similar warning tone. “Certain systems should not be built or deployed,” he said, voicing his support for the implementation of “red lines,” or hard constraints on AI agent capabilities, which was the subject of Nobel Peace Prize Laureate Maria Ressa’s opening address to the General Assembly. According to Aguirre, scientists and politicians should see themselves as equal participants in the red lines agenda. While scientists must commit to sharing their knowledge with politicians, politicians must work towards agreements on acceptable boundaries. This is not a minority position because the global call for red lines represents “a very strong consensus,” he said.

While the prospect of out-of-control AI may seem frightening, Aguirre emphasized that open dialogue can lead to a future where AI’s benefits don’t come with severe costs. “I think there’s a tendency to consider the development of AI like a train on a single track, and our only option is merely to decide how fast or slow we steam ahead,” he said. To the contrary, he said the trajectory of AI is “a proliferation of paths” from which to choose.

Red Lines, Model Oversight, and Safety Institutes

Aguirre continued the red lines discussion with Niki Iliadis, Director of Global AI Governance at The Future Society, and Charbel-Raphaël Segerie, Executive Director of France’s Center for AI Security. Iliadis posed the question of whether red lines would “stop innovation,” as many fear. Aguirre responded that “Unregulated, unbridled AI development” is not equivalent to progress. “We can choose a safer, more responsible, pro-human and trustworthy way,” he argued, drawing comparisons to the history of nuclear technology, where global actors pulled together to prevent the direst outcomes. “International diplomacy has shown in the past that it can do wonders,” Aguirre said. “And I think this is an opportunity for us to do it again.”

The following panel, “Safety on the Frontier of AI,” looked at technical and institutional frameworks to promote safety in general purpose models, such as the LLMs in wide use today. Chris Meserole, Executive Director of the Frontier Model Forum (FMF), described how the organization consults with its member firms—Amazon, Anthropic, Google, Meta, Microsoft, and OpenAI—to identify thresholds of unacceptable risk, with an emphasis on biological and cybersecurity hazards. Coco Zhou, Senior Vice President and Partner of China’s UCloud, picked up on the theme of cybersecurity, saying, “The primary concerns of our customers right now is the leakage of private data,” which is why UCloud designs its cloud-based supercomputing services to accommodate both privacy and open knowledge sharing. Natasha Crampton, Microsoft’s first Chief Responsible AI Officer and a director at the FMF, framed Meserole and Zhou’s participation in this conversation as increasingly necessary, describing forums like AISC and the UN’s Global Dialogue on AI Governance as “young mechanisms” that will require input in the future from “civil society, academia, industry, and government.”

The closing panel focused on the rise of nonprofit AI safety institutes, or AISIs. In response to a question about priority-setting, Deval Pandya, Vice President of AI Engineering at the Vector Institute and an advisor to the Canadian AI Safety Institute (CASI), said that CASI was focusing on synthetic content because the topic hadn’t received due attention from similar organizations. Pandya’s response pointed to the growing links between AISIs across the world: as new organizations emerge, they’re joining forces to identify areas of weakness in their respective agendas. On top of helping scientists account for blind spots, AISI networks allow organizations to find partners for external research review, a priority voiced by Wan Sie Lee, Director of Singapore’s Infocomm Media Development Authority, and Qian Xiao, Vice Dean of Tsinghua University’s Institute for AI International Governance.

While the risks associated with AI are serious, AISC represents a growing community of stakeholders committed to working together to address them. This year saw the publication of the first comprehensive overview of current AI safety literature, which included input from over 100 researchers from 33 countries.

Next year’s India AI Impact Summit, which will host AISC’s third event, aims at broadening the safety community’s international horizons. During his closing speech, Miailhe shared that he and AISC co-founder Hodes “felt a duty to exercise social responsibility” by “building bridges across communities” in this space.

Emma Stamm is a writer, researcher, and educator based in New York City. Her interests are in philosophy and the social study of science and technology.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI安全 人工智能治理 国际合作 AI风险 红线 AI伦理 AISC AI Safety AI Governance International Cooperation AI Risks Red Lines AI Ethics
相关文章