AI News 09月27日 03:06
新型CAMIA攻击揭示AI模型数据记忆漏洞
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

研究人员开发了一种名为CAMIA(Context-Aware Membership Inference Attack)的新型攻击方法,能够更有效地检测AI模型是否“记住”了训练数据中的敏感信息。该方法针对现代生成式AI的逐词生成特性,通过分析模型在不确定情境下的预测行为来识别数据记忆痕迹,准确率远超以往技术。CAMIA的出现凸显了AI模型在处理海量数据时潜在的隐私风险,尤其是在医疗和商业领域,可能导致患者信息或公司机密泄露。研究人员希望此项工作能推动更安全的AI技术发展,平衡AI效用与用户隐私。

💡 **CAMIA攻击揭示AI模型的数据记忆漏洞**:CAMIA(Context-Aware Membership Inference Attack)是一种新型攻击方法,能有效检测AI模型是否将训练数据中的敏感信息“记住”并可能泄露。它通过分析模型在生成文本过程中的行为,比以往方法更精准地识别数据记忆痕迹。

🧠 **基于上下文的不确定性检测记忆**:CAMIA的核心在于识别AI模型在面对不确定情境时对记忆的依赖程度。当模型预测下一个词语时,若在信息模糊的情况下仍表现出高度自信,这表明模型可能依赖于记忆而非泛化能力,从而揭示了潜在的隐私风险。

🚀 **针对生成式AI的突破性进展**:与以往主要针对分类模型的攻击不同,CAMIA专门设计用于现代逐词生成文本的AI模型。它能够追踪模型在生成过程中的不确定性变化,识别出其他方法可能忽略的、由真实记忆而非简单重复引起的微妙模式,显著提高了检测准确率。

🔒 **重要隐私安全警示**:CAMIA的发现强调了在训练大型AI模型时使用海量、未经筛选的数据所带来的隐私风险,尤其在医疗和商业敏感信息处理方面。这项工作呼吁行业开发更具隐私保护意识的技术,以平衡AI的实用性与用户隐私权。

Researchers have developed a new attack that reveals privacy vulnerabilities by determining whether your data was used to train AI models.

The method, named CAMIA (Context-Aware Membership Inference Attack), was developed by researchers from Brave and the National University of Singapore and is far more effective than previous attempts at probing the ‘memory’ of AI models.

There is growing concern of “data memorisation” in AI, where models inadvertently store and can potentially leak sensitive information from their training sets. In healthcare, a model trained on clinical notes could accidentally reveal sensitive patient information. For businesses, if internal emails were used in training, an attacker might be able to trick an LLM into reproducing private company communications.

Such privacy concerns have been amplified by recent announcements, such as LinkedIn’s plan to use user data to improve its generative AI models, raising questions about whether private content might surface in generated text.

To test for this leakage, security experts use Membership Inference Attacks, or MIAs. In simple terms, an MIA asks the model a critical question: “Did you see this example during training?”. If an attacker can reliably figure out the answer, it proves the model is leaking information about its training data, posing a direct privacy risk.

The core idea is that models often behave differently when processing data they were trained on compared to new, unseen data. MIAs are designed to systematically exploit these behavioural gaps.

Until now, most MIAs have been largely ineffective against modern generative AIs. This is because they were originally designed for simpler classification models that give a single output per input. LLMs, however, generate text token-by-token, with each new word being influenced by the words that came before it. This sequential process means that simply looking at the overall confidence for a block of text misses the moment-to-moment dynamics where leakage actually occurs.

The key insight behind the new CAMIA privacy attack is that an AI model’s memorisation is context-dependent. An AI model relies on memorisation most heavily when it’s uncertain about what to say next.

For example, given the prefix “Harry Potter is…written by… The world of Harry…”, in the example below from Brave, a model can easily guess the next token is “Potter” through generalisation, because the context provides strong clues.

In such a case, a confident prediction doesn’t indicate memorisation. However, if the prefix is simply “Harry,” predicting “Potter” becomes far more difficult without having memorised specific training sequences. A low-loss, high-confidence prediction in this ambiguous scenario is a much stronger indicator of memorisation.

CAMIA is the first privacy attack specifically tailored to exploit this generative nature of modern AI models. It tracks how the model’s uncertainty evolves during text generation, allowing it to measure how quickly the AI transitions from “guessing” to “confident recall”. By operating at the token level, it can adjust for situations where low uncertainty is caused by simple repetition and can identify the subtle patterns of true memorisation that other methods miss.

The researchers tested CAMIA on the MIMIR benchmark across several Pythia and GPT-Neo models. When attacking a 2.8B parameter Pythia model on the ArXiv dataset, CAMIA nearly doubled the detection accuracy of prior methods. It increased the true positive rate from 20.11% to 32.00% while maintaining a very low false positive rate of just 1%.

The attack framework is also computationally efficient. On a single A100 GPU, CAMIA can process 1,000 samples in approximately 38 minutes, making it a practical tool for auditing models.

This work reminds the AI industry about the privacy risks in training ever-larger models on vast, unfiltered datasets. The researchers hope their work will spur the development of more privacy-preserving techniques and contribute to ongoing efforts to balance the utility of AI with fundamental user privacy.

See also: Samsung benchmarks real productivity of enterprise AI models

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events, click here for more information.

AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here.

The post CAMIA privacy attack reveals what AI models memorise appeared first on AI News.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

CAMIA AI隐私 数据记忆 Membership Inference Attack AI安全 隐私泄露 生成式AI AI模型 Context-Aware Membership Inference Attack AI security Data memorization Privacy vulnerability Generative AI AI models
相关文章