VentureBeat 10月14日 06:46
模型微调新方法,规避“灾难性遗忘”
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

一项来自伊利诺伊大学厄巴纳-香槟分校的研究提出了一种新的模型再训练方法,旨在解决大型语言模型(LLM)在微调过程中出现的“灾难性遗忘”问题。该方法通过仅微调模型的部分狭窄区域来避免遗忘已学知识,从而降低计算成本。研究团队认为,所谓的“遗忘”并非真正的记忆丢失,而是输出偏差漂移所致。他们发现,仅微调自注意力投影层(SA Proj)可以有效学习新任务,同时保持原有任务的性能,而微调多层感知机(MLP)则可能导致输出偏差。通过调整MLP的特定部分并冻结其他部分,可以在实现新任务学习的同时,最大程度地减少遗忘。

💡 **规避模型“灾难性遗忘”的新方法:** 研究提出了一种针对大型语言模型(LLM)微调的新策略,旨在解决模型在学习新任务时丢失原有知识的“灾难性遗忘”问题。该方法的核心在于,通过选择性地微调模型内部的特定组件,而非对整个模型进行大规模再训练,来达到更新模型能力并保留已有知识的目的,从而显著降低了计算成本和时间投入。

🎯 **“遗忘”实为输出偏差:** 研究团队挑战了“灾难性遗忘”是真实记忆丢失的传统观点,提出这实际上是由于任务分布变化导致的输出偏差漂移。他们观察到,模型在执行特定新任务后,虽然在某些基准测试上性能下降,但在其他相关任务上却能恢复,这表明模型并非真正“遗忘”了信息,而是其输出的概率分布发生了改变。

🔬 **关键在于狭窄区域的微调:** 研究重点关注了模型内部的多层感知机(MLP)和自注意力投影(SA Proj)层。实验发现,仅微调自注意力投影层能够有效学习新任务,且在原有任务上的性能几乎不受影响。相比之下,微调整个MLP层虽然能提升新任务表现,但却容易导致输出偏差和原有任务性能的下降。研究者进一步发现,通过调整MLP的特定投影(如up/gating projections)并冻结其他部分,可以在实现新任务学习的同时,显著减少遗忘现象。

💰 **企业应用价值:** 这种精细化的微调方法对于企业而言具有重要意义。通过精确控制微调的范围,可以避免昂贵的、耗时且碳排放量巨大的全面模型再训练。这不仅节省了企业在模型开发和更新上的成本,还提高了模型更新的可控性和可重复性,使LLM能够更高效、更经济地适应不断变化的应用需求。

Enterprises often find that when they fine-tune models, one effective approach to making a large language model (LLM) fit for purpose and grounded in data is to have the model lose some of its abilities. After fine-tuning, some models “forget” how to perform certain tasks or other tasks they already learned. 

Research from the University of Illinois Urbana-Champaign proposes a new method for retraining models that avoids “catastrophic forgetting,” in which the model loses some of its prior knowledge. The paper focuses on two specific LLMs that generate responses from images: LLaVA and Qwen 2.5-VL.

The approach encourages enterprises to retrain only narrow parts of an LLM to avoid retraining the entire model and incurring a significant increase in compute costs. The team claims that catastrophic forgetting isn’t true memory loss, but rather a side effect of bias drift. 

“Training a new LMM can cost millions of dollars, weeks of time, and emit hundreds of tons of CO2, so finding ways to more efficiently and effectively update existing models is a pressing concern,” the team wrote in the paper. “Guided by this result, we explore tuning recipes that preserve learning while limiting output shift.”

The researchers focused on a multi-layer perceptron (MLP), the model's internal decision-making component. 

Catastrophic forgetting 

The researchers wanted first to verify the existence and the cause of catastrophic forgetting in models. 

To do this, they created a set of target tasks for the models to complete. The models were then fine-tuned and evaluated to determine whether they led to substantial forgetting. But as the process went on, the researchers found that the models were recovering some of their abilities. 

“We also noticed a surprising result, that the model performance would drop significantly in held out benchmarks after training on the counting task, it would mostly recover on PathVQA, another specialized task that is not well represented in the benchmarks,” they said. “Meanwhile, while performing the forgetting mitigation experiments, we also tried separately tuning only the self-attention projection (SA Proj) or MLP layers, motivated by the finding that tuning only the LLM was generally better than tuning the full model. This led to another very surprising result – that tuning only self-attention projection layers led to very good learning of the target tasks with no drop in performance in held out tasks, even after training all five target tasks in a sequence.”

The researchers said they believe that “what looks like forgetting or interference after fine-tuning on a narrow target task is actually bias in the output distribution due to the task distribution shift.”

Narrow retraining

That finding turned out to be the key to the experiment. The researchers noted that tuning the MLP increases the likelihood of “outputting numeric tokens and a highly correlated drop in held out task accuracy.” What it showed is that a model forgetting some of its knowledge is only temporary and not a long-term matter. 

“To avoid biasing the output distribution, we tune the MLP up/gating projections while keeping the down projection frozen, and find that it achieves similar learning to full MLP tuning with little forgetting,” the researchers said. 

This allows for a more straightforward and more reproducible method for fine-tuning a model. 

By focusing on a narrow segment of the model, rather than a wholesale retraining, enterprises can cut compute costs. It also allows better control of output drift. 

However, the research focuses only on two models, specifically those dealing with vision and language. The researchers noted that due to limited resources, they are unable to try the experiment with other models.

Their findings, however, can be extended to other LLMs, especially for different modalities. 

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

LLM 模型微调 灾难性遗忘 AI 自然语言处理 LLM Model Fine-tuning Catastrophic Forgetting AI Natural Language Processing
相关文章