ChatPaper.aiChatPaper

ORBIT:透過源頭調控合併保留生成式檢索中的基礎語言能力

ORBIT: Preserving Foundational Language Capabilities in GenRetrieval via Origin-Regulated Merging

May 12, 2026
作者: Neha Verma, Nikhil Mehta, Shao-Chuan Wang, Naijing Zhang, Alicia Tsai, Li Wei, Lukasz Heldt, Lichan Hong, Ed Chi, Xinyang Yi
cs.AI

摘要

儘管大型語言模型(LLM)的開發進展迅速,但針對特定任務進行微調時,往往會導致其通用的語言推理能力出現災難性遺忘。本研究探討並解決生成式檢索(GenRetrieval)任務中的這一挑戰。在生成式檢索微調過程中,我們發現這種遺忘迅速發生,且與微調後模型與原始模型參數之間的距離相關。基於這些觀察,我們提出ORBIT,一種新穎的方法,透過主動追蹤微調後權重與初始模型權重之間的距離,並在生成式檢索微調過程中,當此模型間距離超過最大閾值時,採用權重平均策略來限制模型漂移。結果顯示,ORBIT能保留大量的文本與檢索效能,優於常見的持續學習基線以及同樣採用權重平均的相關正則化方法。
English
Despite the rapid advancements in large language model (LLM) development, fine-tuning them for specific tasks often results in the catastrophic forgetting of their general, language-based reasoning abilities. This work investigates and addresses this challenge in the context of the Generative Retrieval (GenRetrieval) task. During GenRetrieval fine-tuning, we find this forgetting occurs rapidly and correlates with the distance between the fine-tuned and original model parameters. Given these observations, we propose ORBIT, a novel approach that actively tracks the distance between fine-tuned and initial model weights, and uses a weight averaging strategy to constrain model drift during GenRetrieval fine-tuning when this inter-model distance exceeds a maximum threshold. Our results show that ORBIT retains substantial text and retrieval performance by outperforming both common continual learning baselines and related regularization methods that also employ weight averaging.
PDF11May 14, 2026