LoRAMoE: Alleviating World Knowledge Forgetting in Large Language Models Via MoE-Style Plugin.
Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1 Long Papers)(2024)
摘要
Supervised fine-tuning (SFT) is a crucial step for large language models(LLMs), enabling them to align with human instructions and enhance theircapabilities in downstream tasks. Increasing instruction data substantially isa direct solution to align the model with a broader range of downstream tasksor notably improve its performance on a specific task. However, we find thatlarge-scale increases in instruction data can damage the world knowledgepreviously stored in LLMs. To address this challenge, we propose LoRAMoE, anovelty framework that introduces several low-rank adapters (LoRA) andintegrates them by using a router network, like a plugin version of Mixture ofExperts (MoE). It freezes the backbone model and forces a portion of LoRAs tofocus on leveraging world knowledge to solve downstream tasks, to alleviateworld knowledge-edge forgetting. Experimental results show that, as theinstruction data increases, LoRAMoE can significantly improve the ability toprocess downstream tasks, while maintaining the world knowledge stored in theLLM.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
数据免责声明
页面数据均来自互联网公开来源、合作出版商和通过AI技术自动分析结果,我们不对页面数据的有效性、准确性、正确性、可靠性、完整性和及时性做出任何承诺和保证。若有疑问,可以通过电子邮件方式联系我们:report@aminer.cn