CURLoRA: Stable LLM Continual Fine-tuning and Catastrophic Forgetting Mitigation

CURLoRA: Leveraging CUR Matrix Decomposition for Stable LLM Continual Fine-Tuning and Catastrophic Forgetting Mitigation

Muhammad Fawi ORCID iD icon

This paper introduces CURLoRA, a novel approach to fine-tuning large language models (LLMs) that leverages CUR matrix decomposition in the context of Low-Rank Adaptation (LoRA). Our method addresses two critical challenges in LLM fine-tuning: mitigating catastrophic forgetting during continual learning and reducing the number of trainable parameters. We propose a unique modification to the CUR decomposition process to enable a more efficient and stable way to adapt LLMs to new tasks without compromising any existing knowledge.  We demonstrate through experiments on multiple datasets that CURLoRA outperforms standard LoRA in mitigating catastrophic forgetting. It maintains model stability and performance across tasks while significantly reducing the number of trainable parameters. Our results show that CURLoRA achieves superior accuracy and perplexity scores compared to LoRA upon continual fine-tuning, particularly in scenarios with limited data.

Keywords: Artificial Intelligence, Large Language Models, Catastrophic Forgetting, Matrix Decomposition, Low-Rank Adaptation

Preprint Link: https://arxiv.org/abs/2408.14572

Code link:         https://github.com/MNoorFawi/curlora

License:              MIT License

Share

Leave a Reply

Your email address will not be published. Required fields are marked *