The performance of full-parameter finetuning

#13
by stephenshuang - opened

Will the performance of full-parameter finetuning be better than lora style finetuning?

Alibaba-NLP org

Our experiments reveal that full-parameter fine-tuning tends to yield better results on in-domain test datasets compared to LoRA-style fine-tuning. However, it experiences a significant performance drop when applied to out-of-domain datasets. Consequently, we recommend adopting the LoRA training approach, which not only conserves training resources but also preserves more of the foundational model's capabilities, such as handling long texts and multilingual tasks, among others.

Sign up or log in to comment