Mitigating Forgetting in LLM Supervised Fine-Tuning and Preference Learning

Kavli Affiliate: Yi Zhou | First 5 Authors: Heshan Fernando, Han Shen, Parikshit Ram, Yi Zhou, Horst Samulowitz | Summary: Post-training of pre-trained LLMs, which typically consists of the supervised fine-tuning (SFT) stage and the preference learning (RLHF or DPO) stage, is crucial to effective and safe LLM applications. The widely adopted approach in post-training […]


Continue.. Mitigating Forgetting in LLM Supervised Fine-Tuning and Preference Learning