Replies: 1 comment
-
This is a pretty challenging thing, and despite lots of work on avoiding catastrophic forgetting i nthe literature, there doesn't seem to be a silver bullet for this. That said, the most convincing mechanism (in terms of efficacy and simplicity) that I've seen so far is what people do in RLHF for instance: adding an additional piece to the loss that penalizes the model from producing distributions with a high KL Divergence from the distribution of the "original" (as in previous) model. |
Beta Was this translation helpful? Give feedback.
-
Hi all,
I was wondering if we could fine tune an LLM with 1K samples and saving the model and later add additional 500 samples while preserving the knowledge from previous 1K samples in Ludwig during the next fine tuning phase
Is there any way to do that?
Thanks in advance
Beta Was this translation helpful? Give feedback.
All reactions