MIT researchers unveil a new fine-tuning method that lets enterprises consolidate their "model zoos" into a single, continuously learning agent.
MIT introduces Self-Distillation Fine-Tuning to reduce catastrophic forgetting; it uses student-teacher demonstrations and needs 2.5x compute.
Some results have been hidden because they may be inaccessible to you
Show inaccessible results