Research Reveals How to Teach AI Models New Skills Without Forgetting Old Ones
A recent study explores methods for teaching large multimodal models new abilities while preserving their existing knowledge. Researchers focused on sequential fine-tuning across five specific skills and monitored overall performance across eight different benchmarks. The investigation involved three unique model families. Surprisingly, it was discovered that performance on previously held-out tasks could partially recover after fine-tuning on a different skill, linked to a notable change in the model's output token distribution. A straightforward counting-bias probe indicated that this change correlates with forgetting. The study proposed two effective tuning strategies that enhance learning while minimizing detrimental drift. The first method adjusts only the self-attention projection layers, yielding a +24.9 learning gain and a -0.6 reduction in held-out forgetting. The second method modifies only the MLP Gate and Up projections while keeping the Down projection static, achieving +30.5 and -2.1 gains respectively. Both strategies significantly surpassed full fine-tuning. This research was published on arXiv with the identifier 2510.08564v2, categorized as a 'replace' announcement.
Key facts
- The study focuses on teaching new skills to large multimodal models without catastrophic forgetting.
- Researchers used sequential fine-tuning on five target skills.
- General ability was monitored on eight held-out benchmarks.
- Three different model families were analyzed.
- Performance lost on held-out tasks can partly recover when tuning on a different skill.
- A shift in the output token distribution was identified as a key factor.
- A counting-bias probe showed the shift co-varies with forgetting.
- Two effective tuning recipes were identified: updating only SA Proj. layers and updating only MLP Gate&Up while freezing Down projection.
Entities
Institutions
- arXiv