Routing-Based Continual Learning for Multimodal Large Language Models
📰 ArXiv cs.AI
arXiv:2511.01831v3 Announce Type: replace-cross Abstract: Multimodal Large Language Models (MLLMs) struggle with continual learning, often suffering from catastrophic forgetting when adapting to sequential tasks. We introduce a routing-based architecture that integrates new capabilities while robustly preserving foundational knowledge. While Multi-Task Learning (MTL) offers a theoretical performance upper bound, it incurs a linearly scaling computational overhead as the number of tasks increases
DeepCamp AI