Causally Sufficient and Necessary Feature Expansion for Class-Incremental Learning
📰 ArXiv cs.AI
arXiv:2603.09145v2 Announce Type: replace-cross Abstract: Current expansion-based methods for Class Incremental Learning (CIL) effectively mitigate catastrophic forgetting by freezing old features. However, such task-specific features learned from the new task may collide with the old features. From a causal perspective, spurious feature correlations are the main cause of this collision, manifesting in two scopes: (i) guided by empirical risk minimization (ERM), intra-task spurious correlations
DeepCamp AI